Uses of Class
org.apache.hadoop.mapreduce.TaskAttemptID

Packages that use TaskAttemptID
org.apache.hadoop.mapred   
org.apache.hadoop.mapreduce   
org.apache.hadoop.mapreduce.jobhistory   
org.apache.hadoop.mapreduce.lib.map   
org.apache.hadoop.mapreduce.lib.reduce   
org.apache.hadoop.mapreduce.protocol   
org.apache.hadoop.mapreduce.task   
org.apache.hadoop.mapreduce.task.reduce   
org.apache.hadoop.mapreduce.tools   
 

Uses of TaskAttemptID in org.apache.hadoop.mapred
 

Subclasses of TaskAttemptID in org.apache.hadoop.mapred
 class TaskAttemptID
          TaskAttemptID represents the immutable and unique identifier for a task attempt.
 

Methods in org.apache.hadoop.mapred that return TaskAttemptID
 TaskAttemptID ShuffleConsumerPlugin.Context.getReduceId()
           
 

Methods in org.apache.hadoop.mapred with parameters of type TaskAttemptID
protected static
<INKEY,INVALUE,OUTKEY,OUTVALUE>
Reducer.Context
Task.createReduceContext(Reducer<INKEY,INVALUE,OUTKEY,OUTVALUE> reducer, org.apache.hadoop.conf.Configuration job, TaskAttemptID taskId, RawKeyValueIterator rIter, Counter inputKeyCounter, Counter inputValueCounter, RecordWriter<OUTKEY,OUTVALUE> output, OutputCommitter committer, StatusReporter reporter, org.apache.hadoop.io.RawComparator<INKEY> comparator, Class<INKEY> keyClass, Class<INVALUE> valueClass)
           
static TaskAttemptID TaskAttemptID.downgrade(TaskAttemptID old)
          Downgrade a new TaskAttemptID to an old one
 

Constructors in org.apache.hadoop.mapred with parameters of type TaskAttemptID
BackupStore(org.apache.hadoop.conf.Configuration conf, TaskAttemptID taskid)
           
ShuffleConsumerPlugin.Context(TaskAttemptID reduceId, JobConf jobConf, org.apache.hadoop.fs.FileSystem localFS, TaskUmbilicalProtocol umbilical, org.apache.hadoop.fs.LocalDirAllocator localDirAllocator, Reporter reporter, org.apache.hadoop.io.compress.CompressionCodec codec, Class<? extends Reducer> combinerClass, Task.CombineOutputCollector<K,V> combineCollector, Counters.Counter spilledRecordsCounter, Counters.Counter reduceCombineInputCounter, Counters.Counter shuffledMapsCounter, Counters.Counter reduceShuffleBytes, Counters.Counter failedShuffleCounter, Counters.Counter mergedMapOutputsCounter, TaskStatus status, org.apache.hadoop.util.Progress copyPhase, org.apache.hadoop.util.Progress mergePhase, Task reduceTask, MapOutputFile mapOutputFile)
           
 

Uses of TaskAttemptID in org.apache.hadoop.mapreduce
 

Methods in org.apache.hadoop.mapreduce that return TaskAttemptID
static TaskAttemptID TaskAttemptID.forName(String str)
          Construct a TaskAttemptID object from given string
 TaskAttemptID TaskReport.getSuccessfulTaskAttemptId()
          Get the attempt ID that took this task to completion
 TaskAttemptID TaskCompletionEvent.getTaskAttemptId()
          Returns task id.
 TaskAttemptID TaskAttemptContext.getTaskAttemptID()
          Get the unique name for this task attempt.
 

Methods in org.apache.hadoop.mapreduce that return types with arguments of type TaskAttemptID
 Collection<TaskAttemptID> TaskReport.getRunningTaskAttemptIds()
          Get the running task attempt IDs for this task
 

Methods in org.apache.hadoop.mapreduce with parameters of type TaskAttemptID
 void Job.failTask(TaskAttemptID taskId)
          Fail indicated task attempt.
 LogParams Cluster.getLogParams(JobID jobID, TaskAttemptID taskAttemptID)
          Get log parameters for the specified jobID or taskAttemptID
 String[] Job.getTaskDiagnostics(TaskAttemptID taskid)
          Gets the diagnostic messages for a given task attempt.
 void Job.killTask(TaskAttemptID taskId)
          Kill indicated task attempt.
 boolean Job.killTask(TaskAttemptID taskId, boolean shouldFail)
          Kill indicated task attempt.
protected  void TaskReport.setSuccessfulAttemptId(TaskAttemptID t)
          set successful attempt ID of the task.
protected  void TaskCompletionEvent.setTaskAttemptId(TaskAttemptID taskId)
          Sets task id.
 

Method parameters in org.apache.hadoop.mapreduce with type arguments of type TaskAttemptID
protected  void TaskReport.setRunningTaskAttemptIds(Collection<TaskAttemptID> runningAttempts)
          set running attempt(s) of the task.
 

Constructors in org.apache.hadoop.mapreduce with parameters of type TaskAttemptID
TaskCompletionEvent(int eventId, TaskAttemptID taskId, int idWithinJob, boolean isMap, TaskCompletionEvent.Status status, String taskTrackerHttp)
          Constructor.
 

Uses of TaskAttemptID in org.apache.hadoop.mapreduce.jobhistory
 

Methods in org.apache.hadoop.mapreduce.jobhistory that return TaskAttemptID
 TaskAttemptID ReduceAttemptFinishedEvent.getAttemptId()
          Get the attempt id
 TaskAttemptID TaskAttemptFinishedEvent.getAttemptId()
          Get the task attempt id
 TaskAttemptID JobHistoryParser.TaskAttemptInfo.getAttemptId()
           
 TaskAttemptID MapAttemptFinishedEvent.getAttemptId()
          Get the attempt id
 TaskAttemptID TaskFailedEvent.getFailedAttemptID()
          Get the attempt id due to which the task failed
 TaskAttemptID JobHistoryParser.TaskInfo.getFailedDueToAttemptId()
           
 TaskAttemptID JobHistoryParser.TaskInfo.getSuccessfulAttemptId()
           
 TaskAttemptID TaskFinishedEvent.getSuccessfulTaskAttemptId()
          Get successful task attempt id
 TaskAttemptID TaskAttemptUnsuccessfulCompletionEvent.getTaskAttemptId()
          Get the attempt id
 TaskAttemptID TaskAttemptStartedEvent.getTaskAttemptId()
          Get the attempt id
 

Methods in org.apache.hadoop.mapreduce.jobhistory that return types with arguments of type TaskAttemptID
 Map<TaskAttemptID,JobHistoryParser.TaskAttemptInfo> JobHistoryParser.JobInfo.getAllCompletedTaskAttempts()
           
 Map<TaskAttemptID,JobHistoryParser.TaskAttemptInfo> JobHistoryParser.TaskInfo.getAllTaskAttempts()
           
 

Constructors in org.apache.hadoop.mapreduce.jobhistory with parameters of type TaskAttemptID
MapAttemptFinishedEvent(TaskAttemptID id, TaskType taskType, String taskStatus, long mapFinishTime, long finishTime, String hostname, int port, String rackName, String state, Counters counters, int[][] allSplits)
          Create an event for successful completion of map attempts
MapAttemptFinishedEvent(TaskAttemptID id, TaskType taskType, String taskStatus, long mapFinishTime, long finishTime, String hostname, String state, Counters counters)
          Deprecated. please use the constructor with an additional argument, an array of splits arrays instead. See ProgressSplitsBlock for an explanation of the meaning of that parameter. Create an event for successful completion of map attempts
ReduceAttemptFinishedEvent(TaskAttemptID id, TaskType taskType, String taskStatus, long shuffleFinishTime, long sortFinishTime, long finishTime, String hostname, int port, String rackName, String state, Counters counters, int[][] allSplits)
          Create an event to record completion of a reduce attempt
ReduceAttemptFinishedEvent(TaskAttemptID id, TaskType taskType, String taskStatus, long shuffleFinishTime, long sortFinishTime, long finishTime, String hostname, String state, Counters counters)
          Deprecated. please use the constructor with an additional argument, an array of splits arrays instead. See ProgressSplitsBlock for an explanation of the meaning of that parameter. Create an event to record completion of a reduce attempt
TaskAttemptFinishedEvent(TaskAttemptID id, TaskType taskType, String taskStatus, long finishTime, String rackName, String hostname, String state, Counters counters)
          Create an event to record successful finishes for setup and cleanup attempts
TaskAttemptStartedEvent(TaskAttemptID attemptId, TaskType taskType, long startTime, String trackerName, int httpPort, int shufflePort, org.apache.hadoop.yarn.api.records.ContainerId containerId, String locality, String avataar)
          Create an event to record the start of an attempt
TaskAttemptStartedEvent(TaskAttemptID attemptId, TaskType taskType, long startTime, String trackerName, int httpPort, int shufflePort, String locality, String avataar)
           
TaskAttemptUnsuccessfulCompletionEvent(TaskAttemptID id, TaskType taskType, String status, long finishTime, String hostname, int port, String rackName, String error, Counters counters, int[][] allSplits)
          Create an event to record the unsuccessful completion of attempts
TaskAttemptUnsuccessfulCompletionEvent(TaskAttemptID id, TaskType taskType, String status, long finishTime, String hostname, int port, String rackName, String error, int[][] allSplits)
           
TaskAttemptUnsuccessfulCompletionEvent(TaskAttemptID id, TaskType taskType, String status, long finishTime, String hostname, String error)
          Deprecated. please use the constructor with an additional argument, an array of splits arrays instead. See ProgressSplitsBlock for an explanation of the meaning of that parameter. Create an event to record the unsuccessful completion of attempts
TaskFailedEvent(TaskID id, long finishTime, TaskType taskType, String error, String status, TaskAttemptID failedDueToAttempt)
           
TaskFailedEvent(TaskID id, long finishTime, TaskType taskType, String error, String status, TaskAttemptID failedDueToAttempt, Counters counters)
          Create an event to record task failure
TaskFinishedEvent(TaskID id, TaskAttemptID attemptId, long finishTime, TaskType taskType, String status, Counters counters)
          Create an event to record the successful completion of a task
 

Uses of TaskAttemptID in org.apache.hadoop.mapreduce.lib.map
 

Methods in org.apache.hadoop.mapreduce.lib.map that return TaskAttemptID
 TaskAttemptID WrappedMapper.Context.getTaskAttemptID()
           
 

Uses of TaskAttemptID in org.apache.hadoop.mapreduce.lib.reduce
 

Methods in org.apache.hadoop.mapreduce.lib.reduce that return TaskAttemptID
 TaskAttemptID WrappedReducer.Context.getTaskAttemptID()
           
 

Uses of TaskAttemptID in org.apache.hadoop.mapreduce.protocol
 

Methods in org.apache.hadoop.mapreduce.protocol with parameters of type TaskAttemptID
 LogParams ClientProtocol.getLogFileParams(JobID jobID, TaskAttemptID taskAttemptID)
          Gets the location of the log file for a job if no taskAttemptId is specified, otherwise gets the log location for the taskAttemptId.
 String[] ClientProtocol.getTaskDiagnostics(TaskAttemptID taskId)
          Get the diagnostics for a given task in a given job
 boolean ClientProtocol.killTask(TaskAttemptID taskId, boolean shouldFail)
          Kill indicated task attempt.
 

Uses of TaskAttemptID in org.apache.hadoop.mapreduce.task
 

Methods in org.apache.hadoop.mapreduce.task that return TaskAttemptID
 TaskAttemptID TaskAttemptContextImpl.getTaskAttemptID()
          Get the unique name for this task attempt.
 

Constructors in org.apache.hadoop.mapreduce.task with parameters of type TaskAttemptID
MapContextImpl(org.apache.hadoop.conf.Configuration conf, TaskAttemptID taskid, RecordReader<KEYIN,VALUEIN> reader, RecordWriter<KEYOUT,VALUEOUT> writer, OutputCommitter committer, StatusReporter reporter, InputSplit split)
           
ReduceContextImpl(org.apache.hadoop.conf.Configuration conf, TaskAttemptID taskid, RawKeyValueIterator input, Counter inputKeyCounter, Counter inputValueCounter, RecordWriter<KEYOUT,VALUEOUT> output, OutputCommitter committer, StatusReporter reporter, org.apache.hadoop.io.RawComparator<KEYIN> comparator, Class<KEYIN> keyClass, Class<VALUEIN> valueClass)
           
TaskAttemptContextImpl(org.apache.hadoop.conf.Configuration conf, TaskAttemptID taskId)
           
TaskAttemptContextImpl(org.apache.hadoop.conf.Configuration conf, TaskAttemptID taskId, StatusReporter reporter)
           
TaskInputOutputContextImpl(org.apache.hadoop.conf.Configuration conf, TaskAttemptID taskid, RecordWriter<KEYOUT,VALUEOUT> output, OutputCommitter committer, StatusReporter reporter)
           
 

Uses of TaskAttemptID in org.apache.hadoop.mapreduce.task.reduce
 

Methods in org.apache.hadoop.mapreduce.task.reduce that return TaskAttemptID
 TaskAttemptID MapOutput.getMapId()
           
 

Methods in org.apache.hadoop.mapreduce.task.reduce that return types with arguments of type TaskAttemptID
 List<TaskAttemptID> MapHost.getAndClearKnownMaps()
           
 List<TaskAttemptID> ShuffleSchedulerImpl.getMapsForHost(MapHost host)
           
 

Methods in org.apache.hadoop.mapreduce.task.reduce with parameters of type TaskAttemptID
 void MapHost.addKnownMap(TaskAttemptID mapId)
           
 void ShuffleSchedulerImpl.addKnownMapOutput(String hostName, String hostUrl, TaskAttemptID mapId)
           
 void ShuffleSchedulerImpl.copyFailed(TaskAttemptID mapId, MapHost host, boolean readError, boolean connectExcpt)
           
 void ShuffleSchedulerImpl.copySucceeded(TaskAttemptID mapId, MapHost host, long bytes, long millis, MapOutput<K,V> output)
           
 void ShuffleSchedulerImpl.obsoleteMapOutput(TaskAttemptID mapId)
           
 void ShuffleSchedulerImpl.putBackKnownMapOutput(MapHost host, TaskAttemptID mapId)
           
 MapOutput<K,V> MergeManager.reserve(TaskAttemptID mapId, long requestedSize, int fetcher)
          To reserve resources for data to be shuffled.
 MapOutput<K,V> MergeManagerImpl.reserve(TaskAttemptID mapId, long requestedSize, int fetcher)
           
 

Constructors in org.apache.hadoop.mapreduce.task.reduce with parameters of type TaskAttemptID
InMemoryReader(MergeManagerImpl<K,V> merger, TaskAttemptID taskAttemptId, byte[] data, int start, int length, org.apache.hadoop.conf.Configuration conf)
           
MapOutput(TaskAttemptID mapId, long size, boolean primaryMapOutput)
           
MergeManagerImpl(TaskAttemptID reduceId, JobConf jobConf, org.apache.hadoop.fs.FileSystem localFS, org.apache.hadoop.fs.LocalDirAllocator localDirAllocator, Reporter reporter, org.apache.hadoop.io.compress.CompressionCodec codec, Class<? extends Reducer> combinerClass, Task.CombineOutputCollector<K,V> combineCollector, Counters.Counter spilledRecordsCounter, Counters.Counter reduceCombineInputCounter, Counters.Counter mergedMapOutputsCounter, ExceptionReporter exceptionReporter, org.apache.hadoop.util.Progress mergePhase, MapOutputFile mapOutputFile)
           
ShuffleSchedulerImpl(JobConf job, TaskStatus status, TaskAttemptID reduceId, ExceptionReporter reporter, org.apache.hadoop.util.Progress progress, Counters.Counter shuffledMapsCounter, Counters.Counter reduceShuffleBytes, Counters.Counter failedShuffleCounter)
           
 

Uses of TaskAttemptID in org.apache.hadoop.mapreduce.tools
 

Methods in org.apache.hadoop.mapreduce.tools with parameters of type TaskAttemptID
protected static String CLI.getTaskLogURL(TaskAttemptID taskId, String baseUrl)
           
 



Copyright © 2013 Apache Software Foundation. All Rights Reserved.