|
||||||||||
PREV NEXT | FRAMES NO FRAMES |
Uses of TaskAttemptID in org.apache.hadoop.mapred |
---|
Subclasses of TaskAttemptID in org.apache.hadoop.mapred | |
---|---|
class |
TaskAttemptID
TaskAttemptID represents the immutable and unique identifier for a task attempt. |
Methods in org.apache.hadoop.mapred with parameters of type TaskAttemptID | ||
---|---|---|
protected static
|
Task.createReduceContext(Reducer<INKEY,INVALUE,OUTKEY,OUTVALUE> reducer,
org.apache.hadoop.conf.Configuration job,
TaskAttemptID taskId,
RawKeyValueIterator rIter,
Counter inputKeyCounter,
Counter inputValueCounter,
RecordWriter<OUTKEY,OUTVALUE> output,
OutputCommitter committer,
StatusReporter reporter,
org.apache.hadoop.io.RawComparator<INKEY> comparator,
Class<INKEY> keyClass,
Class<INVALUE> valueClass)
|
|
static TaskAttemptID |
TaskAttemptID.downgrade(TaskAttemptID old)
Downgrade a new TaskAttemptID to an old one |
Constructors in org.apache.hadoop.mapred with parameters of type TaskAttemptID | |
---|---|
BackupStore(org.apache.hadoop.conf.Configuration conf,
TaskAttemptID taskid)
|
Uses of TaskAttemptID in org.apache.hadoop.mapreduce |
---|
Methods in org.apache.hadoop.mapreduce that return TaskAttemptID | |
---|---|
static TaskAttemptID |
TaskAttemptID.forName(String str)
Construct a TaskAttemptID object from given string |
TaskAttemptID |
TaskReport.getSuccessfulTaskAttemptId()
Get the attempt ID that took this task to completion |
TaskAttemptID |
TaskCompletionEvent.getTaskAttemptId()
Returns task id. |
TaskAttemptID |
TaskAttemptContext.getTaskAttemptID()
Get the unique name for this task attempt. |
Methods in org.apache.hadoop.mapreduce that return types with arguments of type TaskAttemptID | |
---|---|
Collection<TaskAttemptID> |
TaskReport.getRunningTaskAttemptIds()
Get the running task attempt IDs for this task |
Methods in org.apache.hadoop.mapreduce with parameters of type TaskAttemptID | |
---|---|
boolean |
Job.failTask(TaskAttemptID taskId)
Fail indicated task attempt. |
LogParams |
Cluster.getLogParams(JobID jobID,
TaskAttemptID taskAttemptID)
Get log parameters for the specified jobID or taskAttemptID |
String[] |
Job.getTaskDiagnostics(TaskAttemptID taskid)
Gets the diagnostic messages for a given task attempt. |
boolean |
Job.killTask(TaskAttemptID taskId)
Kill indicated task attempt. |
protected void |
TaskReport.setSuccessfulAttemptId(TaskAttemptID t)
set successful attempt ID of the task. |
protected void |
TaskCompletionEvent.setTaskAttemptId(TaskAttemptID taskId)
Sets task id. |
Method parameters in org.apache.hadoop.mapreduce with type arguments of type TaskAttemptID | |
---|---|
protected void |
TaskReport.setRunningTaskAttemptIds(Collection<TaskAttemptID> runningAttempts)
set running attempt(s) of the task. |
Constructors in org.apache.hadoop.mapreduce with parameters of type TaskAttemptID | |
---|---|
TaskCompletionEvent(int eventId,
TaskAttemptID taskId,
int idWithinJob,
boolean isMap,
TaskCompletionEvent.Status status,
String taskTrackerHttp)
Constructor. |
Uses of TaskAttemptID in org.apache.hadoop.mapreduce.jobhistory |
---|
Methods in org.apache.hadoop.mapreduce.jobhistory that return TaskAttemptID | |
---|---|
TaskAttemptID |
TaskAttemptFinishedEvent.getAttemptId()
Get the task attempt id |
TaskAttemptID |
ReduceAttemptFinishedEvent.getAttemptId()
Get the attempt id |
TaskAttemptID |
MapAttemptFinishedEvent.getAttemptId()
Get the attempt id |
TaskAttemptID |
JobHistoryParser.TaskAttemptInfo.getAttemptId()
|
TaskAttemptID |
TaskFailedEvent.getFailedAttemptID()
Get the attempt id due to which the task failed |
TaskAttemptID |
JobHistoryParser.TaskInfo.getFailedDueToAttemptId()
|
TaskAttemptID |
JobHistoryParser.TaskInfo.getSuccessfulAttemptId()
|
TaskAttemptID |
TaskFinishedEvent.getSuccessfulTaskAttemptId()
Get successful task attempt id |
TaskAttemptID |
TaskAttemptStartedEvent.getTaskAttemptId()
Get the attempt id |
TaskAttemptID |
TaskAttemptUnsuccessfulCompletionEvent.getTaskAttemptId()
Get the attempt id |
Methods in org.apache.hadoop.mapreduce.jobhistory that return types with arguments of type TaskAttemptID | |
---|---|
Map<TaskAttemptID,JobHistoryParser.TaskAttemptInfo> |
JobHistoryParser.TaskInfo.getAllTaskAttempts()
|
Constructors in org.apache.hadoop.mapreduce.jobhistory with parameters of type TaskAttemptID | |
---|---|
MapAttemptFinishedEvent(TaskAttemptID id,
TaskType taskType,
String taskStatus,
long mapFinishTime,
long finishTime,
String hostname,
int port,
String rackName,
String state,
Counters counters,
int[][] allSplits)
Create an event for successful completion of map attempts |
|
MapAttemptFinishedEvent(TaskAttemptID id,
TaskType taskType,
String taskStatus,
long mapFinishTime,
long finishTime,
String hostname,
String state,
Counters counters)
Deprecated. please use the constructor with an additional argument, an array of splits arrays instead. See ProgressSplitsBlock
for an explanation of the meaning of that parameter.
Create an event for successful completion of map attempts |
|
ReduceAttemptFinishedEvent(TaskAttemptID id,
TaskType taskType,
String taskStatus,
long shuffleFinishTime,
long sortFinishTime,
long finishTime,
String hostname,
int port,
String rackName,
String state,
Counters counters,
int[][] allSplits)
Create an event to record completion of a reduce attempt |
|
ReduceAttemptFinishedEvent(TaskAttemptID id,
TaskType taskType,
String taskStatus,
long shuffleFinishTime,
long sortFinishTime,
long finishTime,
String hostname,
String state,
Counters counters)
Deprecated. please use the constructor with an additional argument, an array of splits arrays instead. See ProgressSplitsBlock
for an explanation of the meaning of that parameter.
Create an event to record completion of a reduce attempt |
|
TaskAttemptFinishedEvent(TaskAttemptID id,
TaskType taskType,
String taskStatus,
long finishTime,
String rackName,
String hostname,
String state,
Counters counters)
Create an event to record successful finishes for setup and cleanup attempts |
|
TaskAttemptStartedEvent(TaskAttemptID attemptId,
TaskType taskType,
long startTime,
String trackerName,
int httpPort,
int shufflePort)
|
|
TaskAttemptStartedEvent(TaskAttemptID attemptId,
TaskType taskType,
long startTime,
String trackerName,
int httpPort,
int shufflePort,
org.apache.hadoop.yarn.api.records.ContainerId containerId)
Create an event to record the start of an attempt |
|
TaskAttemptUnsuccessfulCompletionEvent(TaskAttemptID id,
TaskType taskType,
String status,
long finishTime,
String hostname,
int port,
String rackName,
String error,
int[][] allSplits)
Create an event to record the unsuccessful completion of attempts |
|
TaskAttemptUnsuccessfulCompletionEvent(TaskAttemptID id,
TaskType taskType,
String status,
long finishTime,
String hostname,
String error)
Deprecated. please use the constructor with an additional argument, an array of splits arrays instead. See ProgressSplitsBlock
for an explanation of the meaning of that parameter.
Create an event to record the unsuccessful completion of attempts |
|
TaskFailedEvent(TaskID id,
long finishTime,
TaskType taskType,
String error,
String status,
TaskAttemptID failedDueToAttempt)
Create an event to record task failure |
|
TaskFinishedEvent(TaskID id,
TaskAttemptID attemptId,
long finishTime,
TaskType taskType,
String status,
Counters counters)
Create an event to record the successful completion of a task |
Uses of TaskAttemptID in org.apache.hadoop.mapreduce.lib.map |
---|
Methods in org.apache.hadoop.mapreduce.lib.map that return TaskAttemptID | |
---|---|
TaskAttemptID |
WrappedMapper.Context.getTaskAttemptID()
|
Uses of TaskAttemptID in org.apache.hadoop.mapreduce.lib.reduce |
---|
Methods in org.apache.hadoop.mapreduce.lib.reduce that return TaskAttemptID | |
---|---|
TaskAttemptID |
WrappedReducer.Context.getTaskAttemptID()
|
Uses of TaskAttemptID in org.apache.hadoop.mapreduce.protocol |
---|
Methods in org.apache.hadoop.mapreduce.protocol with parameters of type TaskAttemptID | |
---|---|
LogParams |
ClientProtocol.getLogFileParams(JobID jobID,
TaskAttemptID taskAttemptID)
Gets the location of the log file for a job if no taskAttemptId is specified, otherwise gets the log location for the taskAttemptId. |
String[] |
ClientProtocol.getTaskDiagnostics(TaskAttemptID taskId)
Get the diagnostics for a given task in a given job |
boolean |
ClientProtocol.killTask(TaskAttemptID taskId,
boolean shouldFail)
Kill indicated task attempt. |
Uses of TaskAttemptID in org.apache.hadoop.mapreduce.task |
---|
Methods in org.apache.hadoop.mapreduce.task that return TaskAttemptID | |
---|---|
TaskAttemptID |
TaskAttemptContextImpl.getTaskAttemptID()
Get the unique name for this task attempt. |
Constructors in org.apache.hadoop.mapreduce.task with parameters of type TaskAttemptID | |
---|---|
MapContextImpl(org.apache.hadoop.conf.Configuration conf,
TaskAttemptID taskid,
RecordReader<KEYIN,VALUEIN> reader,
RecordWriter<KEYOUT,VALUEOUT> writer,
OutputCommitter committer,
StatusReporter reporter,
InputSplit split)
|
|
ReduceContextImpl(org.apache.hadoop.conf.Configuration conf,
TaskAttemptID taskid,
RawKeyValueIterator input,
Counter inputKeyCounter,
Counter inputValueCounter,
RecordWriter<KEYOUT,VALUEOUT> output,
OutputCommitter committer,
StatusReporter reporter,
org.apache.hadoop.io.RawComparator<KEYIN> comparator,
Class<KEYIN> keyClass,
Class<VALUEIN> valueClass)
|
|
TaskAttemptContextImpl(org.apache.hadoop.conf.Configuration conf,
TaskAttemptID taskId)
|
|
TaskAttemptContextImpl(org.apache.hadoop.conf.Configuration conf,
TaskAttemptID taskId,
StatusReporter reporter)
|
|
TaskInputOutputContextImpl(org.apache.hadoop.conf.Configuration conf,
TaskAttemptID taskid,
RecordWriter<KEYOUT,VALUEOUT> output,
OutputCommitter committer,
StatusReporter reporter)
|
Uses of TaskAttemptID in org.apache.hadoop.mapreduce.task.reduce |
---|
Methods in org.apache.hadoop.mapreduce.task.reduce with parameters of type TaskAttemptID | |
---|---|
org.apache.hadoop.mapreduce.task.reduce.MapOutput<K,V> |
MergeManager.reserve(TaskAttemptID mapId,
long requestedSize,
int fetcher)
|
Constructors in org.apache.hadoop.mapreduce.task.reduce with parameters of type TaskAttemptID | |
---|---|
InMemoryReader(MergeManager<K,V> merger,
TaskAttemptID taskAttemptId,
byte[] data,
int start,
int length)
|
|
MergeManager(TaskAttemptID reduceId,
JobConf jobConf,
org.apache.hadoop.fs.FileSystem localFS,
org.apache.hadoop.fs.LocalDirAllocator localDirAllocator,
Reporter reporter,
org.apache.hadoop.io.compress.CompressionCodec codec,
Class<? extends Reducer> combinerClass,
Task.CombineOutputCollector<K,V> combineCollector,
Counters.Counter spilledRecordsCounter,
Counters.Counter reduceCombineInputCounter,
Counters.Counter mergedMapOutputsCounter,
org.apache.hadoop.mapreduce.task.reduce.ExceptionReporter exceptionReporter,
org.apache.hadoop.util.Progress mergePhase,
MapOutputFile mapOutputFile)
|
|
Shuffle(TaskAttemptID reduceId,
JobConf jobConf,
org.apache.hadoop.fs.FileSystem localFS,
TaskUmbilicalProtocol umbilical,
org.apache.hadoop.fs.LocalDirAllocator localDirAllocator,
Reporter reporter,
org.apache.hadoop.io.compress.CompressionCodec codec,
Class<? extends Reducer> combinerClass,
Task.CombineOutputCollector<K,V> combineCollector,
Counters.Counter spilledRecordsCounter,
Counters.Counter reduceCombineInputCounter,
Counters.Counter shuffledMapsCounter,
Counters.Counter reduceShuffleBytes,
Counters.Counter failedShuffleCounter,
Counters.Counter mergedMapOutputsCounter,
TaskStatus status,
org.apache.hadoop.util.Progress copyPhase,
org.apache.hadoop.util.Progress mergePhase,
Task reduceTask,
MapOutputFile mapOutputFile)
|
Uses of TaskAttemptID in org.apache.hadoop.mapreduce.tools |
---|
Methods in org.apache.hadoop.mapreduce.tools with parameters of type TaskAttemptID | |
---|---|
protected static String |
CLI.getTaskLogURL(TaskAttemptID taskId,
String baseUrl)
|
|
||||||||||
PREV NEXT | FRAMES NO FRAMES |