Package | Description |
---|---|
org.apache.hadoop.mapred | |
org.apache.hadoop.mapreduce.task.reduce |
Modifier and Type | Field and Description |
---|---|
protected Counters.Counter |
Task.failedShuffleCounter |
protected Counters.Counter |
Task.CombinerRunner.inputCounter |
protected Counters.Counter |
Task.mergedMapOutputsCounter |
protected Counters.Counter |
Task.spilledRecordsCounter |
Modifier and Type | Method and Description |
---|---|
Counters.Counter |
Counters.Group.addCounter(String name,
String displayName,
long value) |
Counters.Counter |
Counters.Group.findCounter(String counterName) |
Counters.Counter |
Counters.Group.findCounter(String counterName,
boolean create) |
Counters.Counter |
Counters.findCounter(String group,
int id,
String name)
Deprecated.
use
Counters.findCounter(String, String) instead |
Counters.Counter |
Counters.findCounter(String group,
String name) |
Counters.Counter |
Counters.Group.findCounter(String counterName,
String displayName) |
Counters.Counter |
Task.TaskReporter.getCounter(Enum<?> name) |
Counters.Counter |
Reporter.getCounter(Enum<?> name)
Get the
Counters.Counter of the given group with the given name. |
Counters.Counter |
Counters.Group.getCounter(int id,
String name)
Deprecated.
use
Counters.Group.findCounter(String) instead |
Counters.Counter |
Task.TaskReporter.getCounter(String group,
String name) |
Counters.Counter |
Reporter.getCounter(String group,
String name)
Get the
Counters.Counter of the given group with the given name. |
Counters.Counter |
Counters.Group.getCounterForName(String name)
Get the counter for the given name and create it if it doesn't exist.
|
Counters.Counter |
ShuffleConsumerPlugin.Context.getFailedShuffleCounter() |
Counters.Counter |
ShuffleConsumerPlugin.Context.getMergedMapOutputsCounter() |
Counters.Counter |
ShuffleConsumerPlugin.Context.getReduceCombineInputCounter() |
Counters.Counter |
ShuffleConsumerPlugin.Context.getReduceShuffleBytes() |
Counters.Counter |
ShuffleConsumerPlugin.Context.getShuffledMapsCounter() |
Counters.Counter |
ShuffleConsumerPlugin.Context.getSpilledRecordsCounter() |
Modifier and Type | Method and Description |
---|---|
CounterGroupBase<Counters.Counter> |
Counters.Group.getUnderlyingGroup() |
Iterator<Counters.Counter> |
Counters.Group.iterator() |
Modifier and Type | Method and Description |
---|---|
void |
Counters.Group.addCounter(Counters.Counter counter) |
boolean |
Counters.Counter.contentEquals(Counters.Counter counter)
Deprecated.
|
static <K,V> Task.CombinerRunner<K,V> |
Task.CombinerRunner.create(JobConf job,
TaskAttemptID taskId,
Counters.Counter inputCounter,
Task.TaskReporter reporter,
OutputCommitter committer) |
static <K,V> RawKeyValueIterator |
Merger.merge(org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.FileSystem fs,
Class<K> keyClass,
Class<V> valueClass,
org.apache.hadoop.io.compress.CompressionCodec codec,
List<Merger.Segment<K,V>> segments,
int mergeFactor,
int inMemSegments,
org.apache.hadoop.fs.Path tmpDir,
org.apache.hadoop.io.RawComparator<K> comparator,
org.apache.hadoop.util.Progressable reporter,
boolean sortSegments,
Counters.Counter readsCounter,
Counters.Counter writesCounter,
org.apache.hadoop.util.Progress mergePhase) |
static <K,V> RawKeyValueIterator |
Merger.merge(org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.FileSystem fs,
Class<K> keyClass,
Class<V> valueClass,
org.apache.hadoop.io.compress.CompressionCodec codec,
List<Merger.Segment<K,V>> segments,
int mergeFactor,
org.apache.hadoop.fs.Path tmpDir,
org.apache.hadoop.io.RawComparator<K> comparator,
org.apache.hadoop.util.Progressable reporter,
boolean sortSegments,
Counters.Counter readsCounter,
Counters.Counter writesCounter,
org.apache.hadoop.util.Progress mergePhase,
TaskType taskType) |
static <K,V> RawKeyValueIterator |
Merger.merge(org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.FileSystem fs,
Class<K> keyClass,
Class<V> valueClass,
org.apache.hadoop.io.compress.CompressionCodec codec,
org.apache.hadoop.fs.Path[] inputs,
boolean deleteInputs,
int mergeFactor,
org.apache.hadoop.fs.Path tmpDir,
org.apache.hadoop.io.RawComparator<K> comparator,
org.apache.hadoop.util.Progressable reporter,
Counters.Counter readsCounter,
Counters.Counter writesCounter,
Counters.Counter mergedMapOutputsCounter,
org.apache.hadoop.util.Progress mergePhase) |
static <K,V> RawKeyValueIterator |
Merger.merge(org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.FileSystem fs,
Class<K> keyClass,
Class<V> valueClass,
org.apache.hadoop.io.compress.CompressionCodec codec,
org.apache.hadoop.fs.Path[] inputs,
boolean deleteInputs,
int mergeFactor,
org.apache.hadoop.fs.Path tmpDir,
org.apache.hadoop.io.RawComparator<K> comparator,
org.apache.hadoop.util.Progressable reporter,
Counters.Counter readsCounter,
Counters.Counter writesCounter,
org.apache.hadoop.util.Progress mergePhase) |
static <K,V> RawKeyValueIterator |
Merger.merge(org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.FileSystem fs,
Class<K> keyClass,
Class<V> valueClass,
List<Merger.Segment<K,V>> segments,
int mergeFactor,
int inMemSegments,
org.apache.hadoop.fs.Path tmpDir,
org.apache.hadoop.io.RawComparator<K> comparator,
org.apache.hadoop.util.Progressable reporter,
boolean sortSegments,
Counters.Counter readsCounter,
Counters.Counter writesCounter,
org.apache.hadoop.util.Progress mergePhase) |
static <K,V> RawKeyValueIterator |
Merger.merge(org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.FileSystem fs,
Class<K> keyClass,
Class<V> valueClass,
List<Merger.Segment<K,V>> segments,
int mergeFactor,
org.apache.hadoop.fs.Path tmpDir,
org.apache.hadoop.io.RawComparator<K> comparator,
org.apache.hadoop.util.Progressable reporter,
boolean sortSegments,
Counters.Counter readsCounter,
Counters.Counter writesCounter,
org.apache.hadoop.util.Progress mergePhase) |
static <K,V> RawKeyValueIterator |
Merger.merge(org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.FileSystem fs,
Class<K> keyClass,
Class<V> valueClass,
List<Merger.Segment<K,V>> segments,
int mergeFactor,
org.apache.hadoop.fs.Path tmpDir,
org.apache.hadoop.io.RawComparator<K> comparator,
org.apache.hadoop.util.Progressable reporter,
Counters.Counter readsCounter,
Counters.Counter writesCounter,
org.apache.hadoop.util.Progress mergePhase) |
Modifier and Type | Method and Description |
---|---|
void |
Counters.Group.incrAllCounters(CounterGroupBase<Counters.Counter> rightGroup) |
Constructor and Description |
---|
CombineOutputCollector(Counters.Counter outCounter,
org.apache.hadoop.util.Progressable progressable,
org.apache.hadoop.conf.Configuration conf) |
CombineValuesIterator(RawKeyValueIterator in,
org.apache.hadoop.io.RawComparator<KEY> comparator,
Class<KEY> keyClass,
Class<VALUE> valClass,
org.apache.hadoop.conf.Configuration conf,
Reporter reporter,
Counters.Counter combineInputCounter) |
Context(TaskAttemptID reduceId,
JobConf jobConf,
org.apache.hadoop.fs.FileSystem localFS,
TaskUmbilicalProtocol umbilical,
org.apache.hadoop.fs.LocalDirAllocator localDirAllocator,
Reporter reporter,
org.apache.hadoop.io.compress.CompressionCodec codec,
Class<? extends Reducer> combinerClass,
Task.CombineOutputCollector<K,V> combineCollector,
Counters.Counter spilledRecordsCounter,
Counters.Counter reduceCombineInputCounter,
Counters.Counter shuffledMapsCounter,
Counters.Counter reduceShuffleBytes,
Counters.Counter failedShuffleCounter,
Counters.Counter mergedMapOutputsCounter,
TaskStatus status,
org.apache.hadoop.util.Progress copyPhase,
org.apache.hadoop.util.Progress mergePhase,
Task reduceTask,
MapOutputFile mapOutputFile,
Map<TaskAttemptID,MapOutputFile> localMapFiles) |
OldCombinerRunner(Class<? extends Reducer<K,V,K,V>> cls,
JobConf conf,
Counters.Counter inputCounter,
Task.TaskReporter reporter) |
Reader(org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.fs.Path file,
org.apache.hadoop.io.compress.CompressionCodec codec,
Counters.Counter readsCounter)
Construct an IFile Reader.
|
Reader(org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.FSDataInputStream in,
long length,
org.apache.hadoop.io.compress.CompressionCodec codec,
Counters.Counter readsCounter)
Construct an IFile Reader.
|
Segment(org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.fs.Path file,
org.apache.hadoop.io.compress.CompressionCodec codec,
boolean preserve,
Counters.Counter mergedMapOutputsCounter) |
Segment(org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.fs.Path file,
org.apache.hadoop.io.compress.CompressionCodec codec,
boolean preserve,
Counters.Counter mergedMapOutputsCounter,
long rawDataLength) |
Segment(org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.fs.Path file,
long segmentOffset,
long segmentLength,
org.apache.hadoop.io.compress.CompressionCodec codec,
boolean preserve,
Counters.Counter mergedMapOutputsCounter) |
Segment(IFile.Reader<K,V> reader,
boolean preserve,
Counters.Counter mapOutputsCounter) |
Writer(org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.FSDataOutputStream out,
Class<K> keyClass,
Class<V> valueClass,
org.apache.hadoop.io.compress.CompressionCodec codec,
Counters.Counter writesCounter) |
Writer(org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.FSDataOutputStream out,
Class<K> keyClass,
Class<V> valueClass,
org.apache.hadoop.io.compress.CompressionCodec codec,
Counters.Counter writesCounter,
boolean ownOutputStream) |
Writer(Counters.Counter writesCounter) |
Constructor and Description |
---|
MergeManagerImpl(TaskAttemptID reduceId,
JobConf jobConf,
org.apache.hadoop.fs.FileSystem localFS,
org.apache.hadoop.fs.LocalDirAllocator localDirAllocator,
Reporter reporter,
org.apache.hadoop.io.compress.CompressionCodec codec,
Class<? extends Reducer> combinerClass,
Task.CombineOutputCollector<K,V> combineCollector,
Counters.Counter spilledRecordsCounter,
Counters.Counter reduceCombineInputCounter,
Counters.Counter mergedMapOutputsCounter,
ExceptionReporter exceptionReporter,
org.apache.hadoop.util.Progress mergePhase,
MapOutputFile mapOutputFile) |
ShuffleSchedulerImpl(JobConf job,
TaskStatus status,
TaskAttemptID reduceId,
ExceptionReporter reporter,
org.apache.hadoop.util.Progress progress,
Counters.Counter shuffledMapsCounter,
Counters.Counter reduceShuffleBytes,
Counters.Counter failedShuffleCounter) |
Copyright © 2017 Apache Software Foundation. All Rights Reserved.