|
||||||||||
PREV NEXT | FRAMES NO FRAMES |
Uses of Reporter in org.apache.hadoop.mapred |
---|
Classes in org.apache.hadoop.mapred that implement Reporter | |
---|---|
protected class |
Task.TaskReporter
|
Fields in org.apache.hadoop.mapred declared as Reporter | |
---|---|
static Reporter |
Reporter.NULL
A constant of Reporter type that does nothing. |
Methods in org.apache.hadoop.mapred with parameters of type Reporter | |
---|---|
void |
RecordWriter.close(Reporter reporter)
Close this RecordWriter to future operations. |
void |
TextOutputFormat.LineRecordWriter.close(Reporter reporter)
|
RecordReader<K,V> |
SequenceFileInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
RecordReader<org.apache.hadoop.io.LongWritable,org.apache.hadoop.io.Text> |
TextInputFormat.getRecordReader(InputSplit genericSplit,
JobConf job,
Reporter reporter)
|
RecordReader<K,V> |
SequenceFileInputFilter.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
Create a record reader for the given split |
abstract RecordReader<K,V> |
MultiFileInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
RecordReader<org.apache.hadoop.io.BytesWritable,org.apache.hadoop.io.BytesWritable> |
SequenceFileAsBinaryInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
RecordReader<K,V> |
InputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
Get the RecordReader for the given InputSplit . |
RecordReader<org.apache.hadoop.io.Text,org.apache.hadoop.io.Text> |
KeyValueTextInputFormat.getRecordReader(InputSplit genericSplit,
JobConf job,
Reporter reporter)
|
abstract RecordReader<K,V> |
FileInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
RecordReader<org.apache.hadoop.io.Text,org.apache.hadoop.io.Text> |
SequenceFileAsTextInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
void |
Task.initialize(JobConf job,
JobID id,
Reporter reporter,
boolean useNewApi)
|
void |
Mapper.map(K1 key,
V1 value,
OutputCollector<K2,V2> output,
Reporter reporter)
Maps a single input key/value pair into an intermediate key/value pair. |
void |
Reducer.reduce(K2 key,
Iterator<V2> values,
OutputCollector<K3,V3> output,
Reporter reporter)
Reduces values for a given key. |
void |
MapRunnable.run(RecordReader<K1,V1> input,
OutputCollector<K2,V2> output,
Reporter reporter)
Start mapping input <key, value> pairs. |
void |
MapRunner.run(RecordReader<K1,V1> input,
OutputCollector<K2,V2> output,
Reporter reporter)
|
Constructors in org.apache.hadoop.mapred with parameters of type Reporter | |
---|---|
Task.CombineValuesIterator(RawKeyValueIterator in,
org.apache.hadoop.io.RawComparator<KEY> comparator,
Class<KEY> keyClass,
Class<VALUE> valClass,
org.apache.hadoop.conf.Configuration conf,
Reporter reporter,
Counters.Counter combineInputCounter)
|
Uses of Reporter in org.apache.hadoop.mapred.join |
---|
Methods in org.apache.hadoop.mapred.join with parameters of type Reporter | |
---|---|
ComposableRecordReader<K,V> |
ComposableInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
ComposableRecordReader<K,TupleWritable> |
CompositeInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
Construct a CompositeRecordReader for the children of this InputFormat as defined in the init expression. |
Uses of Reporter in org.apache.hadoop.mapred.lib |
---|
Fields in org.apache.hadoop.mapred.lib declared as Reporter | |
---|---|
protected Reporter |
CombineFileRecordReader.reporter
|
Methods in org.apache.hadoop.mapred.lib with parameters of type Reporter | |
---|---|
void |
FilterOutputFormat.FilterRecordWriter.close(Reporter reporter)
|
OutputCollector |
MultipleOutputs.getCollector(String namedOutput,
Reporter reporter)
Gets the output collector for a named output. |
OutputCollector |
MultipleOutputs.getCollector(String namedOutput,
String multiName,
Reporter reporter)
Gets the output collector for a multi named output. |
RecordReader<org.apache.hadoop.io.LongWritable,org.apache.hadoop.io.Text> |
NLineInputFormat.getRecordReader(InputSplit genericSplit,
JobConf job,
Reporter reporter)
|
abstract RecordReader<K,V> |
CombineFileInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
This is not implemented yet. |
RecordReader<K,V> |
DelegatingInputFormat.getRecordReader(InputSplit split,
JobConf conf,
Reporter reporter)
|
void |
DelegatingMapper.map(K1 key,
V1 value,
OutputCollector<K2,V2> outputCollector,
Reporter reporter)
|
void |
TokenCountMapper.map(K key,
org.apache.hadoop.io.Text value,
OutputCollector<org.apache.hadoop.io.Text,org.apache.hadoop.io.LongWritable> output,
Reporter reporter)
|
void |
RegexMapper.map(K key,
org.apache.hadoop.io.Text value,
OutputCollector<org.apache.hadoop.io.Text,org.apache.hadoop.io.LongWritable> output,
Reporter reporter)
|
void |
IdentityMapper.map(K key,
V val,
OutputCollector<K,V> output,
Reporter reporter)
The identify function. |
void |
FieldSelectionMapReduce.map(K key,
V val,
OutputCollector<org.apache.hadoop.io.Text,org.apache.hadoop.io.Text> output,
Reporter reporter)
The identify function. |
void |
InverseMapper.map(K key,
V value,
OutputCollector<V,K> output,
Reporter reporter)
The inverse function. |
void |
ChainMapper.map(Object key,
Object value,
OutputCollector output,
Reporter reporter)
Chains the map(...) methods of the Mappers in the chain. |
void |
LongSumReducer.reduce(K key,
Iterator<org.apache.hadoop.io.LongWritable> values,
OutputCollector<K,org.apache.hadoop.io.LongWritable> output,
Reporter reporter)
|
void |
IdentityReducer.reduce(K key,
Iterator<V> values,
OutputCollector<K,V> output,
Reporter reporter)
Writes all keys and values directly to output. |
void |
ChainReducer.reduce(Object key,
Iterator values,
OutputCollector output,
Reporter reporter)
Chains the reduce(...) method of the Reducer with the
map(...) methods of the Mappers in the chain. |
void |
FieldSelectionMapReduce.reduce(org.apache.hadoop.io.Text key,
Iterator<org.apache.hadoop.io.Text> values,
OutputCollector<org.apache.hadoop.io.Text,org.apache.hadoop.io.Text> output,
Reporter reporter)
|
void |
MultithreadedMapRunner.run(RecordReader<K1,V1> input,
OutputCollector<K2,V2> output,
Reporter reporter)
|
Constructors in org.apache.hadoop.mapred.lib with parameters of type Reporter | |
---|---|
CombineFileRecordReader(JobConf job,
CombineFileSplit split,
Reporter reporter,
Class<RecordReader<K,V>> rrClass)
A generic RecordReader that can hand out different recordReaders for each chunk in the CombineFileSplit. |
Uses of Reporter in org.apache.hadoop.mapred.lib.aggregate |
---|
Methods in org.apache.hadoop.mapred.lib.aggregate with parameters of type Reporter | |
---|---|
void |
ValueAggregatorReducer.map(K1 arg0,
V1 arg1,
OutputCollector<org.apache.hadoop.io.Text,org.apache.hadoop.io.Text> arg2,
Reporter arg3)
Do nothing. |
void |
ValueAggregatorCombiner.map(K1 arg0,
V1 arg1,
OutputCollector<org.apache.hadoop.io.Text,org.apache.hadoop.io.Text> arg2,
Reporter arg3)
Do nothing. |
void |
ValueAggregatorMapper.map(K1 key,
V1 value,
OutputCollector<org.apache.hadoop.io.Text,org.apache.hadoop.io.Text> output,
Reporter reporter)
the map function. |
void |
ValueAggregatorReducer.reduce(org.apache.hadoop.io.Text key,
Iterator<org.apache.hadoop.io.Text> values,
OutputCollector<org.apache.hadoop.io.Text,org.apache.hadoop.io.Text> output,
Reporter reporter)
|
void |
ValueAggregatorCombiner.reduce(org.apache.hadoop.io.Text key,
Iterator<org.apache.hadoop.io.Text> values,
OutputCollector<org.apache.hadoop.io.Text,org.apache.hadoop.io.Text> output,
Reporter reporter)
Combines values for a given key. |
void |
ValueAggregatorMapper.reduce(org.apache.hadoop.io.Text arg0,
Iterator<org.apache.hadoop.io.Text> arg1,
OutputCollector<org.apache.hadoop.io.Text,org.apache.hadoop.io.Text> arg2,
Reporter arg3)
Do nothing. |
Uses of Reporter in org.apache.hadoop.mapred.lib.db |
---|
Methods in org.apache.hadoop.mapred.lib.db with parameters of type Reporter | |
---|---|
void |
DBOutputFormat.DBRecordWriter.close(Reporter reporter)
Close this RecordWriter to future operations. |
RecordReader<org.apache.hadoop.io.LongWritable,T> |
DBInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
Get the RecordReader for the given InputSplit . |
Uses of Reporter in org.apache.hadoop.mapreduce.task.reduce |
---|
Constructors in org.apache.hadoop.mapreduce.task.reduce with parameters of type Reporter | |
---|---|
MergeManager(TaskAttemptID reduceId,
JobConf jobConf,
org.apache.hadoop.fs.FileSystem localFS,
org.apache.hadoop.fs.LocalDirAllocator localDirAllocator,
Reporter reporter,
org.apache.hadoop.io.compress.CompressionCodec codec,
Class<? extends Reducer> combinerClass,
Task.CombineOutputCollector<K,V> combineCollector,
Counters.Counter spilledRecordsCounter,
Counters.Counter reduceCombineInputCounter,
Counters.Counter mergedMapOutputsCounter,
org.apache.hadoop.mapreduce.task.reduce.ExceptionReporter exceptionReporter,
org.apache.hadoop.util.Progress mergePhase,
MapOutputFile mapOutputFile)
|
|
Shuffle(TaskAttemptID reduceId,
JobConf jobConf,
org.apache.hadoop.fs.FileSystem localFS,
TaskUmbilicalProtocol umbilical,
org.apache.hadoop.fs.LocalDirAllocator localDirAllocator,
Reporter reporter,
org.apache.hadoop.io.compress.CompressionCodec codec,
Class<? extends Reducer> combinerClass,
Task.CombineOutputCollector<K,V> combineCollector,
Counters.Counter spilledRecordsCounter,
Counters.Counter reduceCombineInputCounter,
Counters.Counter shuffledMapsCounter,
Counters.Counter reduceShuffleBytes,
Counters.Counter failedShuffleCounter,
Counters.Counter mergedMapOutputsCounter,
TaskStatus status,
org.apache.hadoop.util.Progress copyPhase,
org.apache.hadoop.util.Progress mergePhase,
Task reduceTask,
MapOutputFile mapOutputFile)
|
|
||||||||||
PREV NEXT | FRAMES NO FRAMES |