Uses of Interface
org.apache.hadoop.mapred.Reporter

Packages that use Reporter
org.apache.hadoop.mapred   
org.apache.hadoop.mapred.join   
org.apache.hadoop.mapred.lib   
org.apache.hadoop.mapred.lib.aggregate   
org.apache.hadoop.mapred.lib.db   
org.apache.hadoop.mapreduce.task.reduce   
 

Uses of Reporter in org.apache.hadoop.mapred
 

Classes in org.apache.hadoop.mapred that implement Reporter
protected  class Task.TaskReporter
           
 

Fields in org.apache.hadoop.mapred declared as Reporter
static Reporter Reporter.NULL
          A constant of Reporter type that does nothing.
 

Methods in org.apache.hadoop.mapred with parameters of type Reporter
 void RecordWriter.close(Reporter reporter)
          Close this RecordWriter to future operations.
 void TextOutputFormat.LineRecordWriter.close(Reporter reporter)
           
 RecordReader<K,V> SequenceFileInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
 RecordReader<org.apache.hadoop.io.LongWritable,org.apache.hadoop.io.Text> TextInputFormat.getRecordReader(InputSplit genericSplit, JobConf job, Reporter reporter)
           
 RecordReader<K,V> SequenceFileInputFilter.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
          Create a record reader for the given split
abstract  RecordReader<K,V> MultiFileInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
 RecordReader<org.apache.hadoop.io.BytesWritable,org.apache.hadoop.io.BytesWritable> SequenceFileAsBinaryInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
 RecordReader<K,V> InputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
          Get the RecordReader for the given InputSplit.
 RecordReader<org.apache.hadoop.io.Text,org.apache.hadoop.io.Text> KeyValueTextInputFormat.getRecordReader(InputSplit genericSplit, JobConf job, Reporter reporter)
           
abstract  RecordReader<K,V> FileInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
 RecordReader<org.apache.hadoop.io.Text,org.apache.hadoop.io.Text> SequenceFileAsTextInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
 void Task.initialize(JobConf job, JobID id, Reporter reporter, boolean useNewApi)
           
 void Mapper.map(K1 key, V1 value, OutputCollector<K2,V2> output, Reporter reporter)
          Maps a single input key/value pair into an intermediate key/value pair.
 void Reducer.reduce(K2 key, Iterator<V2> values, OutputCollector<K3,V3> output, Reporter reporter)
          Reduces values for a given key.
 void MapRunnable.run(RecordReader<K1,V1> input, OutputCollector<K2,V2> output, Reporter reporter)
          Start mapping input <key, value> pairs.
 void MapRunner.run(RecordReader<K1,V1> input, OutputCollector<K2,V2> output, Reporter reporter)
           
 

Constructors in org.apache.hadoop.mapred with parameters of type Reporter
Task.CombineValuesIterator(RawKeyValueIterator in, org.apache.hadoop.io.RawComparator<KEY> comparator, Class<KEY> keyClass, Class<VALUE> valClass, org.apache.hadoop.conf.Configuration conf, Reporter reporter, Counters.Counter combineInputCounter)
           
 

Uses of Reporter in org.apache.hadoop.mapred.join
 

Methods in org.apache.hadoop.mapred.join with parameters of type Reporter
 ComposableRecordReader<K,V> ComposableInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
 ComposableRecordReader<K,TupleWritable> CompositeInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
          Construct a CompositeRecordReader for the children of this InputFormat as defined in the init expression.
 

Uses of Reporter in org.apache.hadoop.mapred.lib
 

Fields in org.apache.hadoop.mapred.lib declared as Reporter
protected  Reporter CombineFileRecordReader.reporter
           
 

Methods in org.apache.hadoop.mapred.lib with parameters of type Reporter
 void FilterOutputFormat.FilterRecordWriter.close(Reporter reporter)
           
 OutputCollector MultipleOutputs.getCollector(String namedOutput, Reporter reporter)
          Gets the output collector for a named output.
 OutputCollector MultipleOutputs.getCollector(String namedOutput, String multiName, Reporter reporter)
          Gets the output collector for a multi named output.
 RecordReader<org.apache.hadoop.io.LongWritable,org.apache.hadoop.io.Text> NLineInputFormat.getRecordReader(InputSplit genericSplit, JobConf job, Reporter reporter)
           
abstract  RecordReader<K,V> CombineFileInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
          This is not implemented yet.
 RecordReader<K,V> DelegatingInputFormat.getRecordReader(InputSplit split, JobConf conf, Reporter reporter)
           
 void DelegatingMapper.map(K1 key, V1 value, OutputCollector<K2,V2> outputCollector, Reporter reporter)
           
 void TokenCountMapper.map(K key, org.apache.hadoop.io.Text value, OutputCollector<org.apache.hadoop.io.Text,org.apache.hadoop.io.LongWritable> output, Reporter reporter)
           
 void RegexMapper.map(K key, org.apache.hadoop.io.Text value, OutputCollector<org.apache.hadoop.io.Text,org.apache.hadoop.io.LongWritable> output, Reporter reporter)
           
 void IdentityMapper.map(K key, V val, OutputCollector<K,V> output, Reporter reporter)
          The identify function.
 void FieldSelectionMapReduce.map(K key, V val, OutputCollector<org.apache.hadoop.io.Text,org.apache.hadoop.io.Text> output, Reporter reporter)
          The identify function.
 void InverseMapper.map(K key, V value, OutputCollector<V,K> output, Reporter reporter)
          The inverse function.
 void ChainMapper.map(Object key, Object value, OutputCollector output, Reporter reporter)
          Chains the map(...) methods of the Mappers in the chain.
 void LongSumReducer.reduce(K key, Iterator<org.apache.hadoop.io.LongWritable> values, OutputCollector<K,org.apache.hadoop.io.LongWritable> output, Reporter reporter)
           
 void IdentityReducer.reduce(K key, Iterator<V> values, OutputCollector<K,V> output, Reporter reporter)
          Writes all keys and values directly to output.
 void ChainReducer.reduce(Object key, Iterator values, OutputCollector output, Reporter reporter)
          Chains the reduce(...) method of the Reducer with the map(...) methods of the Mappers in the chain.
 void FieldSelectionMapReduce.reduce(org.apache.hadoop.io.Text key, Iterator<org.apache.hadoop.io.Text> values, OutputCollector<org.apache.hadoop.io.Text,org.apache.hadoop.io.Text> output, Reporter reporter)
           
 void MultithreadedMapRunner.run(RecordReader<K1,V1> input, OutputCollector<K2,V2> output, Reporter reporter)
           
 

Constructors in org.apache.hadoop.mapred.lib with parameters of type Reporter
CombineFileRecordReader(JobConf job, CombineFileSplit split, Reporter reporter, Class<RecordReader<K,V>> rrClass)
          A generic RecordReader that can hand out different recordReaders for each chunk in the CombineFileSplit.
 

Uses of Reporter in org.apache.hadoop.mapred.lib.aggregate
 

Methods in org.apache.hadoop.mapred.lib.aggregate with parameters of type Reporter
 void ValueAggregatorReducer.map(K1 arg0, V1 arg1, OutputCollector<org.apache.hadoop.io.Text,org.apache.hadoop.io.Text> arg2, Reporter arg3)
          Do nothing.
 void ValueAggregatorCombiner.map(K1 arg0, V1 arg1, OutputCollector<org.apache.hadoop.io.Text,org.apache.hadoop.io.Text> arg2, Reporter arg3)
          Do nothing.
 void ValueAggregatorMapper.map(K1 key, V1 value, OutputCollector<org.apache.hadoop.io.Text,org.apache.hadoop.io.Text> output, Reporter reporter)
          the map function.
 void ValueAggregatorReducer.reduce(org.apache.hadoop.io.Text key, Iterator<org.apache.hadoop.io.Text> values, OutputCollector<org.apache.hadoop.io.Text,org.apache.hadoop.io.Text> output, Reporter reporter)
           
 void ValueAggregatorCombiner.reduce(org.apache.hadoop.io.Text key, Iterator<org.apache.hadoop.io.Text> values, OutputCollector<org.apache.hadoop.io.Text,org.apache.hadoop.io.Text> output, Reporter reporter)
          Combines values for a given key.
 void ValueAggregatorMapper.reduce(org.apache.hadoop.io.Text arg0, Iterator<org.apache.hadoop.io.Text> arg1, OutputCollector<org.apache.hadoop.io.Text,org.apache.hadoop.io.Text> arg2, Reporter arg3)
          Do nothing.
 

Uses of Reporter in org.apache.hadoop.mapred.lib.db
 

Methods in org.apache.hadoop.mapred.lib.db with parameters of type Reporter
 void DBOutputFormat.DBRecordWriter.close(Reporter reporter)
          Close this RecordWriter to future operations.
 RecordReader<org.apache.hadoop.io.LongWritable,T> DBInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
          Get the RecordReader for the given InputSplit.
 

Uses of Reporter in org.apache.hadoop.mapreduce.task.reduce
 

Constructors in org.apache.hadoop.mapreduce.task.reduce with parameters of type Reporter
MergeManager(TaskAttemptID reduceId, JobConf jobConf, org.apache.hadoop.fs.FileSystem localFS, org.apache.hadoop.fs.LocalDirAllocator localDirAllocator, Reporter reporter, org.apache.hadoop.io.compress.CompressionCodec codec, Class<? extends Reducer> combinerClass, Task.CombineOutputCollector<K,V> combineCollector, Counters.Counter spilledRecordsCounter, Counters.Counter reduceCombineInputCounter, Counters.Counter mergedMapOutputsCounter, org.apache.hadoop.mapreduce.task.reduce.ExceptionReporter exceptionReporter, org.apache.hadoop.util.Progress mergePhase, MapOutputFile mapOutputFile)
           
Shuffle(TaskAttemptID reduceId, JobConf jobConf, org.apache.hadoop.fs.FileSystem localFS, TaskUmbilicalProtocol umbilical, org.apache.hadoop.fs.LocalDirAllocator localDirAllocator, Reporter reporter, org.apache.hadoop.io.compress.CompressionCodec codec, Class<? extends Reducer> combinerClass, Task.CombineOutputCollector<K,V> combineCollector, Counters.Counter spilledRecordsCounter, Counters.Counter reduceCombineInputCounter, Counters.Counter shuffledMapsCounter, Counters.Counter reduceShuffleBytes, Counters.Counter failedShuffleCounter, Counters.Counter mergedMapOutputsCounter, TaskStatus status, org.apache.hadoop.util.Progress copyPhase, org.apache.hadoop.util.Progress mergePhase, Task reduceTask, MapOutputFile mapOutputFile)
           
 



Copyright © 2012 Apache Software Foundation. All Rights Reserved.