Uses of Interface
org.apache.hadoop.mapred.Reporter

Packages that use Reporter
org.apache.hadoop.contrib.utils.join   
org.apache.hadoop.examples Hadoop example code. 
org.apache.hadoop.examples.dancing This package is a distributed implementation of Knuth's dancing links algorithm that can run under Hadoop. 
org.apache.hadoop.hbase.mapred   
org.apache.hadoop.mapred A software framework for easily writing applications which process vast amounts of data (multi-terabyte data-sets) parallelly on large clusters (thousands of nodes) built of commodity hardware in a reliable, fault-tolerant manner. 
org.apache.hadoop.mapred.join Given a set of sorted datasets keyed with the same class and yielding equal partitions, it is possible to effect a join of those datasets prior to the map. 
org.apache.hadoop.mapred.lib Library of generally useful mappers, reducers, and partitioners. 
org.apache.hadoop.mapred.lib.aggregate Classes for performing various counting and aggregations. 
org.apache.hadoop.streaming Hadoop Streaming is a utility which allows users to create and run Map-Reduce jobs with any executables (e.g. 
org.apache.hadoop.tools   
org.apache.hadoop.util Common utilities. 
 

Uses of Reporter in org.apache.hadoop.contrib.utils.join
 

Fields in org.apache.hadoop.contrib.utils.join declared as Reporter
protected  Reporter DataJoinReducerBase.reporter
           
protected  Reporter DataJoinMapperBase.reporter
           
 

Methods in org.apache.hadoop.contrib.utils.join with parameters of type Reporter
protected  void DataJoinReducerBase.collect(WritableComparable key, TaggedMapOutput aRecord, OutputCollector output, Reporter reporter)
          The subclass can overwrite this method to perform additional filtering and/or other processing logic before a value is collected.
 void DataJoinReducerBase.map(WritableComparable arg0, Writable arg1, OutputCollector arg2, Reporter arg3)
           
 void DataJoinMapperBase.map(WritableComparable key, Writable value, OutputCollector output, Reporter reporter)
           
 void DataJoinReducerBase.reduce(WritableComparable key, Iterator values, OutputCollector output, Reporter reporter)
           
 void DataJoinMapperBase.reduce(WritableComparable arg0, Iterator arg1, OutputCollector arg2, Reporter arg3)
           
 

Uses of Reporter in org.apache.hadoop.examples
 

Methods in org.apache.hadoop.examples with parameters of type Reporter
 void SleepJob.map(IntWritable key, IntWritable value, OutputCollector<IntWritable,IntWritable> output, Reporter reporter)
           
 void WordCount.MapClass.map(LongWritable key, Text value, OutputCollector<Text,IntWritable> output, Reporter reporter)
           
 void PiEstimator.PiMapper.map(LongWritable key, Writable val, OutputCollector<LongWritable,LongWritable> out, Reporter reporter)
          Map method.
 void SleepJob.reduce(IntWritable key, Iterator<IntWritable> values, OutputCollector<IntWritable,IntWritable> output, Reporter reporter)
           
 void PiEstimator.PiReducer.reduce(LongWritable key, Iterator<LongWritable> values, OutputCollector<WritableComparable,Writable> output, Reporter reporter)
          Reduce method.
 void WordCount.Reduce.reduce(Text key, Iterator<IntWritable> values, OutputCollector<Text,IntWritable> output, Reporter reporter)
           
 

Uses of Reporter in org.apache.hadoop.examples.dancing
 

Methods in org.apache.hadoop.examples.dancing with parameters of type Reporter
 void DistributedPentomino.PentMap.map(WritableComparable key, Text value, OutputCollector<Text,Text> output, Reporter reporter)
          Break the prefix string into moves (a sequence of integer row ids that will be selected for each column in order).
 

Uses of Reporter in org.apache.hadoop.hbase.mapred
 

Methods in org.apache.hadoop.hbase.mapred with parameters of type Reporter
 void TableOutputFormat.TableRecordWriter.close(Reporter reporter)
          Close this RecordWriter to future operations.
 RecordReader<HStoreKey,MapWritable> TableInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
abstract  void TableMap.map(HStoreKey key, MapWritable value, OutputCollector<K,V> output, Reporter reporter)
          Call a user defined function on a single HBase record, represented by a key and its associated record value.
 void GroupingTableMap.map(HStoreKey key, MapWritable value, OutputCollector<Text,MapWritable> output, Reporter reporter)
          Extract the grouping columns from value to construct a new key.
 void IdentityTableMap.map(HStoreKey key, MapWritable value, OutputCollector<Text,MapWritable> output, Reporter reporter)
          Pass the key, value to reduce
abstract  void TableReduce.reduce(K key, Iterator<V> values, OutputCollector<Text,MapWritable> output, Reporter reporter)
           
 void IndexTableReduce.reduce(Text key, Iterator<MapWritable> values, OutputCollector<Text,org.apache.hadoop.hbase.mapred.LuceneDocumentWrapper> output, Reporter reporter)
           
 void IdentityTableReduce.reduce(Text key, Iterator<MapWritable> values, OutputCollector<Text,MapWritable> output, Reporter reporter)
          No aggregation, output pairs of (key, record)
 

Uses of Reporter in org.apache.hadoop.mapred
 

Fields in org.apache.hadoop.mapred declared as Reporter
static Reporter Reporter.NULL
          A constant of Reporter type that does nothing.
 

Methods in org.apache.hadoop.mapred with parameters of type Reporter
 void TextOutputFormat.LineRecordWriter.close(Reporter reporter)
           
 void RecordWriter.close(Reporter reporter)
          Close this RecordWriter to future operations.
 RecordReader<LongWritable,Text> TextInputFormat.getRecordReader(InputSplit genericSplit, JobConf job, Reporter reporter)
           
 RecordReader<K,V> SequenceFileInputFilter.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
          Create a record reader for the given split
 RecordReader<K,V> SequenceFileInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
 RecordReader<Text,Text> SequenceFileAsTextInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
abstract  RecordReader<K,V> MultiFileInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
 RecordReader<Text,Text> KeyValueTextInputFormat.getRecordReader(InputSplit genericSplit, JobConf job, Reporter reporter)
           
abstract  RecordReader<K,V> FileInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
 RecordReader<K,V> InputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
          Get the RecordReader for the given InputSplit.
 void Mapper.map(K1 key, V1 value, OutputCollector<K2,V2> output, Reporter reporter)
          Maps a single input key/value pair into an intermediate key/value pair.
 void Reducer.reduce(K2 key, Iterator<V2> values, OutputCollector<K3,V3> output, Reporter reporter)
          Reduces values for a given key.
 void MapRunner.run(RecordReader<K1,V1> input, OutputCollector<K2,V2> output, Reporter reporter)
           
 void MapRunnable.run(RecordReader<K1,V1> input, OutputCollector<K2,V2> output, Reporter reporter)
          Start mapping input <key, value> pairs.
 

Uses of Reporter in org.apache.hadoop.mapred.join
 

Methods in org.apache.hadoop.mapred.join with parameters of type Reporter
 org.apache.hadoop.mapred.join.ComposableRecordReader<K,TupleWritable> CompositeInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
          Construct a CompositeRecordReader for the children of this InputFormat as defined in the init expression.
 org.apache.hadoop.mapred.join.ComposableRecordReader<K,V> ComposableInputFormat.getRecordReader(InputSplit split, JobConf job, Reporter reporter)
           
 

Uses of Reporter in org.apache.hadoop.mapred.lib
 

Methods in org.apache.hadoop.mapred.lib with parameters of type Reporter
 void TokenCountMapper.map(K key, Text value, OutputCollector<Text,LongWritable> output, Reporter reporter)
           
 void RegexMapper.map(K key, Text value, OutputCollector<Text,LongWritable> output, Reporter reporter)
           
 void IdentityMapper.map(K key, V val, OutputCollector<K,V> output, Reporter reporter)
          The identify function.
 void FieldSelectionMapReduce.map(K key, V val, OutputCollector<Text,Text> output, Reporter reporter)
          The identify function.
 void InverseMapper.map(K key, V value, OutputCollector<V,K> output, Reporter reporter)
          The inverse function.
 void LongSumReducer.reduce(K key, Iterator<LongWritable> values, OutputCollector<K,LongWritable> output, Reporter reporter)
           
 void IdentityReducer.reduce(K key, Iterator<V> values, OutputCollector<K,V> output, Reporter reporter)
          Writes all keys and values directly to output.
 void FieldSelectionMapReduce.reduce(Text key, Iterator<Text> values, OutputCollector<Text,Text> output, Reporter reporter)
           
 void MultithreadedMapRunner.run(RecordReader<K1,V1> input, OutputCollector<K2,V2> output, Reporter reporter)
           
 

Uses of Reporter in org.apache.hadoop.mapred.lib.aggregate
 

Methods in org.apache.hadoop.mapred.lib.aggregate with parameters of type Reporter
 void ValueAggregatorReducer.map(K1 arg0, V1 arg1, OutputCollector<Text,Text> arg2, Reporter arg3)
          Do nothing.
 void ValueAggregatorMapper.map(K1 key, V1 value, OutputCollector<Text,Text> output, Reporter reporter)
          the map function.
 void ValueAggregatorCombiner.map(K1 arg0, V1 arg1, OutputCollector<Text,Text> arg2, Reporter arg3)
          Do nothing.
 void ValueAggregatorReducer.reduce(Text key, Iterator<Text> values, OutputCollector<Text,Text> output, Reporter reporter)
           
 void ValueAggregatorMapper.reduce(Text arg0, Iterator<Text> arg1, OutputCollector<Text,Text> arg2, Reporter arg3)
          Do nothing.
 void ValueAggregatorCombiner.reduce(Text key, Iterator<Text> values, OutputCollector<Text,Text> output, Reporter reporter)
          Combines values for a given key.
 

Uses of Reporter in org.apache.hadoop.streaming
 

Methods in org.apache.hadoop.streaming with parameters of type Reporter
 RecordReader<Text,Text> StreamInputFormat.getRecordReader(InputSplit genericSplit, JobConf job, Reporter reporter)
           
 void PipeMapper.map(WritableComparable key, Writable value, OutputCollector output, Reporter reporter)
           
 void PipeReducer.reduce(WritableComparable key, Iterator values, OutputCollector output, Reporter reporter)
           
 

Constructors in org.apache.hadoop.streaming with parameters of type Reporter
StreamBaseRecordReader(FSDataInputStream in, FileSplit split, Reporter reporter, JobConf job, FileSystem fs)
           
StreamXmlRecordReader(FSDataInputStream in, FileSplit split, Reporter reporter, JobConf job, FileSystem fs)
           
 

Uses of Reporter in org.apache.hadoop.tools
 

Methods in org.apache.hadoop.tools with parameters of type Reporter
 void Logalyzer.LogRegexMapper.map(K key, Text value, OutputCollector<Text,LongWritable> output, Reporter reporter)
           
 

Uses of Reporter in org.apache.hadoop.util
 

Methods in org.apache.hadoop.util with parameters of type Reporter
 void CopyFiles.FSCopyFilesMapper.map(LongWritable key, org.apache.hadoop.util.CopyFiles.FilePair value, OutputCollector<WritableComparable,Text> out, Reporter reporter)
          Map method.
 



Copyright © 2006 The Apache Software Foundation