|
||||||||||
PREV NEXT | FRAMES NO FRAMES |
Packages that use Reporter | |
---|---|
org.apache.hadoop.contrib.utils.join | |
org.apache.hadoop.examples | Hadoop example code. |
org.apache.hadoop.examples.dancing | This package is a distributed implementation of Knuth's dancing links algorithm that can run under Hadoop. |
org.apache.hadoop.hbase.mapred | |
org.apache.hadoop.mapred | A software framework for easily writing applications which process vast amounts of data (multi-terabyte data-sets) parallelly on large clusters (thousands of nodes) built of commodity hardware in a reliable, fault-tolerant manner. |
org.apache.hadoop.mapred.join | Given a set of sorted datasets keyed with the same class and yielding equal partitions, it is possible to effect a join of those datasets prior to the map. |
org.apache.hadoop.mapred.lib | Library of generally useful mappers, reducers, and partitioners. |
org.apache.hadoop.mapred.lib.aggregate | Classes for performing various counting and aggregations. |
org.apache.hadoop.streaming | Hadoop Streaming is a utility which allows users to create and run Map-Reduce jobs with any executables (e.g. |
org.apache.hadoop.tools | |
org.apache.hadoop.util | Common utilities. |
Uses of Reporter in org.apache.hadoop.contrib.utils.join |
---|
Fields in org.apache.hadoop.contrib.utils.join declared as Reporter | |
---|---|
protected Reporter |
DataJoinReducerBase.reporter
|
protected Reporter |
DataJoinMapperBase.reporter
|
Methods in org.apache.hadoop.contrib.utils.join with parameters of type Reporter | |
---|---|
protected void |
DataJoinReducerBase.collect(WritableComparable key,
TaggedMapOutput aRecord,
OutputCollector output,
Reporter reporter)
The subclass can overwrite this method to perform additional filtering and/or other processing logic before a value is collected. |
void |
DataJoinReducerBase.map(WritableComparable arg0,
Writable arg1,
OutputCollector arg2,
Reporter arg3)
|
void |
DataJoinMapperBase.map(WritableComparable key,
Writable value,
OutputCollector output,
Reporter reporter)
|
void |
DataJoinReducerBase.reduce(WritableComparable key,
Iterator values,
OutputCollector output,
Reporter reporter)
|
void |
DataJoinMapperBase.reduce(WritableComparable arg0,
Iterator arg1,
OutputCollector arg2,
Reporter arg3)
|
Uses of Reporter in org.apache.hadoop.examples |
---|
Methods in org.apache.hadoop.examples with parameters of type Reporter | |
---|---|
void |
SleepJob.map(IntWritable key,
IntWritable value,
OutputCollector<IntWritable,IntWritable> output,
Reporter reporter)
|
void |
WordCount.MapClass.map(LongWritable key,
Text value,
OutputCollector<Text,IntWritable> output,
Reporter reporter)
|
void |
PiEstimator.PiMapper.map(LongWritable key,
Writable val,
OutputCollector<LongWritable,LongWritable> out,
Reporter reporter)
Map method. |
void |
SleepJob.reduce(IntWritable key,
Iterator<IntWritable> values,
OutputCollector<IntWritable,IntWritable> output,
Reporter reporter)
|
void |
PiEstimator.PiReducer.reduce(LongWritable key,
Iterator<LongWritable> values,
OutputCollector<WritableComparable,Writable> output,
Reporter reporter)
Reduce method. |
void |
WordCount.Reduce.reduce(Text key,
Iterator<IntWritable> values,
OutputCollector<Text,IntWritable> output,
Reporter reporter)
|
Uses of Reporter in org.apache.hadoop.examples.dancing |
---|
Methods in org.apache.hadoop.examples.dancing with parameters of type Reporter | |
---|---|
void |
DistributedPentomino.PentMap.map(WritableComparable key,
Text value,
OutputCollector<Text,Text> output,
Reporter reporter)
Break the prefix string into moves (a sequence of integer row ids that will be selected for each column in order). |
Uses of Reporter in org.apache.hadoop.hbase.mapred |
---|
Methods in org.apache.hadoop.hbase.mapred with parameters of type Reporter | |
---|---|
void |
TableOutputFormat.TableRecordWriter.close(Reporter reporter)
Close this RecordWriter to future operations. |
RecordReader<HStoreKey,MapWritable> |
TableInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
abstract void |
TableMap.map(HStoreKey key,
MapWritable value,
OutputCollector<K,V> output,
Reporter reporter)
Call a user defined function on a single HBase record, represented by a key and its associated record value. |
void |
GroupingTableMap.map(HStoreKey key,
MapWritable value,
OutputCollector<Text,MapWritable> output,
Reporter reporter)
Extract the grouping columns from value to construct a new key. |
void |
IdentityTableMap.map(HStoreKey key,
MapWritable value,
OutputCollector<Text,MapWritable> output,
Reporter reporter)
Pass the key, value to reduce |
abstract void |
TableReduce.reduce(K key,
Iterator<V> values,
OutputCollector<Text,MapWritable> output,
Reporter reporter)
|
void |
IndexTableReduce.reduce(Text key,
Iterator<MapWritable> values,
OutputCollector<Text,org.apache.hadoop.hbase.mapred.LuceneDocumentWrapper> output,
Reporter reporter)
|
void |
IdentityTableReduce.reduce(Text key,
Iterator<MapWritable> values,
OutputCollector<Text,MapWritable> output,
Reporter reporter)
No aggregation, output pairs of (key, record) |
Uses of Reporter in org.apache.hadoop.mapred |
---|
Fields in org.apache.hadoop.mapred declared as Reporter | |
---|---|
static Reporter |
Reporter.NULL
A constant of Reporter type that does nothing. |
Methods in org.apache.hadoop.mapred with parameters of type Reporter | |
---|---|
void |
TextOutputFormat.LineRecordWriter.close(Reporter reporter)
|
void |
RecordWriter.close(Reporter reporter)
Close this RecordWriter to future operations. |
RecordReader<LongWritable,Text> |
TextInputFormat.getRecordReader(InputSplit genericSplit,
JobConf job,
Reporter reporter)
|
RecordReader<K,V> |
SequenceFileInputFilter.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
Create a record reader for the given split |
RecordReader<K,V> |
SequenceFileInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
RecordReader<Text,Text> |
SequenceFileAsTextInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
abstract RecordReader<K,V> |
MultiFileInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
RecordReader<Text,Text> |
KeyValueTextInputFormat.getRecordReader(InputSplit genericSplit,
JobConf job,
Reporter reporter)
|
abstract RecordReader<K,V> |
FileInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
RecordReader<K,V> |
InputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
Get the RecordReader for the given InputSplit . |
void |
Mapper.map(K1 key,
V1 value,
OutputCollector<K2,V2> output,
Reporter reporter)
Maps a single input key/value pair into an intermediate key/value pair. |
void |
Reducer.reduce(K2 key,
Iterator<V2> values,
OutputCollector<K3,V3> output,
Reporter reporter)
Reduces values for a given key. |
void |
MapRunner.run(RecordReader<K1,V1> input,
OutputCollector<K2,V2> output,
Reporter reporter)
|
void |
MapRunnable.run(RecordReader<K1,V1> input,
OutputCollector<K2,V2> output,
Reporter reporter)
Start mapping input <key, value> pairs. |
Uses of Reporter in org.apache.hadoop.mapred.join |
---|
Methods in org.apache.hadoop.mapred.join with parameters of type Reporter | |
---|---|
org.apache.hadoop.mapred.join.ComposableRecordReader<K,TupleWritable> |
CompositeInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
Construct a CompositeRecordReader for the children of this InputFormat as defined in the init expression. |
org.apache.hadoop.mapred.join.ComposableRecordReader<K,V> |
ComposableInputFormat.getRecordReader(InputSplit split,
JobConf job,
Reporter reporter)
|
Uses of Reporter in org.apache.hadoop.mapred.lib |
---|
Methods in org.apache.hadoop.mapred.lib with parameters of type Reporter | |
---|---|
void |
TokenCountMapper.map(K key,
Text value,
OutputCollector<Text,LongWritable> output,
Reporter reporter)
|
void |
RegexMapper.map(K key,
Text value,
OutputCollector<Text,LongWritable> output,
Reporter reporter)
|
void |
IdentityMapper.map(K key,
V val,
OutputCollector<K,V> output,
Reporter reporter)
The identify function. |
void |
FieldSelectionMapReduce.map(K key,
V val,
OutputCollector<Text,Text> output,
Reporter reporter)
The identify function. |
void |
InverseMapper.map(K key,
V value,
OutputCollector<V,K> output,
Reporter reporter)
The inverse function. |
void |
LongSumReducer.reduce(K key,
Iterator<LongWritable> values,
OutputCollector<K,LongWritable> output,
Reporter reporter)
|
void |
IdentityReducer.reduce(K key,
Iterator<V> values,
OutputCollector<K,V> output,
Reporter reporter)
Writes all keys and values directly to output. |
void |
FieldSelectionMapReduce.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
|
void |
MultithreadedMapRunner.run(RecordReader<K1,V1> input,
OutputCollector<K2,V2> output,
Reporter reporter)
|
Uses of Reporter in org.apache.hadoop.mapred.lib.aggregate |
---|
Methods in org.apache.hadoop.mapred.lib.aggregate with parameters of type Reporter | |
---|---|
void |
ValueAggregatorReducer.map(K1 arg0,
V1 arg1,
OutputCollector<Text,Text> arg2,
Reporter arg3)
Do nothing. |
void |
ValueAggregatorMapper.map(K1 key,
V1 value,
OutputCollector<Text,Text> output,
Reporter reporter)
the map function. |
void |
ValueAggregatorCombiner.map(K1 arg0,
V1 arg1,
OutputCollector<Text,Text> arg2,
Reporter arg3)
Do nothing. |
void |
ValueAggregatorReducer.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
|
void |
ValueAggregatorMapper.reduce(Text arg0,
Iterator<Text> arg1,
OutputCollector<Text,Text> arg2,
Reporter arg3)
Do nothing. |
void |
ValueAggregatorCombiner.reduce(Text key,
Iterator<Text> values,
OutputCollector<Text,Text> output,
Reporter reporter)
Combines values for a given key. |
Uses of Reporter in org.apache.hadoop.streaming |
---|
Methods in org.apache.hadoop.streaming with parameters of type Reporter | |
---|---|
RecordReader<Text,Text> |
StreamInputFormat.getRecordReader(InputSplit genericSplit,
JobConf job,
Reporter reporter)
|
void |
PipeMapper.map(WritableComparable key,
Writable value,
OutputCollector output,
Reporter reporter)
|
void |
PipeReducer.reduce(WritableComparable key,
Iterator values,
OutputCollector output,
Reporter reporter)
|
Constructors in org.apache.hadoop.streaming with parameters of type Reporter | |
---|---|
StreamBaseRecordReader(FSDataInputStream in,
FileSplit split,
Reporter reporter,
JobConf job,
FileSystem fs)
|
|
StreamXmlRecordReader(FSDataInputStream in,
FileSplit split,
Reporter reporter,
JobConf job,
FileSystem fs)
|
Uses of Reporter in org.apache.hadoop.tools |
---|
Methods in org.apache.hadoop.tools with parameters of type Reporter | |
---|---|
void |
Logalyzer.LogRegexMapper.map(K key,
Text value,
OutputCollector<Text,LongWritable> output,
Reporter reporter)
|
Uses of Reporter in org.apache.hadoop.util |
---|
Methods in org.apache.hadoop.util with parameters of type Reporter | |
---|---|
void |
CopyFiles.FSCopyFilesMapper.map(LongWritable key,
org.apache.hadoop.util.CopyFiles.FilePair value,
OutputCollector<WritableComparable,Text> out,
Reporter reporter)
Map method. |
|
||||||||||
PREV NEXT | FRAMES NO FRAMES |