org.apache.hadoop.hive.ql.io
Class HiveInputFormat<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable>
java.lang.Object
org.apache.hadoop.hive.ql.io.HiveInputFormat<K,V>
- All Implemented Interfaces:
- org.apache.hadoop.mapred.InputFormat<K,V>, org.apache.hadoop.mapred.JobConfigurable
- Direct Known Subclasses:
- CombineHiveInputFormat
public class HiveInputFormat<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable>
- extends Object
- implements org.apache.hadoop.mapred.InputFormat<K,V>, org.apache.hadoop.mapred.JobConfigurable
HiveInputFormat is a parameterized InputFormat which looks at the path name and determine
the correct InputFormat for that path name from mapredPlan.pathToPartitionInfo().
It can be used to read files with different input format in the same map-reduce job.
Field Summary |
static org.apache.commons.logging.Log |
LOG
|
Method Summary |
void |
configure(org.apache.hadoop.mapred.JobConf job)
|
protected static partitionDesc |
getPartitionDescFromPath(Map<String,partitionDesc> pathToPartitionInfo,
org.apache.hadoop.fs.Path dir)
|
org.apache.hadoop.mapred.RecordReader |
getRecordReader(org.apache.hadoop.mapred.InputSplit split,
org.apache.hadoop.mapred.JobConf job,
org.apache.hadoop.mapred.Reporter reporter)
|
org.apache.hadoop.mapred.InputSplit[] |
getSplits(org.apache.hadoop.mapred.JobConf job,
int numSplits)
|
protected void |
init(org.apache.hadoop.mapred.JobConf job)
|
protected void |
initColumnsNeeded(org.apache.hadoop.mapred.JobConf jobConf,
Class inputFormatClass,
String splitPath,
String splitPathWithNoSchema)
|
void |
validateInput(org.apache.hadoop.mapred.JobConf job)
|
Methods inherited from class java.lang.Object |
clone, equals, finalize, getClass, hashCode, notify, notifyAll, toString, wait, wait, wait |
LOG
public static final org.apache.commons.logging.Log LOG
HiveInputFormat
public HiveInputFormat()
configure
public void configure(org.apache.hadoop.mapred.JobConf job)
- Specified by:
configure
in interface org.apache.hadoop.mapred.JobConfigurable
getRecordReader
public org.apache.hadoop.mapred.RecordReader getRecordReader(org.apache.hadoop.mapred.InputSplit split,
org.apache.hadoop.mapred.JobConf job,
org.apache.hadoop.mapred.Reporter reporter)
throws IOException
- Specified by:
getRecordReader
in interface org.apache.hadoop.mapred.InputFormat<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable>
- Throws:
IOException
init
protected void init(org.apache.hadoop.mapred.JobConf job)
getSplits
public org.apache.hadoop.mapred.InputSplit[] getSplits(org.apache.hadoop.mapred.JobConf job,
int numSplits)
throws IOException
- Specified by:
getSplits
in interface org.apache.hadoop.mapred.InputFormat<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable>
- Throws:
IOException
validateInput
public void validateInput(org.apache.hadoop.mapred.JobConf job)
throws IOException
- Throws:
IOException
getPartitionDescFromPath
protected static partitionDesc getPartitionDescFromPath(Map<String,partitionDesc> pathToPartitionInfo,
org.apache.hadoop.fs.Path dir)
throws IOException
- Throws:
IOException
initColumnsNeeded
protected void initColumnsNeeded(org.apache.hadoop.mapred.JobConf jobConf,
Class inputFormatClass,
String splitPath,
String splitPathWithNoSchema)
Copyright © 2009 The Apache Software Foundation