public class WholeTextFileInputFormat extends org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat<String,String> implements Configurable
CombineFileInputFormat for
reading whole text files. Each file is read as key-value pair, where the key is the file path and
the value is the entire content of file.| Constructor and Description |
|---|
WholeTextFileInputFormat() |
| Modifier and Type | Method and Description |
|---|---|
org.apache.hadoop.mapreduce.RecordReader<String,String> |
createRecordReader(org.apache.hadoop.mapreduce.InputSplit split,
org.apache.hadoop.mapreduce.TaskAttemptContext context) |
void |
setMinPartitions(org.apache.hadoop.mapreduce.JobContext context,
int minPartitions)
Allow minPartitions set by end-user in order to keep compatibility with old Hadoop API,
which is set through setMaxSplitSize
|
addInputPath, addInputPaths, getInputPathFilter, getInputPaths, getMaxSplitSize, getMinSplitSize, setInputPathFilter, setInputPaths, setInputPaths, setMaxInputSplitSize, setMinInputSplitSizeequals, getClass, hashCode, notify, notifyAll, toString, wait, wait, waitconf, getConf, setConfpublic org.apache.hadoop.mapreduce.RecordReader<String,String> createRecordReader(org.apache.hadoop.mapreduce.InputSplit split,
org.apache.hadoop.mapreduce.TaskAttemptContext context)
createRecordReader in class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat<String,String>public void setMinPartitions(org.apache.hadoop.mapreduce.JobContext context,
int minPartitions)