public class WholeTextFileInputFormat
extends org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat<String,String>
implements org.apache.hadoop.conf.Configurable
CombineFileInputFormat
for
reading whole text files. Each file is read as key-value pair, where the key is the file path and
the value is the entire content of file.Constructor and Description |
---|
WholeTextFileInputFormat() |
Modifier and Type | Method and Description |
---|---|
org.apache.hadoop.mapreduce.RecordReader<String,String> |
createRecordReader(org.apache.hadoop.mapreduce.InputSplit split,
org.apache.hadoop.mapreduce.TaskAttemptContext context) |
org.apache.hadoop.conf.Configuration |
getConf() |
void |
setConf(org.apache.hadoop.conf.Configuration c) |
void |
setMinPartitions(org.apache.hadoop.mapreduce.JobContext context,
int minPartitions)
Allow minPartitions set by end-user in order to keep compatibility with old Hadoop API,
which is set through setMaxSplitSize
|
public void setConf(org.apache.hadoop.conf.Configuration c)
setConf
in interface org.apache.hadoop.conf.Configurable
public org.apache.hadoop.conf.Configuration getConf()
getConf
in interface org.apache.hadoop.conf.Configurable
public org.apache.hadoop.mapreduce.RecordReader<String,String> createRecordReader(org.apache.hadoop.mapreduce.InputSplit split, org.apache.hadoop.mapreduce.TaskAttemptContext context)
createRecordReader
in class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat<String,String>
public void setMinPartitions(org.apache.hadoop.mapreduce.JobContext context, int minPartitions)