public static class MultiFileWordCount.MyInputFormat extends org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat<MultiFileWordCount.WordOffset,Text>
CombineFileInputFormat, one should extend it, to return a
(custom) RecordReader. CombineFileInputFormat uses
CombineFileSplits.org.apache.hadoop.mapreduce.lib.input.FileInputFormat.CounterSPLIT_MINSIZE_PERNODE, SPLIT_MINSIZE_PERRACKDEFAULT_LIST_STATUS_NUM_THREADS, INPUT_DIR, INPUT_DIR_RECURSIVE, LIST_STATUS_NUM_THREADS, NUM_INPUT_FILES, PATHFILTER_CLASS, SPLIT_MAXSIZE, SPLIT_MINSIZE| Constructor and Description |
|---|
MyInputFormat() |
| Modifier and Type | Method and Description |
|---|---|
org.apache.hadoop.mapreduce.RecordReader<MultiFileWordCount.WordOffset,Text> |
createRecordReader(org.apache.hadoop.mapreduce.InputSplit split,
org.apache.hadoop.mapreduce.TaskAttemptContext context) |
createPool, createPool, getFileBlockLocations, getSplits, isSplitable, setMaxSplitSize, setMinSplitSizeNode, setMinSplitSizeRackaddInputPath, addInputPathRecursively, addInputPaths, computeSplitSize, getBlockIndex, getFormatMinSplitSize, getInputDirRecursive, getInputPathFilter, getInputPaths, getMaxSplitSize, getMinSplitSize, listStatus, makeSplit, makeSplit, setInputDirRecursive, setInputPathFilter, setInputPaths, setInputPaths, setMaxInputSplitSize, setMinInputSplitSizepublic org.apache.hadoop.mapreduce.RecordReader<MultiFileWordCount.WordOffset,Text> createRecordReader(org.apache.hadoop.mapreduce.InputSplit split, org.apache.hadoop.mapreduce.TaskAttemptContext context) throws IOException
createRecordReader in class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat<MultiFileWordCount.WordOffset,Text>IOExceptionCopyright © 2018 CERN. All Rights Reserved.