public class TeraInputFormat extends org.apache.hadoop.mapreduce.lib.input.FileInputFormat<Text,Text>
org.apache.hadoop.mapreduce.lib.input.FileInputFormat.CounterDEFAULT_LIST_STATUS_NUM_THREADS, INPUT_DIR, INPUT_DIR_RECURSIVE, LIST_STATUS_NUM_THREADS, NUM_INPUT_FILES, PATHFILTER_CLASS, SPLIT_MAXSIZE, SPLIT_MINSIZE| Constructor and Description |
|---|
TeraInputFormat() |
| Modifier and Type | Method and Description |
|---|---|
org.apache.hadoop.mapreduce.RecordReader<Text,Text> |
createRecordReader(org.apache.hadoop.mapreduce.InputSplit split,
org.apache.hadoop.mapreduce.TaskAttemptContext context) |
List<org.apache.hadoop.mapreduce.InputSplit> |
getSplits(org.apache.hadoop.mapreduce.JobContext job) |
static void |
writePartitionFile(org.apache.hadoop.mapreduce.JobContext job,
Path partFile)
Use the input splits to take samples of the input and generate sample
keys.
|
addInputPath, addInputPathRecursively, addInputPaths, computeSplitSize, getBlockIndex, getFormatMinSplitSize, getInputDirRecursive, getInputPathFilter, getInputPaths, getMaxSplitSize, getMinSplitSize, isSplitable, listStatus, makeSplit, makeSplit, setInputDirRecursive, setInputPathFilter, setInputPaths, setInputPaths, setMaxInputSplitSize, setMinInputSplitSizepublic static void writePartitionFile(org.apache.hadoop.mapreduce.JobContext job, Path partFile) throws Throwable
job - the job to samplepartFile - where to write the output file toThrowable - if something goes wrongpublic org.apache.hadoop.mapreduce.RecordReader<Text,Text> createRecordReader(org.apache.hadoop.mapreduce.InputSplit split, org.apache.hadoop.mapreduce.TaskAttemptContext context) throws IOException
createRecordReader in class org.apache.hadoop.mapreduce.InputFormat<Text,Text>IOExceptionpublic List<org.apache.hadoop.mapreduce.InputSplit> getSplits(org.apache.hadoop.mapreduce.JobContext job) throws IOException
getSplits in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat<Text,Text>IOExceptionCopyright © 2018 CERN. All Rights Reserved.