org.apache.hadoop.examples
Class BaileyBorweinPlouffe.BbpInputFormat

java.lang.Object
  extended by org.apache.hadoop.mapreduce.InputFormat<org.apache.hadoop.io.LongWritable,org.apache.hadoop.io.IntWritable>
      extended by org.apache.hadoop.examples.BaileyBorweinPlouffe.BbpInputFormat
Enclosing class:
BaileyBorweinPlouffe

public static class BaileyBorweinPlouffe.BbpInputFormat
extends InputFormat<org.apache.hadoop.io.LongWritable,org.apache.hadoop.io.IntWritable>

Input format for the BaileyBorweinPlouffe.BbpMapper. Keys and values represent offsets and sizes, respectively.


Constructor Summary
BaileyBorweinPlouffe.BbpInputFormat()
           
 
Method Summary
 RecordReader<org.apache.hadoop.io.LongWritable,org.apache.hadoop.io.IntWritable> createRecordReader(InputSplit generic, TaskAttemptContext context)
          Create a record reader for a given split.
 List<InputSplit> getSplits(JobContext context)
          Logically split the set of input files for the job.
 
Methods inherited from class java.lang.Object
clone, equals, finalize, getClass, hashCode, notify, notifyAll, toString, wait, wait, wait
 

Constructor Detail

BaileyBorweinPlouffe.BbpInputFormat

public BaileyBorweinPlouffe.BbpInputFormat()
Method Detail

getSplits

public List<InputSplit> getSplits(JobContext context)
Logically split the set of input files for the job.

Each InputSplit is then assigned to an individual Mapper for processing.

Note: The split is a logical split of the inputs and the input files are not physically split into chunks. For e.g. a split could be <input-file-path, start, offset> tuple. The InputFormat also creates the RecordReader to read the InputSplit.

Specified by:
getSplits in class InputFormat<org.apache.hadoop.io.LongWritable,org.apache.hadoop.io.IntWritable>
Parameters:
context - job configuration.
Returns:
an array of InputSplits for the job.

createRecordReader

public RecordReader<org.apache.hadoop.io.LongWritable,org.apache.hadoop.io.IntWritable> createRecordReader(InputSplit generic,
                                                                                                           TaskAttemptContext context)
Create a record reader for a given split. The framework will call RecordReader.initialize(InputSplit, TaskAttemptContext) before the split is used.

Specified by:
createRecordReader in class InputFormat<org.apache.hadoop.io.LongWritable,org.apache.hadoop.io.IntWritable>
Parameters:
generic - the split to be read
context - the information about the task
Returns:
a new record reader


Copyright © 2009 The Apache Software Foundation