A B C D E F G H I J K L M N O P Q R S T U V W

A

abort() - Method in class org.apache.hadoop.mapred.gridmix.Statistics
 
abortJob(JobContext, int) - Method in class org.apache.hadoop.mapred.FileOutputCommitter
 
abortJob(JobContext, int) - Method in class org.apache.hadoop.mapred.OutputCommitter
For aborting an unsuccessful job's output.
abortJob(JobContext, JobStatus.State) - Method in class org.apache.hadoop.mapred.OutputCommitter
This method implements the new interface by calling the old method.
abortJob(JobContext, JobStatus.State) - Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
Delete the temporary directory, including all of the work directories.
abortJob(JobContext, JobStatus.State) - Method in class org.apache.hadoop.mapreduce.OutputCommitter
For aborting an unsuccessful job's output.
abortTask(TaskAttemptContext) - Method in class org.apache.hadoop.mapred.FileOutputCommitter
 
abortTask(TaskAttemptContext) - Method in class org.apache.hadoop.mapred.OutputCommitter
Discard the task output
abortTask(TaskAttemptContext) - Method in class org.apache.hadoop.mapred.OutputCommitter
This method implements the new interface by calling the old method.
abortTask(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
Delete the work directory
abortTask(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.OutputCommitter
Discard the task output
accept(CompositeRecordReader.JoinCollector, K) - Method in interface org.apache.hadoop.mapred.join.ComposableRecordReader
While key-value pairs from this RecordReader match the given key, register them with the JoinCollector provided.
accept(CompositeRecordReader.JoinCollector, K) - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
If key provided matches that of this Composite, give JoinCollector iterator over values it may emit.
accept(CompositeRecordReader.JoinCollector, K) - Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
Add an iterator to the collector at the position occupied by this RecordReader over the values in this stream paired with the key provided (ie register a stream of values from this source matching K with a collector).
accept(Path) - Method in class org.apache.hadoop.mapred.OutputLogFilter
Deprecated.  
accept(Object) - Method in class org.apache.hadoop.mapred.SequenceFileInputFilter.MD5Filter
Filtering method If MD5(key) % frequency==0, return true; otherwise return false
accept(Object) - Method in class org.apache.hadoop.mapred.SequenceFileInputFilter.PercentFilter
Filtering method If record# % frequency==0, return true; otherwise return false
accept(Object) - Method in class org.apache.hadoop.mapred.SequenceFileInputFilter.RegexFilter
Filtering method If key matches the regex, return true; otherwise return false
accept(Path) - Method in class org.apache.hadoop.mapred.Utils.OutputFileUtils.OutputFilesFilter
 
accept(Path) - Method in class org.apache.hadoop.mapred.Utils.OutputFileUtils.OutputLogFilter
 
accept(Object) - Method in interface org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.Filter
filter function Decide if a record should be filtered or not
accept(Object) - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.MD5Filter
Filtering method If MD5(key) % frequency==0, return true; otherwise return false
accept(Object) - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.PercentFilter
Filtering method If record# % frequency==0, return true; otherwise return false
accept(Object) - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.RegexFilter
Filtering method If key matches the regex, return true; otherwise return false
accept(CompositeRecordReader.JoinCollector, K) - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
If key provided matches that of this Composite, give JoinCollector iterator over values it may emit.
accept(CompositeRecordReader.JoinCollector, K) - Method in class org.apache.hadoop.mapreduce.lib.join.WrappedRecordReader
Add an iterator to the collector at the position occupied by this RecordReader over the values in this stream paired with the key provided (ie register a stream of values from this source matching K with a collector).
accuracy(long, boolean) - Static method in class org.apache.hadoop.examples.pi.Util
Estimate accuracy.
activateOptions() - Method in class org.apache.hadoop.mapred.TaskLogAppender
 
add(Object) - Method in class org.apache.hadoop.contrib.utils.join.ArrayListBackedIterator
 
add(Object) - Method in interface org.apache.hadoop.contrib.utils.join.ResetableIterator
 
add(Job) - Method in class org.apache.hadoop.mapred.gridmix.Statistics
Used by JobMonitor to add the completed job.
add(InputSplit) - Method in class org.apache.hadoop.mapred.join.CompositeInputSplit
Add an InputSplit to this collection.
add(ComposableRecordReader<K, ? extends V>) - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
Add a RecordReader to this collection.
add(TupleWritable) - Method in class org.apache.hadoop.mapred.join.JoinRecordReader.JoinDelegationIterator
 
add(V) - Method in class org.apache.hadoop.mapred.join.MultiFilterRecordReader.MultiFilterDelegationIterator
 
add(X) - Method in class org.apache.hadoop.mapreduce.lib.join.ArrayListBackedIterator
 
add(InputSplit) - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeInputSplit
Add an InputSplit to this collection.
add(ComposableRecordReader<K, ? extends V>) - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
Add a RecordReader to this collection.
add(int, ResetableIterator<X>) - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader.JoinCollector
Register a given iterator at position id.
add(TupleWritable) - Method in class org.apache.hadoop.mapreduce.lib.join.JoinRecordReader.JoinDelegationIterator
 
add(V) - Method in class org.apache.hadoop.mapreduce.lib.join.MultiFilterRecordReader.MultiFilterDelegationIterator
 
add(T) - Method in interface org.apache.hadoop.mapreduce.lib.join.ResetableIterator
Add an element to the collection of elements to iterate over.
add(U) - Method in class org.apache.hadoop.mapreduce.lib.join.ResetableIterator.EMPTY
 
add(X) - Method in class org.apache.hadoop.mapreduce.lib.join.StreamBackedIterator
 
addArchiveToClassPath(Path) - Method in class org.apache.hadoop.mapreduce.Job
Add an archive path to the current set of classpath entries.
addCacheArchive(URI) - Method in class org.apache.hadoop.mapreduce.Job
Add a archives to be localized
addCacheFile(URI) - Method in class org.apache.hadoop.mapreduce.Job
Add a file to be localized
addClusterStatsObservers(StatListener<Statistics.ClusterStats>) - Method in class org.apache.hadoop.mapred.gridmix.Statistics
 
addColumn(ColumnName, boolean) - Method in class org.apache.hadoop.examples.dancing.DancingLinks
Add a column to the table
addColumn(ColumnName) - Method in class org.apache.hadoop.examples.dancing.DancingLinks
Add a column to the table
addCounter(Counter) - Method in class org.apache.hadoop.mapreduce.CounterGroup
Add a counter to this group.
addDefaults() - Method in class org.apache.hadoop.mapred.join.CompositeInputFormat
Adds the default set of identifiers to the parser.
addDefaults() - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeInputFormat
Adds the default set of identifiers to the parser.
addDelegationToken(String, Token<? extends TokenIdentifier>) - Static method in class org.apache.hadoop.mapreduce.security.TokenCache
auxiliary method to add a delegation token
addDependingJob(ControlledJob) - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
Add a job to this jobs' dependency list.
addDoubleValue(Object, double) - Method in class org.apache.hadoop.contrib.utils.join.JobBase
Increment the given counter by the given incremental value If the counter does not exist, one is created with value 0.
addFileToClassPath(Path) - Method in class org.apache.hadoop.mapreduce.Job
Add an file path to the current set of classpath entries It adds the file to cache as well.
addGroup(CounterGroup) - Method in class org.apache.hadoop.mapreduce.Counters
Add a group.
addIdentifier(String, Class<?>[], Class<? extends Parser.Node>, Class<? extends ComposableRecordReader>) - Static method in class org.apache.hadoop.mapred.join.Parser.Node
For a given identifier, add a mapping to the nodetype for the parse tree and to the ComposableRecordReader to be created, including the formals required to invoke the constructor.
addIdentifier(String, Class<?>[], Class<? extends Parser.Node>, Class<? extends ComposableRecordReader>) - Static method in class org.apache.hadoop.mapreduce.lib.join.Parser.Node
For a given identifier, add a mapping to the nodetype for the parse tree and to the ComposableRecordReader to be created, including the formals required to invoke the constructor.
addInputPath(JobConf, Path) - Static method in class org.apache.hadoop.mapred.FileInputFormat
Add a Path to the list of inputs for the map-reduce job.
addInputPath(JobConf, Path, Class<? extends InputFormat>) - Static method in class org.apache.hadoop.mapred.lib.MultipleInputs
Add a Path with a custom InputFormat to the list of inputs for the map-reduce job.
addInputPath(JobConf, Path, Class<? extends InputFormat>, Class<? extends Mapper>) - Static method in class org.apache.hadoop.mapred.lib.MultipleInputs
Add a Path with a custom InputFormat and Mapper to the list of inputs for the map-reduce job.
addInputPath(Job, Path) - Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
Add a Path to the list of inputs for the map-reduce job.
addInputPath(Job, Path, Class<? extends InputFormat>) - Static method in class org.apache.hadoop.mapreduce.lib.input.MultipleInputs
Add a Path with a custom InputFormat to the list of inputs for the map-reduce job.
addInputPath(Job, Path, Class<? extends InputFormat>, Class<? extends Mapper>) - Static method in class org.apache.hadoop.mapreduce.lib.input.MultipleInputs
Add a Path with a custom InputFormat and Mapper to the list of inputs for the map-reduce job.
addInputPathRecursively(List<FileStatus>, FileSystem, Path, PathFilter) - Method in class org.apache.hadoop.mapred.FileInputFormat
Add files in the input path recursively into the results.
addInputPaths(JobConf, String) - Static method in class org.apache.hadoop.mapred.FileInputFormat
Add the given comma separated paths to the list of inputs for the map-reduce job.
addInputPaths(Job, String) - Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
Add the given comma separated paths to the list of inputs for the map-reduce job.
additionalConfSpec_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
addJob(ControlledJob) - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl
Add a new job.
addJobCollection(Collection<ControlledJob>) - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl
Add a collection of jobs
addJobs(Collection<Job>) - Method in class org.apache.hadoop.mapred.jobcontrol.JobControl
Add a collection of jobs
addJobStatsListeners(StatListener<Statistics.JobStats>) - Method in class org.apache.hadoop.mapred.gridmix.Statistics
 
addLongValue(Object, long) - Method in class org.apache.hadoop.contrib.utils.join.JobBase
Increment the given counter by the given incremental value If the counter does not exist, one is created with value 0.
addMapper(JobConf, Class<? extends Mapper<K1, V1, K2, V2>>, Class<? extends K1>, Class<? extends V1>, Class<? extends K2>, Class<? extends V2>, boolean, JobConf) - Static method in class org.apache.hadoop.mapred.lib.ChainMapper
Adds a Mapper class to the chain job's JobConf.
addMapper(JobConf, Class<? extends Mapper<K1, V1, K2, V2>>, Class<? extends K1>, Class<? extends V1>, Class<? extends K2>, Class<? extends V2>, boolean, JobConf) - Static method in class org.apache.hadoop.mapred.lib.ChainReducer
Adds a Mapper class to the chain job's JobConf.
addMapper(Job, Class<? extends Mapper>, Class<?>, Class<?>, Class<?>, Class<?>, Configuration) - Static method in class org.apache.hadoop.mapreduce.lib.chain.ChainMapper
Adds a Mapper class to the chain mapper.
addMapper(Job, Class<? extends Mapper>, Class<?>, Class<?>, Class<?>, Class<?>, Configuration) - Static method in class org.apache.hadoop.mapreduce.lib.chain.ChainReducer
Adds a Mapper class to the chain reducer.
addMod(double, double) - Static method in class org.apache.hadoop.examples.pi.math.Modular
Given x in [0,1) and a in (-1,1), return (x, a) mod 1.0.
addMultiNamedOutput(JobConf, String, Class<? extends OutputFormat>, Class<?>, Class<?>) - Static method in class org.apache.hadoop.mapred.lib.MultipleOutputs
Adds a multi named output for the job.
addNamedOutput(JobConf, String, Class<? extends OutputFormat>, Class<?>, Class<?>) - Static method in class org.apache.hadoop.mapred.lib.MultipleOutputs
Adds a named output for the job.
addNamedOutput(Job, String, Class<? extends OutputFormat>, Class<?>, Class<?>) - Static method in class org.apache.hadoop.mapreduce.lib.output.MultipleOutputs
Adds a named output for the job.
addNextValue(Object) - Method in class org.apache.hadoop.mapreduce.lib.aggregate.DoubleValueSum
add a value to the aggregator
addNextValue(double) - Method in class org.apache.hadoop.mapreduce.lib.aggregate.DoubleValueSum
add a value to the aggregator
addNextValue(Object) - Method in class org.apache.hadoop.mapreduce.lib.aggregate.LongValueMax
add a value to the aggregator
addNextValue(long) - Method in class org.apache.hadoop.mapreduce.lib.aggregate.LongValueMax
add a value to the aggregator
addNextValue(Object) - Method in class org.apache.hadoop.mapreduce.lib.aggregate.LongValueMin
add a value to the aggregator
addNextValue(long) - Method in class org.apache.hadoop.mapreduce.lib.aggregate.LongValueMin
add a value to the aggregator
addNextValue(Object) - Method in class org.apache.hadoop.mapreduce.lib.aggregate.LongValueSum
add a value to the aggregator
addNextValue(long) - Method in class org.apache.hadoop.mapreduce.lib.aggregate.LongValueSum
add a value to the aggregator
addNextValue(Object) - Method in class org.apache.hadoop.mapreduce.lib.aggregate.StringValueMax
add a value to the aggregator
addNextValue(Object) - Method in class org.apache.hadoop.mapreduce.lib.aggregate.StringValueMin
add a value to the aggregator
addNextValue(Object) - Method in class org.apache.hadoop.mapreduce.lib.aggregate.UniqValueCount
add a value to the aggregator
addNextValue(Object) - Method in interface org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregator
add a value to the aggregator
addNextValue(Object) - Method in class org.apache.hadoop.mapreduce.lib.aggregate.ValueHistogram
add the given val to the aggregator.
addRow(boolean[]) - Method in class org.apache.hadoop.examples.dancing.DancingLinks
Add a row to the table.
addSecretKey(Text, byte[]) - Static method in class org.apache.hadoop.mapreduce.security.TokenCache
auxiliary methods to store user' s secret keys
addTaskEnvironment_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
AggregateWordCount - Class in org.apache.hadoop.examples
This is an example Aggregated Hadoop Map/Reduce application.
AggregateWordCount() - Constructor for class org.apache.hadoop.examples.AggregateWordCount
 
AggregateWordCount.WordCountPlugInClass - Class in org.apache.hadoop.examples
 
AggregateWordCount.WordCountPlugInClass() - Constructor for class org.apache.hadoop.examples.AggregateWordCount.WordCountPlugInClass
 
AggregateWordHistogram - Class in org.apache.hadoop.examples
This is an example Aggregated Hadoop Map/Reduce application.
AggregateWordHistogram() - Constructor for class org.apache.hadoop.examples.AggregateWordHistogram
 
AggregateWordHistogram.AggregateWordHistogramPlugin - Class in org.apache.hadoop.examples
 
AggregateWordHistogram.AggregateWordHistogramPlugin() - Constructor for class org.apache.hadoop.examples.AggregateWordHistogram.AggregateWordHistogramPlugin
 
aggregatorDescriptorList - Variable in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJobBase
 
aggregatorDescriptorList - Static variable in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorJobBase
 
allFinished() - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl
 
append(LoggingEvent) - Method in class org.apache.hadoop.mapred.TaskLogAppender
 
appendTo(StringBuilder) - Method in class org.apache.hadoop.mapreduce.JobID
Add the stuff after the "job" prefix to the given builder.
appendTo(StringBuilder) - Method in class org.apache.hadoop.mapreduce.TaskAttemptID
Add the unique string to the StringBuilder
appendTo(StringBuilder) - Method in class org.apache.hadoop.mapreduce.TaskID
Add the unique string to the given builder.
archiveURIs - Variable in class org.apache.hadoop.streaming.StreamJob
 
argv_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
ArithmeticProgression - Class in org.apache.hadoop.examples.pi.math
An arithmetic progression
ArithmeticProgression(char, long, long, long) - Constructor for class org.apache.hadoop.examples.pi.math.ArithmeticProgression
Constructor
ArrayListBackedIterator - Class in org.apache.hadoop.contrib.utils.join
This class provides an implementation of ResetableIterator.
ArrayListBackedIterator() - Constructor for class org.apache.hadoop.contrib.utils.join.ArrayListBackedIterator
 
ArrayListBackedIterator(ArrayList<Object>) - Constructor for class org.apache.hadoop.contrib.utils.join.ArrayListBackedIterator
 
ArrayListBackedIterator<X extends org.apache.hadoop.io.Writable> - Class in org.apache.hadoop.mapred.join
This class provides an implementation of ResetableIterator.
ArrayListBackedIterator() - Constructor for class org.apache.hadoop.mapred.join.ArrayListBackedIterator
 
ArrayListBackedIterator(ArrayList<X>) - Constructor for class org.apache.hadoop.mapred.join.ArrayListBackedIterator
 
ArrayListBackedIterator<X extends org.apache.hadoop.io.Writable> - Class in org.apache.hadoop.mapreduce.lib.join
This class provides an implementation of ResetableIterator.
ArrayListBackedIterator() - Constructor for class org.apache.hadoop.mapreduce.lib.join.ArrayListBackedIterator
 
ArrayListBackedIterator(ArrayList<X>) - Constructor for class org.apache.hadoop.mapreduce.lib.join.ArrayListBackedIterator
 
ATTEMPT - Static variable in class org.apache.hadoop.mapreduce.TaskAttemptID
 
AutoInputFormat - Class in org.apache.hadoop.streaming
An InputFormat that tries to deduce the types of the input files automatically.
AutoInputFormat() - Constructor for class org.apache.hadoop.streaming.AutoInputFormat
 
available() - Method in class org.apache.hadoop.mapred.TaskLog.Reader
 

B

BaileyBorweinPlouffe - Class in org.apache.hadoop.examples
A map/reduce program that uses Bailey-Borwein-Plouffe to compute exact digits of Pi.
BaileyBorweinPlouffe() - Constructor for class org.apache.hadoop.examples.BaileyBorweinPlouffe
 
BaileyBorweinPlouffe.BbpInputFormat - Class in org.apache.hadoop.examples
Input format for the BaileyBorweinPlouffe.BbpMapper.
BaileyBorweinPlouffe.BbpInputFormat() - Constructor for class org.apache.hadoop.examples.BaileyBorweinPlouffe.BbpInputFormat
 
BaileyBorweinPlouffe.BbpMapper - Class in org.apache.hadoop.examples
Mapper class computing digits of Pi.
BaileyBorweinPlouffe.BbpMapper() - Constructor for class org.apache.hadoop.examples.BaileyBorweinPlouffe.BbpMapper
 
BaileyBorweinPlouffe.BbpReducer - Class in org.apache.hadoop.examples
Reducer for concatenating map outputs.
BaileyBorweinPlouffe.BbpReducer() - Constructor for class org.apache.hadoop.examples.BaileyBorweinPlouffe.BbpReducer
 
BaileyBorweinPlouffe.BbpSplit - Class in org.apache.hadoop.examples
Input split for the BaileyBorweinPlouffe.BbpInputFormat.
BaileyBorweinPlouffe.BbpSplit() - Constructor for class org.apache.hadoop.examples.BaileyBorweinPlouffe.BbpSplit
Public default constructor for the Writable interface.
BASE_OUTPUT_NAME - Static variable in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
 
baseOut - Variable in class org.apache.hadoop.mapred.lib.FilterOutputFormat
 
baseOut - Variable in class org.apache.hadoop.mapreduce.lib.output.FilterOutputFormat
 
Bellard - Class in org.apache.hadoop.examples.pi.math
Bellard's BBP-type Pi formula 1/2^6 \sum_{n=0}^\infty (-1)^n/2^{10n} (-2^5/(4n+1) -1/(4n+3) +2^8/(10n+1) -2^6/(10n+3) -2^2/(10n+5) -2^2/(10n+7) +1/(10n+9)) References: [1] David H.
Bellard() - Constructor for class org.apache.hadoop.examples.pi.math.Bellard
 
Bellard.Parameter - Enum in org.apache.hadoop.examples.pi.math
Parameters for the sums
Bellard.Sum - Class in org.apache.hadoop.examples.pi.math
The sums in the Bellard's formula
BigDecimalSplitter - Class in org.apache.hadoop.mapreduce.lib.db
Implement DBSplitter over BigDecimal values.
BigDecimalSplitter() - Constructor for class org.apache.hadoop.mapreduce.lib.db.BigDecimalSplitter
 
BinaryPartitioner<V> - Class in org.apache.hadoop.mapred.lib
Partition BinaryComparable keys using a configurable part of the bytes array returned by BinaryComparable.getBytes().
BinaryPartitioner() - Constructor for class org.apache.hadoop.mapred.lib.BinaryPartitioner
 
BinaryPartitioner<V> - Class in org.apache.hadoop.mapreduce.lib.partition
Partition BinaryComparable keys using a configurable part of the bytes array returned by BinaryComparable.getBytes().
BinaryPartitioner() - Constructor for class org.apache.hadoop.mapreduce.lib.partition.BinaryPartitioner
 
bit2terms(long) - Static method in class org.apache.hadoop.examples.pi.math.Bellard
Estimate the number of terms.
BooleanSplitter - Class in org.apache.hadoop.mapreduce.lib.db
Implement DBSplitter over boolean values.
BooleanSplitter() - Constructor for class org.apache.hadoop.mapreduce.lib.db.BooleanSplitter
 
BYTES_PER_MAP - Static variable in class org.apache.hadoop.examples.RandomTextWriter
 
BYTES_PER_MAP - Static variable in class org.apache.hadoop.examples.RandomWriter
 
BYTES_READ - Static variable in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
 

C

cacheArchives - Variable in class org.apache.hadoop.streaming.StreamJob
 
cacheFiles - Variable in class org.apache.hadoop.streaming.StreamJob
 
cancelDelegationToken(Token<DelegationTokenIdentifier>) - Method in class org.apache.hadoop.mapred.JobClient
Cancel a delegation token from the JobTracker
cancelDelegationToken(Token<DelegationTokenIdentifier>) - Method in class org.apache.hadoop.mapreduce.Cluster
Cancel a delegation token from the JobTracker
ChainMapper - Class in org.apache.hadoop.mapred.lib
The ChainMapper class allows to use multiple Mapper classes within a single Map task.
ChainMapper() - Constructor for class org.apache.hadoop.mapred.lib.ChainMapper
Constructor.
ChainMapper<KEYIN,VALUEIN,KEYOUT,VALUEOUT> - Class in org.apache.hadoop.mapreduce.lib.chain
The ChainMapper class allows to use multiple Mapper classes within a single Map task.
ChainMapper() - Constructor for class org.apache.hadoop.mapreduce.lib.chain.ChainMapper
 
ChainReducer - Class in org.apache.hadoop.mapred.lib
The ChainReducer class allows to chain multiple Mapper classes after a Reducer within the Reducer task.
ChainReducer() - Constructor for class org.apache.hadoop.mapred.lib.ChainReducer
Constructor.
ChainReducer<KEYIN,VALUEIN,KEYOUT,VALUEOUT> - Class in org.apache.hadoop.mapreduce.lib.chain
The ChainReducer class allows to chain multiple Mapper classes after a Reducer within the Reducer task.
ChainReducer() - Constructor for class org.apache.hadoop.mapreduce.lib.chain.ChainReducer
 
checkDirectory(File) - Static method in class org.apache.hadoop.examples.pi.Util
Check local directory.
checkOutputSpecs(JobContext) - Method in class org.apache.hadoop.examples.terasort.TeraOutputFormat
 
checkOutputSpecs(FileSystem, JobConf) - Method in class org.apache.hadoop.mapred.FileOutputFormat
 
checkOutputSpecs(FileSystem, JobConf) - Method in class org.apache.hadoop.mapred.lib.db.DBOutputFormat
Check for validity of the output-specification for the job.
checkOutputSpecs(FileSystem, JobConf) - Method in class org.apache.hadoop.mapred.lib.FilterOutputFormat
 
checkOutputSpecs(FileSystem, JobConf) - Method in class org.apache.hadoop.mapred.lib.LazyOutputFormat
 
checkOutputSpecs(FileSystem, JobConf) - Method in class org.apache.hadoop.mapred.lib.NullOutputFormat
 
checkOutputSpecs(FileSystem, JobConf) - Method in interface org.apache.hadoop.mapred.OutputFormat
Check for validity of the output-specification for the job.
checkOutputSpecs(FileSystem, JobConf) - Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat
 
checkOutputSpecs(JobContext) - Method in class org.apache.hadoop.mapreduce.lib.db.DBOutputFormat
 
checkOutputSpecs(JobContext) - Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
 
checkOutputSpecs(JobContext) - Method in class org.apache.hadoop.mapreduce.lib.output.FilterOutputFormat
 
checkOutputSpecs(JobContext) - Method in class org.apache.hadoop.mapreduce.lib.output.LazyOutputFormat
 
checkOutputSpecs(JobContext) - Method in class org.apache.hadoop.mapreduce.lib.output.NullOutputFormat
 
checkOutputSpecs(JobContext) - Method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat
 
checkOutputSpecs(JobContext) - Method in class org.apache.hadoop.mapreduce.OutputFormat
Check for validity of the output-specification for the job.
chooseShardForDelete(DocumentID) - Method in class org.apache.hadoop.contrib.index.example.HashingDistributionPolicy
 
chooseShardForDelete(DocumentID) - Method in class org.apache.hadoop.contrib.index.example.RoundRobinDistributionPolicy
 
chooseShardForDelete(DocumentID) - Method in interface org.apache.hadoop.contrib.index.mapred.IDistributionPolicy
Choose a shard or all shards to send a delete request.
chooseShardForInsert(DocumentID) - Method in class org.apache.hadoop.contrib.index.example.HashingDistributionPolicy
 
chooseShardForInsert(DocumentID) - Method in class org.apache.hadoop.contrib.index.example.RoundRobinDistributionPolicy
 
chooseShardForInsert(DocumentID) - Method in interface org.apache.hadoop.contrib.index.mapred.IDistributionPolicy
Choose a shard to send an insert request.
CLASS - Static variable in class org.apache.hadoop.examples.dancing.Pentomino
 
cleanup(Reducer<LongWritable, BytesWritable, LongWritable, BytesWritable>.Context) - Method in class org.apache.hadoop.examples.BaileyBorweinPlouffe.BbpReducer
Write output to files.
cleanup(Reducer<BooleanWritable, LongWritable, WritableComparable<?>, Writable>.Context) - Method in class org.apache.hadoop.examples.QuasiMonteCarlo.QmcReducer
Reduce task done, write output to a file.
cleanup(Mapper<LongWritable, NullWritable, Text, Text>.Context) - Method in class org.apache.hadoop.examples.terasort.TeraGen.SortGenMapper
 
cleanup(Mapper<KEYIN, VALUEIN, KEYOUT, VALUEOUT>.Context) - Method in class org.apache.hadoop.mapreduce.Mapper
Called once at the end of the task.
cleanup(Reducer<KEYIN, VALUEIN, KEYOUT, VALUEOUT>.Context) - Method in class org.apache.hadoop.mapreduce.Reducer
Called once at the end of the task.
cleanupJob(JobContext) - Method in class org.apache.hadoop.mapred.FileOutputCommitter
Deprecated. 
cleanupJob(JobContext) - Method in class org.apache.hadoop.mapred.OutputCommitter
Deprecated. Use OutputCommitter.commitJob(JobContext) or OutputCommitter.abortJob(JobContext, int) instead.
cleanupJob(JobContext) - Method in class org.apache.hadoop.mapred.OutputCommitter
Deprecated. Use OutputCommitter.commitJob(org.apache.hadoop.mapreduce.JobContext) or OutputCommitter.abortJob(org.apache.hadoop.mapreduce.JobContext, org.apache.hadoop.mapreduce.JobStatus.State) instead.
cleanupJob(JobContext) - Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
Deprecated. 
cleanupJob(JobContext) - Method in class org.apache.hadoop.mapreduce.OutputCommitter
Deprecated. Use OutputCommitter.commitJob(JobContext) or OutputCommitter.abortJob(JobContext, JobStatus.State) instead.
cleanupProgress() - Method in class org.apache.hadoop.mapred.JobStatus
 
cleanupProgress() - Method in interface org.apache.hadoop.mapred.RunningJob
Get the progress of the job's cleanup-tasks, as a float between 0.0 and 1.0.
cleanupProgress() - Method in class org.apache.hadoop.mapreduce.Job
Get the progress of the job's cleanup-tasks, as a float between 0.0 and 1.0.
clear() - Method in class org.apache.hadoop.mapred.join.JoinRecordReader.JoinDelegationIterator
 
clear() - Method in class org.apache.hadoop.mapred.join.MultiFilterRecordReader.MultiFilterDelegationIterator
 
clear() - Method in class org.apache.hadoop.mapreduce.lib.join.ArrayListBackedIterator
 
clear() - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader.JoinCollector
Clear all state information.
clear() - Method in class org.apache.hadoop.mapreduce.lib.join.JoinRecordReader.JoinDelegationIterator
 
clear() - Method in class org.apache.hadoop.mapreduce.lib.join.MultiFilterRecordReader.MultiFilterDelegationIterator
 
clear() - Method in interface org.apache.hadoop.mapreduce.lib.join.ResetableIterator
Close datasources, but do not release internal resources.
clear() - Method in class org.apache.hadoop.mapreduce.lib.join.ResetableIterator.EMPTY
 
clear() - Method in class org.apache.hadoop.mapreduce.lib.join.StreamBackedIterator
 
clearMark() - Method in class org.apache.hadoop.mapreduce.MarkableIterator
 
clearMark() - Method in class org.apache.hadoop.mapreduce.task.ReduceContextImpl.ValueIterator
 
CLI - Class in org.apache.hadoop.mapreduce.tools
Interprets the map reduce cli options
CLI() - Constructor for class org.apache.hadoop.mapreduce.tools.CLI
 
CLI(Configuration) - Constructor for class org.apache.hadoop.mapreduce.tools.CLI
 
clone(JobConf) - Method in class org.apache.hadoop.contrib.utils.join.TaggedMapOutput
 
clone() - Method in class org.apache.hadoop.mapreduce.JobStatus
 
close() - Method in class org.apache.hadoop.contrib.index.example.IdentityLocalAnalysis
 
close() - Method in class org.apache.hadoop.contrib.index.example.LineDocLocalAnalysis
 
close() - Method in class org.apache.hadoop.contrib.index.example.LineDocRecordReader
 
close() - Method in class org.apache.hadoop.contrib.index.lucene.FileSystemDirectory
 
close() - Method in class org.apache.hadoop.contrib.index.lucene.ShardWriter
Close the shard writer.
close() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateCombiner
 
close() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateMapper
 
close() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateReducer
 
close() - Method in class org.apache.hadoop.contrib.utils.join.ArrayListBackedIterator
 
close() - Method in class org.apache.hadoop.contrib.utils.join.DataJoinMapperBase
 
close() - Method in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
 
close() - Method in interface org.apache.hadoop.contrib.utils.join.ResetableIterator
 
close() - Method in class org.apache.hadoop.examples.MultiFileWordCount.CombineFileLineRecordReader
 
close() - Method in class org.apache.hadoop.mapred.JobClient
Close the JobClient.
close() - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
Close all child RRs.
close() - Method in class org.apache.hadoop.mapred.join.JoinRecordReader.JoinDelegationIterator
 
close() - Method in class org.apache.hadoop.mapred.join.MultiFilterRecordReader.MultiFilterDelegationIterator
 
close() - Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
Forward close request to proxied RR.
close() - Method in class org.apache.hadoop.mapred.KeyValueLineRecordReader
 
close() - Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorCombiner
Do nothing.
close() - Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJobBase
 
close() - Method in class org.apache.hadoop.mapred.lib.ChainMapper
Closes the ChainMapper and all the Mappers in the chain.
close() - Method in class org.apache.hadoop.mapred.lib.ChainReducer
Closes the ChainReducer, the Reducer and all the Mappers in the chain.
close() - Method in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
 
close(Reporter) - Method in class org.apache.hadoop.mapred.lib.db.DBOutputFormat.DBRecordWriter
Close this RecordWriter to future operations.
close() - Method in class org.apache.hadoop.mapred.lib.FieldSelectionMapReduce
 
close(Reporter) - Method in class org.apache.hadoop.mapred.lib.FilterOutputFormat.FilterRecordWriter
 
close() - Method in class org.apache.hadoop.mapred.lib.MultipleOutputs
Closes all the opened named outputs.
close() - Method in class org.apache.hadoop.mapred.MapReduceBase
Default implementation that does nothing.
close() - Method in interface org.apache.hadoop.mapred.RecordReader
Close this InputSplit to future operations.
close(Reporter) - Method in interface org.apache.hadoop.mapred.RecordWriter
Close this RecordWriter to future operations.
close() - Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
 
close() - Method in class org.apache.hadoop.mapred.SequenceFileAsTextRecordReader
 
close() - Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
 
close() - Method in class org.apache.hadoop.mapred.TaskLog.Reader
 
close() - Method in class org.apache.hadoop.mapred.TaskLogAppender
 
close(Reporter) - Method in class org.apache.hadoop.mapred.TextOutputFormat.LineRecordWriter
 
close() - Method in class org.apache.hadoop.mapreduce.Cluster
Close the Cluster.
close(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.db.DBOutputFormat.DBRecordWriter
Close this RecordWriter to future operations.
close() - Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
Close the record reader.
close() - Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader
 
close() - Method in class org.apache.hadoop.mapreduce.lib.input.KeyValueLineRecordReader
 
close() - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
 
close() - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsTextRecordReader
 
close() - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileRecordReader
 
close() - Method in class org.apache.hadoop.mapreduce.lib.join.ArrayListBackedIterator
 
close() - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
Close all child RRs.
close() - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader.JoinCollector
Close all child iterators.
close() - Method in class org.apache.hadoop.mapreduce.lib.join.JoinRecordReader.JoinDelegationIterator
 
close() - Method in class org.apache.hadoop.mapreduce.lib.join.MultiFilterRecordReader.MultiFilterDelegationIterator
 
close() - Method in interface org.apache.hadoop.mapreduce.lib.join.ResetableIterator
Close datasources and release resources.
close() - Method in class org.apache.hadoop.mapreduce.lib.join.ResetableIterator.EMPTY
 
close() - Method in class org.apache.hadoop.mapreduce.lib.join.StreamBackedIterator
 
close() - Method in class org.apache.hadoop.mapreduce.lib.join.WrappedRecordReader
Forward close request to proxied RR.
close(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.output.FilterOutputFormat.FilterRecordWriter
 
close() - Method in class org.apache.hadoop.mapreduce.lib.output.MultipleOutputs
Closes all the opened outputs.
close(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat.LineRecordWriter
 
close() - Method in class org.apache.hadoop.mapreduce.RecordReader
Close the record reader.
close(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.RecordWriter
Close this RecordWriter to future operations.
close() - Method in class org.apache.hadoop.streaming.PipeMapper
 
close() - Method in class org.apache.hadoop.streaming.PipeReducer
 
close() - Method in class org.apache.hadoop.streaming.StreamBaseRecordReader
Close this to future operations.
closeConnection() - Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat
 
closeWriter() - Method in class org.apache.hadoop.contrib.index.mapred.IntermediateForm
Close the Lucene index writer associated with the intermediate form, if created.
Cluster - Class in org.apache.hadoop.mapreduce
Provides a way to access information about the map/reduce cluster.
Cluster(Configuration) - Constructor for class org.apache.hadoop.mapreduce.Cluster
 
Cluster(InetSocketAddress, Configuration) - Constructor for class org.apache.hadoop.mapreduce.Cluster
 
ClusterMetrics - Class in org.apache.hadoop.mapreduce
Status information on the current state of the Map-Reduce cluster.
ClusterMetrics() - Constructor for class org.apache.hadoop.mapreduce.ClusterMetrics
 
ClusterMetrics(int, int, int, int, int, int, int, int, int, int, int, int) - Constructor for class org.apache.hadoop.mapreduce.ClusterMetrics
 
ClusterStatus - Class in org.apache.hadoop.mapred
Status information on the current state of the Map-Reduce cluster.
ClusterStatus.BlackListInfo - Class in org.apache.hadoop.mapred
Class which encapsulates information about a blacklisted tasktracker.
cmp - Variable in class org.apache.hadoop.mapreduce.lib.join.WrappedRecordReader
 
cmpcl - Variable in class org.apache.hadoop.mapred.join.Parser.Node
 
cmpcl - Variable in class org.apache.hadoop.mapreduce.lib.join.Parser.Node
 
collate(Object[], String) - Static method in class org.apache.hadoop.streaming.StreamUtil
 
collate(List, String) - Static method in class org.apache.hadoop.streaming.StreamUtil
 
collect(Object, TaggedMapOutput, OutputCollector, Reporter) - Method in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
The subclass can overwrite this method to perform additional filtering and/or other processing logic before a value is collected.
collect(K, V) - Method in interface org.apache.hadoop.mapred.OutputCollector
Adds a key/value pair to the output.
collected - Variable in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
 
Combinable<T> - Interface in org.apache.hadoop.examples.pi
A class is Combinable if its object can be combined with other objects.
combine(Object[], Object[]) - Method in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
 
combine(T) - Method in interface org.apache.hadoop.examples.pi.Combinable
Combine this with that.
combine(Summation) - Method in class org.apache.hadoop.examples.pi.math.Summation
Combine this with that.
combine(TaskResult) - Method in class org.apache.hadoop.examples.pi.TaskResult
Combine this with that.
combine(Collection<T>) - Static method in class org.apache.hadoop.examples.pi.Util
Combine a list of items.
combine(Object[], TupleWritable) - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
 
combine(Object[], TupleWritable) - Method in class org.apache.hadoop.mapred.join.InnerJoinRecordReader
Return true iff the tuple is full (all data sources contain this key).
combine(Object[], TupleWritable) - Method in class org.apache.hadoop.mapred.join.MultiFilterRecordReader
Default implementation offers MultiFilterRecordReader.emit(org.apache.hadoop.mapred.join.TupleWritable) every Tuple from the collector (the outer join of child RRs).
combine(Object[], TupleWritable) - Method in class org.apache.hadoop.mapred.join.OuterJoinRecordReader
Emit everything from the collector.
combine(Object[], TupleWritable) - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
 
combine(Object[], TupleWritable) - Method in class org.apache.hadoop.mapreduce.lib.join.InnerJoinRecordReader
Return true iff the tuple is full (all data sources contain this key).
combine(Object[], TupleWritable) - Method in class org.apache.hadoop.mapreduce.lib.join.MultiFilterRecordReader
Default implementation offers MultiFilterRecordReader.emit(org.apache.hadoop.mapreduce.lib.join.TupleWritable) every Tuple from the collector (the outer join of child RRs).
combine(Object[], TupleWritable) - Method in class org.apache.hadoop.mapreduce.lib.join.OuterJoinRecordReader
Emit everything from the collector.
CombineFileInputFormat<K,V> - Class in org.apache.hadoop.mapred.lib
An abstract InputFormat that returns CombineFileSplit's in InputFormat.getSplits(JobConf, int) method.
CombineFileInputFormat() - Constructor for class org.apache.hadoop.mapred.lib.CombineFileInputFormat
default constructor
CombineFileInputFormat<K,V> - Class in org.apache.hadoop.mapreduce.lib.input
An abstract InputFormat that returns CombineFileSplit's in InputFormat.getSplits(JobContext) method.
CombineFileInputFormat() - Constructor for class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat
default constructor
CombineFileRecordReader<K,V> - Class in org.apache.hadoop.mapred.lib
A generic RecordReader that can hand out different recordReaders for each chunk in a CombineFileSplit.
CombineFileRecordReader(JobConf, CombineFileSplit, Reporter, Class<RecordReader<K, V>>) - Constructor for class org.apache.hadoop.mapred.lib.CombineFileRecordReader
A generic RecordReader that can hand out different recordReaders for each chunk in the CombineFileSplit.
CombineFileRecordReader<K,V> - Class in org.apache.hadoop.mapreduce.lib.input
A generic RecordReader that can hand out different recordReaders for each chunk in a CombineFileSplit.
CombineFileRecordReader(CombineFileSplit, TaskAttemptContext, Class<? extends RecordReader<K, V>>) - Constructor for class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader
A generic RecordReader that can hand out different recordReaders for each chunk in the CombineFileSplit.
CombineFileSplit - Class in org.apache.hadoop.mapred.lib
 
CombineFileSplit() - Constructor for class org.apache.hadoop.mapred.lib.CombineFileSplit
 
CombineFileSplit(JobConf, Path[], long[], long[], String[]) - Constructor for class org.apache.hadoop.mapred.lib.CombineFileSplit
 
CombineFileSplit(JobConf, Path[], long[]) - Constructor for class org.apache.hadoop.mapred.lib.CombineFileSplit
 
CombineFileSplit(CombineFileSplit) - Constructor for class org.apache.hadoop.mapred.lib.CombineFileSplit
Copy constructor
CombineFileSplit - Class in org.apache.hadoop.mapreduce.lib.input
A sub-collection of input files.
CombineFileSplit() - Constructor for class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
default constructor
CombineFileSplit(Path[], long[], long[], String[]) - Constructor for class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
 
CombineFileSplit(Path[], long[]) - Constructor for class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
 
CombineFileSplit(CombineFileSplit) - Constructor for class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
Copy constructor
comCmd_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
commitJob(JobContext) - Method in class org.apache.hadoop.examples.terasort.TeraOutputFormat.TeraOutputCommitter
 
commitJob(JobContext) - Method in class org.apache.hadoop.mapred.FileOutputCommitter
 
commitJob(JobContext) - Method in class org.apache.hadoop.mapred.OutputCommitter
For committing job's output after successful job completion.
commitJob(JobContext) - Method in class org.apache.hadoop.mapred.OutputCommitter
This method implements the new interface by calling the old method.
commitJob(JobContext) - Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
Delete the temporary directory, including all of the work directories.
commitJob(JobContext) - Method in class org.apache.hadoop.mapreduce.OutputCommitter
For committing job's output after successful job completion.
commitTask(TaskAttemptContext) - Method in class org.apache.hadoop.mapred.FileOutputCommitter
 
commitTask(TaskAttemptContext) - Method in class org.apache.hadoop.mapred.OutputCommitter
To promote the task's temporary output to final output location The task's output is moved to the job's output directory.
commitTask(TaskAttemptContext) - Method in class org.apache.hadoop.mapred.OutputCommitter
This method implements the new interface by calling the old method.
commitTask(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
Move the files from the work directory to the job output directory
commitTask(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.OutputCommitter
To promote the task's temporary output to final output location The task's output is moved to the job's output directory.
COMPARATOR_OPTIONS - Static variable in class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedComparator
 
compare(byte[], int, int, byte[], int, int) - Method in class org.apache.hadoop.examples.SecondarySort.FirstGroupingComparator
 
compare(SecondarySort.IntPair, SecondarySort.IntPair) - Method in class org.apache.hadoop.examples.SecondarySort.FirstGroupingComparator
 
compare(byte[], int, int, byte[], int, int) - Method in class org.apache.hadoop.examples.SecondarySort.IntPair.Comparator
 
compare(byte[], int, int, byte[], int, int) - Method in class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedComparator
 
compareTo(Object) - Method in class org.apache.hadoop.contrib.index.mapred.DocumentID
 
compareTo(Object) - Method in class org.apache.hadoop.contrib.index.mapred.Shard
 
compareTo(Shard) - Method in class org.apache.hadoop.contrib.index.mapred.Shard
Compare to another shard.
compareTo(Object) - Method in class org.apache.hadoop.examples.MultiFileWordCount.WordOffset
 
compareTo(ArithmeticProgression) - Method in class org.apache.hadoop.examples.pi.math.ArithmeticProgression
compareTo(Summation) - Method in class org.apache.hadoop.examples.pi.math.Summation
compareTo(SummationWritable) - Method in class org.apache.hadoop.examples.pi.SummationWritable
compareTo(TaskResult) - Method in class org.apache.hadoop.examples.pi.TaskResult
compareTo(SecondarySort.IntPair) - Method in class org.apache.hadoop.examples.SecondarySort.IntPair
 
compareTo(ComposableRecordReader<K, ?>) - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
Implement Comparable contract (compare key of join or head of heap with that of another).
compareTo(ComposableRecordReader<K, ?>) - Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
Implement Comparable contract (compare key at head of proxied RR with that of another).
compareTo(ID) - Method in class org.apache.hadoop.mapreduce.ID
Compare IDs by associated numbers
compareTo(ID) - Method in class org.apache.hadoop.mapreduce.JobID
Compare JobIds by first jtIdentifiers, then by job numbers
compareTo(ComposableRecordReader<K, ?>) - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
Implement Comparable contract (compare key of join or head of heap with that of another).
compareTo(ComposableRecordReader<K, ?>) - Method in class org.apache.hadoop.mapreduce.lib.join.WrappedRecordReader
Implement Comparable contract (compare key at head of proxied RR with that of another).
compareTo(ID) - Method in class org.apache.hadoop.mapreduce.TaskAttemptID
Compare TaskIds by first tipIds, then by task numbers.
compareTo(ID) - Method in class org.apache.hadoop.mapreduce.TaskID
Compare TaskInProgressIds by first jobIds, then by tip numbers.
COMPLETION_POLL_INTERVAL_KEY - Static variable in class org.apache.hadoop.mapreduce.Job
Key in mapred-*.xml that sets completionPollInvervalMillis
ComposableInputFormat<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> - Interface in org.apache.hadoop.mapred.join
Refinement of InputFormat requiring implementors to provide ComposableRecordReader instead of RecordReader.
ComposableInputFormat<K extends org.apache.hadoop.io.WritableComparable<?>,V extends org.apache.hadoop.io.Writable> - Class in org.apache.hadoop.mapreduce.lib.join
Refinement of InputFormat requiring implementors to provide ComposableRecordReader instead of RecordReader.
ComposableInputFormat() - Constructor for class org.apache.hadoop.mapreduce.lib.join.ComposableInputFormat
 
ComposableRecordReader<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> - Interface in org.apache.hadoop.mapred.join
Additional operations required of a RecordReader to participate in a join.
ComposableRecordReader<K extends org.apache.hadoop.io.WritableComparable<?>,V extends org.apache.hadoop.io.Writable> - Class in org.apache.hadoop.mapreduce.lib.join
Additional operations required of a RecordReader to participate in a join.
ComposableRecordReader() - Constructor for class org.apache.hadoop.mapreduce.lib.join.ComposableRecordReader
 
compose(Class<? extends InputFormat>, String) - Static method in class org.apache.hadoop.mapred.join.CompositeInputFormat
Convenience method for constructing composite formats.
compose(String, Class<? extends InputFormat>, String...) - Static method in class org.apache.hadoop.mapred.join.CompositeInputFormat
Convenience method for constructing composite formats.
compose(String, Class<? extends InputFormat>, Path...) - Static method in class org.apache.hadoop.mapred.join.CompositeInputFormat
Convenience method for constructing composite formats.
compose(Class<? extends InputFormat>, String) - Static method in class org.apache.hadoop.mapreduce.lib.join.CompositeInputFormat
Convenience method for constructing composite formats.
compose(String, Class<? extends InputFormat>, String...) - Static method in class org.apache.hadoop.mapreduce.lib.join.CompositeInputFormat
Convenience method for constructing composite formats.
compose(String, Class<? extends InputFormat>, Path...) - Static method in class org.apache.hadoop.mapreduce.lib.join.CompositeInputFormat
Convenience method for constructing composite formats.
CompositeInputFormat<K extends org.apache.hadoop.io.WritableComparable> - Class in org.apache.hadoop.mapred.join
An InputFormat capable of performing joins over a set of data sources sorted and partitioned the same way.
CompositeInputFormat() - Constructor for class org.apache.hadoop.mapred.join.CompositeInputFormat
 
CompositeInputFormat<K extends org.apache.hadoop.io.WritableComparable> - Class in org.apache.hadoop.mapreduce.lib.join
An InputFormat capable of performing joins over a set of data sources sorted and partitioned the same way.
CompositeInputFormat() - Constructor for class org.apache.hadoop.mapreduce.lib.join.CompositeInputFormat
 
CompositeInputSplit - Class in org.apache.hadoop.mapred.join
This InputSplit contains a set of child InputSplits.
CompositeInputSplit() - Constructor for class org.apache.hadoop.mapred.join.CompositeInputSplit
 
CompositeInputSplit(int) - Constructor for class org.apache.hadoop.mapred.join.CompositeInputSplit
 
CompositeInputSplit - Class in org.apache.hadoop.mapreduce.lib.join
This InputSplit contains a set of child InputSplits.
CompositeInputSplit() - Constructor for class org.apache.hadoop.mapreduce.lib.join.CompositeInputSplit
 
CompositeInputSplit(int) - Constructor for class org.apache.hadoop.mapreduce.lib.join.CompositeInputSplit
 
CompositeRecordReader<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable,X extends org.apache.hadoop.io.Writable> - Class in org.apache.hadoop.mapred.join
A RecordReader that can effect joins of RecordReaders sharing a common key type and partitioning.
CompositeRecordReader(int, int, Class<? extends WritableComparator>) - Constructor for class org.apache.hadoop.mapred.join.CompositeRecordReader
Create a RecordReader with capacity children to position id in the parent reader.
CompositeRecordReader<K extends org.apache.hadoop.io.WritableComparable<?>,V extends org.apache.hadoop.io.Writable,X extends org.apache.hadoop.io.Writable> - Class in org.apache.hadoop.mapreduce.lib.join
A RecordReader that can effect joins of RecordReaders sharing a common key type and partitioning.
CompositeRecordReader(int, int, Class<? extends WritableComparator>) - Constructor for class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
Create a RecordReader with capacity children to position id in the parent reader.
CompositeRecordReader.JoinCollector - Class in org.apache.hadoop.mapreduce.lib.join
Collector for join values.
CompositeRecordReader.JoinCollector(int) - Constructor for class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader.JoinCollector
Construct a collector capable of handling the specified number of children.
COMPRESS - Static variable in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
 
COMPRESS_CODEC - Static variable in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
 
COMPRESS_TYPE - Static variable in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
 
compute() - Method in class org.apache.hadoop.examples.pi.math.Summation
Compute the value of the summation.
computePi(long, Map<Bellard.Parameter, T>) - Static method in class org.apache.hadoop.examples.pi.math.Bellard
Compute bits of Pi from the results.
computePi(long) - Static method in class org.apache.hadoop.examples.pi.math.Bellard
Compute bits of Pi in the local machine.
computeSplitSize(long, long, long) - Method in class org.apache.hadoop.mapred.FileInputFormat
 
computeSplitSize(long, long, long) - Method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
 
conf - Variable in class org.apache.hadoop.mapred.SequenceFileRecordReader
 
conf - Variable in class org.apache.hadoop.mapreduce.lib.input.SequenceFileRecordReader
 
conf - Variable in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
 
config_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
configure(JobConf) - Method in class org.apache.hadoop.contrib.index.example.IdentityLocalAnalysis
 
configure(JobConf) - Method in class org.apache.hadoop.contrib.index.example.LineDocLocalAnalysis
 
configure(JobConf) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateCombiner
 
configure(JobConf) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateMapper
 
configure(JobConf) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdatePartitioner
 
configure(JobConf) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateReducer
 
configure(IndexUpdateConfiguration) - Method in class org.apache.hadoop.contrib.index.mapred.IntermediateForm
Configure using an index update configuration.
configure(JobConf) - Method in class org.apache.hadoop.contrib.utils.join.DataJoinMapperBase
 
configure(JobConf) - Method in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
 
configure(JobConf) - Method in class org.apache.hadoop.contrib.utils.join.JobBase
Initializes a new instance from a JobConf.
configure(JobConf) - Method in interface org.apache.hadoop.mapred.JobConfigurable
Initializes a new instance from a JobConf.
configure(JobConf) - Method in class org.apache.hadoop.mapred.KeyValueTextInputFormat
 
configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.aggregate.UserDefinedValueAggregatorDescriptor
Do nothing.
configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor
get the input file name.
configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorCombiner
Combiner does not need to configure.
configure(JobConf) - Method in interface org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorDescriptor
Configure the object
configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJobBase
 
configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.BinaryPartitioner
 
configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.ChainMapper
Configures the ChainMapper and all the Mappers in the chain.
configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.ChainReducer
Configures the ChainReducer, the Reducer and all the Mappers in the chain.
configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat
Initializes a new instance from a JobConf.
configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.FieldSelectionMapReduce
 
configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.HashPartitioner
 
configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.KeyFieldBasedComparator
 
configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.KeyFieldBasedPartitioner
 
configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.MultithreadedMapRunner
 
configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.NLineInputFormat
 
configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.RegexMapper
 
configure(JobConf) - Method in class org.apache.hadoop.mapred.lib.TotalOrderPartitioner
 
configure(JobConf) - Method in class org.apache.hadoop.mapred.MapReduceBase
Default implementation that does nothing.
configure(JobConf) - Method in class org.apache.hadoop.mapred.MapRunner
 
configure(JobConf) - Method in class org.apache.hadoop.mapred.TextInputFormat
 
configure(Configuration) - Method in class org.apache.hadoop.mapreduce.lib.aggregate.UserDefinedValueAggregatorDescriptor
Do nothing.
configure(Configuration) - Method in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorBaseDescriptor
get the input file name.
configure(Configuration) - Method in interface org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorDescriptor
Configure the object
configure(JobConf) - Method in class org.apache.hadoop.streaming.AutoInputFormat
 
configure(JobConf) - Method in class org.apache.hadoop.streaming.PipeMapper
 
configure(JobConf) - Method in class org.apache.hadoop.streaming.PipeMapRed
 
configure(JobConf) - Method in class org.apache.hadoop.streaming.PipeReducer
 
configureDB(JobConf, String, String, String, String) - Static method in class org.apache.hadoop.mapred.lib.db.DBConfiguration
Sets the DB access related fields in the JobConf.
configureDB(JobConf, String, String) - Static method in class org.apache.hadoop.mapred.lib.db.DBConfiguration
Sets the DB access related fields in the JobConf.
configureDB(Configuration, String, String, String, String) - Static method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
Sets the DB access related fields in the Configuration.
configureDB(Configuration, String, String) - Static method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
Sets the DB access related fields in the JobConf.
constructQuery(String, String[]) - Method in class org.apache.hadoop.mapreduce.lib.db.DBOutputFormat
Constructs the query used as the prepared statement to insert data.
Container<T> - Interface in org.apache.hadoop.examples.pi
A class is a Container if it contains an element.
contains(Summation) - Method in class org.apache.hadoop.examples.pi.math.Summation
Does this contains that?
context - Variable in class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader
 
ControlledJob - Class in org.apache.hadoop.mapreduce.lib.jobcontrol
This class encapsulates a MapReduce job and its dependency.
ControlledJob(Job, List<ControlledJob>) - Constructor for class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
Construct a job.
ControlledJob(Configuration) - Constructor for class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
Construct a job.
ControlledJob.State - Enum in org.apache.hadoop.mapreduce.lib.jobcontrol
 
countCounters() - Method in class org.apache.hadoop.mapreduce.Counters
Returns the total number of counters, by summing the number of counters in each group.
Counter - Class in org.apache.hadoop.mapreduce
A named counter that tracks the progress of a map/reduce job.
Counter() - Constructor for class org.apache.hadoop.mapreduce.Counter
 
Counter(String, String) - Constructor for class org.apache.hadoop.mapreduce.Counter
 
Counter(String, String, long) - Constructor for class org.apache.hadoop.mapreduce.Counter
Create a counter.
COUNTER_GROUP - Static variable in class org.apache.hadoop.mapred.SkipBadRecords
Special counters which are written by the application and are used by the framework for detecting bad records.
COUNTER_GROUP - Static variable in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
 
COUNTER_MAP_PROCESSED_RECORDS - Static variable in class org.apache.hadoop.mapred.SkipBadRecords
Number of processed map records.
COUNTER_REDUCE_PROCESSED_GROUPS - Static variable in class org.apache.hadoop.mapred.SkipBadRecords
Number of processed reduce groups.
CounterGroup - Class in org.apache.hadoop.mapreduce
A group of Counters that logically belong together.
CounterGroup(String) - Constructor for class org.apache.hadoop.mapreduce.CounterGroup
 
CounterGroup(String, String) - Constructor for class org.apache.hadoop.mapreduce.CounterGroup
Create a CounterGroup.
Counters - Class in org.apache.hadoop.mapred
A set of named counters.
Counters() - Constructor for class org.apache.hadoop.mapred.Counters
 
Counters - Class in org.apache.hadoop.mapreduce
 
Counters() - Constructor for class org.apache.hadoop.mapreduce.Counters
 
Counters(Counters) - Constructor for class org.apache.hadoop.mapreduce.Counters
Utility method to create a Counters object from the org.apache.hadoop.mapred counters
Counters.Counter - Class in org.apache.hadoop.mapred
A counter record, comprising its name and value.
Counters.Group - Class in org.apache.hadoop.mapred
Group of counters, comprising of counters from a particular counter Enum class.
CREATE_DIR - Static variable in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
 
createDataJoinJob(String[]) - Static method in class org.apache.hadoop.contrib.utils.join.DataJoinJob
 
createDBRecordReader(DBInputFormat.DBInputSplit, Configuration) - Method in class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat
 
createDBRecordReader(DBInputFormat.DBInputSplit, Configuration) - Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat
 
createDBRecordReader(DBInputFormat.DBInputSplit, Configuration) - Method in class org.apache.hadoop.mapreduce.lib.db.OracleDataDrivenDBInputFormat
 
createInputStream(String) - Method in class org.apache.hadoop.mapred.gridmix.Gridmix
 
createInstance(String) - Static method in class org.apache.hadoop.mapred.lib.aggregate.UserDefinedValueAggregatorDescriptor
Create an instance of the given class
createInstance(String) - Static method in class org.apache.hadoop.mapreduce.lib.aggregate.UserDefinedValueAggregatorDescriptor
Create an instance of the given class
createInternalValue() - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
Create a value to be used internally for joins.
createJob(String[]) - Static method in class org.apache.hadoop.streaming.StreamJob
This method creates a streaming job from the given argument list.
createJobFactory(JobSubmitter, String, Path, Configuration, CountDownLatch) - Method in class org.apache.hadoop.mapred.gridmix.Gridmix
 
createJobMonitor(Statistics) - Method in class org.apache.hadoop.mapred.gridmix.Gridmix
 
createJobSubmitter(JobMonitor, int, int, FilePool) - Method in class org.apache.hadoop.mapred.gridmix.Gridmix
 
createKey() - Method in class org.apache.hadoop.contrib.index.example.LineDocRecordReader
 
createKey() - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
Create a new key value common to all child RRs.
createKey() - Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
Request new key from proxied RR.
createKey() - Method in class org.apache.hadoop.mapred.KeyValueLineRecordReader
 
createKey() - Method in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
 
createKey() - Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBRecordReader
Create an object of the appropriate type to be used as a key.
createKey() - Method in interface org.apache.hadoop.mapred.RecordReader
Create an object of the appropriate type to be used as a key.
createKey() - Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
 
createKey() - Method in class org.apache.hadoop.mapred.SequenceFileAsTextRecordReader
 
createKey() - Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
 
createKey() - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
Create a new key common to all child RRs.
createKey() - Method in class org.apache.hadoop.mapreduce.lib.join.WrappedRecordReader
Request new key from proxied RR.
createKey() - Method in class org.apache.hadoop.streaming.StreamBaseRecordReader
 
createOutput(String) - Method in class org.apache.hadoop.contrib.index.lucene.FileSystemDirectory
 
createPool(JobConf, List<PathFilter>) - Method in class org.apache.hadoop.mapred.lib.CombineFileInputFormat
Deprecated. Use CombineFileInputFormat.createPool(List).
createPool(JobConf, PathFilter...) - Method in class org.apache.hadoop.mapred.lib.CombineFileInputFormat
Deprecated. Use CombineFileInputFormat.createPool(PathFilter...).
createPool(List<PathFilter>) - Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat
Create a new pool and add the filters to it.
createPool(PathFilter...) - Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat
Create a new pool and add the filters to it.
createReaderThread() - Method in class org.apache.hadoop.mapred.gridmix.SerialJobFactory
 
createReaderThread() - Method in class org.apache.hadoop.mapred.gridmix.StressJobFactory
 
createRecordReader(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.examples.BaileyBorweinPlouffe.BbpInputFormat
Create a record reader for a given split.
createRecordReader(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.examples.MultiFileWordCount.MyInputFormat
 
createRecordReader(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.examples.pi.DistSum.Machine.AbstractInputFormat
Specify how to read the records
createRecordReader(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.examples.terasort.TeraInputFormat
 
createRecordReader(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.mapred.lib.CombineFileInputFormat
 
createRecordReader(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.InputFormat
Create a record reader for a given split.
createRecordReader(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat
Create a record reader for a given split.
createRecordReader(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat
This is not implemented yet.
createRecordReader(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.input.KeyValueTextInputFormat
 
createRecordReader(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.input.NLineInputFormat
 
createRecordReader(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsBinaryInputFormat
 
createRecordReader(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsTextInputFormat
 
createRecordReader(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter
Create a record reader for the given split
createRecordReader(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFormat
 
createRecordReader(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.input.TextInputFormat
 
createRecordReader(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.join.ComposableInputFormat
 
createRecordReader(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeInputFormat
Construct a CompositeRecordReader for the children of this InputFormat as defined in the init expression.
createResetableIterator() - Method in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
The subclass can provide a different implementation on ResetableIterator.
createSymlink() - Method in class org.apache.hadoop.mapreduce.Job
This method allows you to create symlinks in the current working directory of the task to all the cache files/archives
createTupleWritable() - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
Create a value to be used internally for joins.
createValue() - Method in class org.apache.hadoop.contrib.index.example.LineDocRecordReader
 
createValue() - Method in class org.apache.hadoop.mapred.join.JoinRecordReader
Create an object of the appropriate type to be used as a value.
createValue() - Method in class org.apache.hadoop.mapred.join.MultiFilterRecordReader
Create an object of the appropriate type to be used as a value.
createValue() - Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
Request new value from proxied RR.
createValue() - Method in class org.apache.hadoop.mapred.KeyValueLineRecordReader
 
createValue() - Method in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
 
createValue() - Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBRecordReader
Create an object of the appropriate type to be used as a value.
createValue() - Method in interface org.apache.hadoop.mapred.RecordReader
Create an object of the appropriate type to be used as a value.
createValue() - Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
 
createValue() - Method in class org.apache.hadoop.mapred.SequenceFileAsTextRecordReader
 
createValue() - Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
 
createValue() - Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
Deprecated.  
createValue() - Method in class org.apache.hadoop.mapreduce.lib.join.JoinRecordReader
 
createValue() - Method in class org.apache.hadoop.mapreduce.lib.join.OverrideRecordReader
 
createValue() - Method in class org.apache.hadoop.mapreduce.lib.join.WrappedRecordReader
 
createValue() - Method in class org.apache.hadoop.streaming.StreamBaseRecordReader
 
createValueAggregatorJob(String[]) - Static method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJob
Create an Aggregate based map/reduce job.
createValueAggregatorJob(String[], Class<? extends ValueAggregatorDescriptor>[]) - Static method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJob
 
createValueAggregatorJob(Configuration, String[]) - Static method in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorJob
Create an Aggregate based map/reduce job.
createValueAggregatorJob(String[], Class<? extends ValueAggregatorDescriptor>[]) - Static method in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorJob
 
createValueAggregatorJobs(String[], Class<? extends ValueAggregatorDescriptor>[]) - Static method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJob
 
createValueAggregatorJobs(String[]) - Static method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJob
 
createValueAggregatorJobs(String[], Class<? extends ValueAggregatorDescriptor>[]) - Static method in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorJob
 
createValueAggregatorJobs(String[]) - Static method in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorJob
 
createWriter(File, String) - Static method in class org.apache.hadoop.examples.pi.Util
Create a writer of a local file.
curReader - Variable in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
 
curReader - Variable in class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader
 

D

DancingLinks<ColumnName> - Class in org.apache.hadoop.examples.dancing
A generic solver for tile laying problems using Knuth's dancing link algorithm.
DancingLinks() - Constructor for class org.apache.hadoop.examples.dancing.DancingLinks
 
DancingLinks.SolutionAcceptor<ColumnName> - Interface in org.apache.hadoop.examples.dancing
Applications should implement this to receive the solutions to their problems.
DATA_FIELD_SEPERATOR - Static variable in class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionHelper
 
DataDrivenDBInputFormat<T extends DBWritable> - Class in org.apache.hadoop.mapreduce.lib.db
A InputFormat that reads input data from an SQL table.
DataDrivenDBInputFormat() - Constructor for class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat
 
DataDrivenDBInputFormat.DataDrivenDBInputSplit - Class in org.apache.hadoop.mapreduce.lib.db
A InputSplit that spans a set of rows
DataDrivenDBInputFormat.DataDrivenDBInputSplit() - Constructor for class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat.DataDrivenDBInputSplit
Default Constructor
DataDrivenDBInputFormat.DataDrivenDBInputSplit(String, String) - Constructor for class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat.DataDrivenDBInputSplit
Convenience Constructor
DataDrivenDBRecordReader<T extends DBWritable> - Class in org.apache.hadoop.mapreduce.lib.db
A RecordReader that reads records from a SQL table, using data-driven WHERE clause splits.
DataDrivenDBRecordReader(DBInputFormat.DBInputSplit, Class<T>, Configuration, Connection, DBConfiguration, String, String[], String, String) - Constructor for class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBRecordReader
 
DataJoinJob - Class in org.apache.hadoop.contrib.utils.join
This class implements the main function for creating a map/reduce job to join data of different sources.
DataJoinJob() - Constructor for class org.apache.hadoop.contrib.utils.join.DataJoinJob
 
DataJoinMapperBase - Class in org.apache.hadoop.contrib.utils.join
This abstract class serves as the base class for the mapper class of a data join job.
DataJoinMapperBase() - Constructor for class org.apache.hadoop.contrib.utils.join.DataJoinMapperBase
 
DataJoinReducerBase - Class in org.apache.hadoop.contrib.utils.join
This abstract class serves as the base class for the reducer class of a data join job.
DataJoinReducerBase() - Constructor for class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
 
DateSplitter - Class in org.apache.hadoop.mapreduce.lib.db
Implement DBSplitter over date/time values.
DateSplitter() - Constructor for class org.apache.hadoop.mapreduce.lib.db.DateSplitter
 
dateToString(Date) - Method in class org.apache.hadoop.mapreduce.lib.db.DateSplitter
Given a Date 'd', format it as a string for use in a SQL date comparison operation.
dateToString(Date) - Method in class org.apache.hadoop.mapreduce.lib.db.OracleDateSplitter
 
DBConfiguration - Class in org.apache.hadoop.mapred.lib.db
 
DBConfiguration - Class in org.apache.hadoop.mapreduce.lib.db
A container for configuration property names for jobs with DB input/output.
DBConfiguration(Configuration) - Constructor for class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
 
DBCountPageView - Class in org.apache.hadoop.examples
This is a demonstrative program, which uses DBInputFormat for reading the input data from a database, and DBOutputFormat for writing the data to the database.
DBCountPageView() - Constructor for class org.apache.hadoop.examples.DBCountPageView
 
DBInputFormat<T extends DBWritable> - Class in org.apache.hadoop.mapred.lib.db
 
DBInputFormat() - Constructor for class org.apache.hadoop.mapred.lib.db.DBInputFormat
 
DBInputFormat<T extends DBWritable> - Class in org.apache.hadoop.mapreduce.lib.db
A InputFormat that reads input data from an SQL table.
DBInputFormat() - Constructor for class org.apache.hadoop.mapreduce.lib.db.DBInputFormat
 
DBInputFormat.DBInputSplit - Class in org.apache.hadoop.mapred.lib.db
A InputSplit that spans a set of rows
DBInputFormat.DBInputSplit() - Constructor for class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBInputSplit
Default Constructor
DBInputFormat.DBInputSplit(long, long) - Constructor for class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBInputSplit
Convenience Constructor
DBInputFormat.DBInputSplit - Class in org.apache.hadoop.mapreduce.lib.db
A InputSplit that spans a set of rows
DBInputFormat.DBInputSplit() - Constructor for class org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit
Default Constructor
DBInputFormat.DBInputSplit(long, long) - Constructor for class org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit
Convenience Constructor
DBInputFormat.DBRecordReader - Class in org.apache.hadoop.mapred.lib.db
A RecordReader that reads records from a SQL table.
DBInputFormat.DBRecordReader(DBInputFormat.DBInputSplit, Class<T>, JobConf, Connection, DBConfiguration, String, String[], String) - Constructor for class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBRecordReader
 
DBInputFormat.NullDBWritable - Class in org.apache.hadoop.mapred.lib.db
A Class that does nothing, implementing DBWritable
DBInputFormat.NullDBWritable() - Constructor for class org.apache.hadoop.mapred.lib.db.DBInputFormat.NullDBWritable
 
DBInputFormat.NullDBWritable - Class in org.apache.hadoop.mapreduce.lib.db
A Class that does nothing, implementing DBWritable
DBInputFormat.NullDBWritable() - Constructor for class org.apache.hadoop.mapreduce.lib.db.DBInputFormat.NullDBWritable
 
DBOutputFormat<K extends DBWritable,V> - Class in org.apache.hadoop.mapred.lib.db
 
DBOutputFormat() - Constructor for class org.apache.hadoop.mapred.lib.db.DBOutputFormat
 
DBOutputFormat<K extends DBWritable,V> - Class in org.apache.hadoop.mapreduce.lib.db
A OutputFormat that sends the reduce output to a SQL table.
DBOutputFormat() - Constructor for class org.apache.hadoop.mapreduce.lib.db.DBOutputFormat
 
DBOutputFormat.DBRecordWriter - Class in org.apache.hadoop.mapred.lib.db
A RecordWriter that writes the reduce output to a SQL table
DBOutputFormat.DBRecordWriter(Connection, PreparedStatement) - Constructor for class org.apache.hadoop.mapred.lib.db.DBOutputFormat.DBRecordWriter
 
DBOutputFormat.DBRecordWriter - Class in org.apache.hadoop.mapreduce.lib.db
A RecordWriter that writes the reduce output to a SQL table
DBOutputFormat.DBRecordWriter() - Constructor for class org.apache.hadoop.mapreduce.lib.db.DBOutputFormat.DBRecordWriter
 
DBOutputFormat.DBRecordWriter(Connection, PreparedStatement) - Constructor for class org.apache.hadoop.mapreduce.lib.db.DBOutputFormat.DBRecordWriter
 
DBRecordReader<T extends DBWritable> - Class in org.apache.hadoop.mapreduce.lib.db
A RecordReader that reads records from a SQL table.
DBRecordReader(DBInputFormat.DBInputSplit, Class<T>, Configuration, Connection, DBConfiguration, String, String[], String) - Constructor for class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
 
DBSplitter - Interface in org.apache.hadoop.mapreduce.lib.db
DBSplitter will generate DBInputSplits to use with DataDrivenDBInputFormat.
DBWritable - Interface in org.apache.hadoop.mapred.lib.db
 
DBWritable - Interface in org.apache.hadoop.mapreduce.lib.db
Objects that are read from/written to a database should implement DBWritable.
debug_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
DEFAULT_LOG_LEVEL - Static variable in class org.apache.hadoop.mapred.JobConf
Default logging level for map/reduce tasks.
DEFAULT_MAPRED_TASK_JAVA_OPTS - Static variable in class org.apache.hadoop.mapred.JobConf
 
DEFAULT_PATH - Static variable in class org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner
 
DEFAULT_QUEUE_NAME - Static variable in class org.apache.hadoop.mapred.JobConf
Name of the queue to which jobs will be submitted, if no queue name is mentioned.
DELETE - Static variable in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp.Op
 
deleteFile(String) - Method in class org.apache.hadoop.contrib.index.lucene.FileSystemDirectory
 
deleteLocalFiles() - Method in class org.apache.hadoop.mapred.JobConf
Deprecated. 
deleteLocalFiles(String) - Method in class org.apache.hadoop.mapred.JobConf
 
deleteTermIterator() - Method in class org.apache.hadoop.contrib.index.mapred.IntermediateForm
Get an iterator for the delete terms in the intermediate form.
delta - Variable in class org.apache.hadoop.examples.pi.math.ArithmeticProgression
Difference between terms
DEPENDENT_FAILED - Static variable in class org.apache.hadoop.mapred.jobcontrol.Job
 
DEPTH - Static variable in class org.apache.hadoop.examples.dancing.Pentomino
 
DESCRIPTION - Static variable in class org.apache.hadoop.examples.BaileyBorweinPlouffe
 
DESCRIPTION - Static variable in class org.apache.hadoop.examples.pi.DistBbp
 
DESCRIPTOR - Static variable in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorJobBase
 
DESCRIPTOR_NUM - Static variable in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorJobBase
 
detailedUsage_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
dfmt(double) - Static method in class org.apache.hadoop.streaming.StreamUtil
 
DIR_FORMATS - Static variable in class org.apache.hadoop.mapreduce.lib.input.MultipleInputs
 
DIR_MAPPERS - Static variable in class org.apache.hadoop.mapreduce.lib.input.MultipleInputs
 
DISABLED_MEMORY_LIMIT - Static variable in class org.apache.hadoop.mapred.JobConf
A value which if set for memory related configuration options, indicates that the options are turned off.
displayJobList(Job[]) - Method in class org.apache.hadoop.mapreduce.tools.CLI
 
displayTasks(JobID, String, String) - Method in class org.apache.hadoop.mapred.JobClient
Display the information about a job's tasks, of a particular type and in a particular state
displayTasks(Job, String, String) - Method in class org.apache.hadoop.mapreduce.tools.CLI
Display the information about a job's tasks, of a particular type and in a particular state
DistBbp - Class in org.apache.hadoop.examples.pi
A map/reduce program that uses a BBP-type method to compute exact binary digits of Pi.
DistBbp() - Constructor for class org.apache.hadoop.examples.pi.DistBbp
 
DistributedCache - Class in org.apache.hadoop.filecache
Deprecated. Use methods on Job.
DistributedCache() - Constructor for class org.apache.hadoop.filecache.DistributedCache
Deprecated.  
DistributedPentomino - Class in org.apache.hadoop.examples.dancing
Launch a distributed pentomino solver.
DistributedPentomino() - Constructor for class org.apache.hadoop.examples.dancing.DistributedPentomino
 
DistributedPentomino.PentMap - Class in org.apache.hadoop.examples.dancing
Each map takes a line, which represents a prefix move and finds all of the solutions that start with that prefix.
DistributedPentomino.PentMap() - Constructor for class org.apache.hadoop.examples.dancing.DistributedPentomino.PentMap
 
DistSum - Class in org.apache.hadoop.examples.pi
The main class for computing sums using map/reduce jobs.
DistSum() - Constructor for class org.apache.hadoop.examples.pi.DistSum
 
DistSum.Machine - Class in org.apache.hadoop.examples.pi
Abstract machine for job execution.
DistSum.Machine() - Constructor for class org.apache.hadoop.examples.pi.DistSum.Machine
 
DistSum.Machine.AbstractInputFormat - Class in org.apache.hadoop.examples.pi
An abstract InputFormat for the jobs
DistSum.Machine.AbstractInputFormat() - Constructor for class org.apache.hadoop.examples.pi.DistSum.Machine.AbstractInputFormat
 
DistSum.Machine.SummationSplit - Class in org.apache.hadoop.examples.pi
Split for the summations
DistSum.Machine.SummationSplit() - Constructor for class org.apache.hadoop.examples.pi.DistSum.Machine.SummationSplit
 
DistSum.MapSide - Class in org.apache.hadoop.examples.pi
A machine which does computation on the map side.
DistSum.MapSide() - Constructor for class org.apache.hadoop.examples.pi.DistSum.MapSide
 
DistSum.MapSide.PartitionInputFormat - Class in org.apache.hadoop.examples.pi
An InputFormat which partitions a summation
DistSum.MapSide.PartitionInputFormat() - Constructor for class org.apache.hadoop.examples.pi.DistSum.MapSide.PartitionInputFormat
 
DistSum.MapSide.SummingMapper - Class in org.apache.hadoop.examples.pi
A mapper which computes sums
DistSum.MapSide.SummingMapper() - Constructor for class org.apache.hadoop.examples.pi.DistSum.MapSide.SummingMapper
 
DistSum.MixMachine - Class in org.apache.hadoop.examples.pi
A machine which chooses Machine in runtime according to the cluster status
DistSum.MixMachine() - Constructor for class org.apache.hadoop.examples.pi.DistSum.MixMachine
 
DistSum.ReduceSide - Class in org.apache.hadoop.examples.pi
A machine which does computation on the reduce side.
DistSum.ReduceSide() - Constructor for class org.apache.hadoop.examples.pi.DistSum.ReduceSide
 
DistSum.ReduceSide.IndexPartitioner - Class in org.apache.hadoop.examples.pi
Use the index for partitioning.
DistSum.ReduceSide.IndexPartitioner() - Constructor for class org.apache.hadoop.examples.pi.DistSum.ReduceSide.IndexPartitioner
 
DistSum.ReduceSide.PartitionMapper - Class in org.apache.hadoop.examples.pi
A Mapper which partitions a summation
DistSum.ReduceSide.PartitionMapper() - Constructor for class org.apache.hadoop.examples.pi.DistSum.ReduceSide.PartitionMapper
 
DistSum.ReduceSide.SummationInputFormat - Class in org.apache.hadoop.examples.pi
An InputFormat which returns a single summation.
DistSum.ReduceSide.SummationInputFormat() - Constructor for class org.apache.hadoop.examples.pi.DistSum.ReduceSide.SummationInputFormat
 
DistSum.ReduceSide.SummingReducer - Class in org.apache.hadoop.examples.pi
A Reducer which computes sums
DistSum.ReduceSide.SummingReducer() - Constructor for class org.apache.hadoop.examples.pi.DistSum.ReduceSide.SummingReducer
 
DocumentAndOp - Class in org.apache.hadoop.contrib.index.mapred
This class represents an indexing operation.
DocumentAndOp() - Constructor for class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
Constructor for no operation.
DocumentAndOp(DocumentAndOp.Op, Document) - Constructor for class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
Constructor for an insert operation.
DocumentAndOp(DocumentAndOp.Op, Term) - Constructor for class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
Constructor for a delete operation.
DocumentAndOp(DocumentAndOp.Op, Document, Term) - Constructor for class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
Constructor for an insert, a delete or an update operation.
DocumentAndOp.Op - Class in org.apache.hadoop.contrib.index.mapred
This class represents the type of an operation - an insert, a delete or an update.
DocumentID - Class in org.apache.hadoop.contrib.index.mapred
The class represents a document id, which is of type text.
DocumentID() - Constructor for class org.apache.hadoop.contrib.index.mapred.DocumentID
Constructor.
DOUBLE_VALUE_SUM - Static variable in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor
 
DOUBLE_VALUE_SUM - Static variable in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorBaseDescriptor
 
DoubleValueSum - Class in org.apache.hadoop.mapred.lib.aggregate
This class implements a value aggregator that sums up a sequence of double values.
DoubleValueSum() - Constructor for class org.apache.hadoop.mapred.lib.aggregate.DoubleValueSum
 
DoubleValueSum - Class in org.apache.hadoop.mapreduce.lib.aggregate
This class implements a value aggregator that sums up a sequence of double values.
DoubleValueSum() - Constructor for class org.apache.hadoop.mapreduce.lib.aggregate.DoubleValueSum
The default constructor
downgrade(JobID) - Static method in class org.apache.hadoop.mapred.JobID
Downgrade a new JobID to an old one
downgrade(JobStatus) - Static method in class org.apache.hadoop.mapred.JobStatus
 
downgrade(TaskAttemptID) - Static method in class org.apache.hadoop.mapred.TaskAttemptID
Downgrade a new TaskAttemptID to an old one
downgrade(TaskID) - Static method in class org.apache.hadoop.mapred.TaskID
Downgrade a new TaskID to an old one
DRIVER_CLASS_PROPERTY - Static variable in class org.apache.hadoop.mapred.lib.db.DBConfiguration
The JDBC Driver class name
DRIVER_CLASS_PROPERTY - Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
The JDBC Driver class name
DumpTypedBytes - Class in org.apache.hadoop.streaming
Utility program that fetches all files that match a given pattern and dumps their content to stdout as typed bytes.
DumpTypedBytes(Configuration) - Constructor for class org.apache.hadoop.streaming.DumpTypedBytes
 
DumpTypedBytes() - Constructor for class org.apache.hadoop.streaming.DumpTypedBytes
 

E

E - Variable in class org.apache.hadoop.examples.pi.math.Summation
Variable e in the summation.
emit(TupleWritable) - Method in class org.apache.hadoop.mapred.join.MultiFilterRecordReader
For each tuple emitted, return a value (typically one of the values in the tuple).
emit(TupleWritable) - Method in class org.apache.hadoop.mapred.join.OverrideRecordReader
Emit the value with the highest position in the tuple.
emit(TupleWritable) - Method in class org.apache.hadoop.mapreduce.lib.join.MultiFilterRecordReader
For each tuple emitted, return a value (typically one of the values in the tuple).
emit(TupleWritable) - Method in class org.apache.hadoop.mapreduce.lib.join.OverrideRecordReader
Emit the value with the highest position in the tuple.
empty - Variable in class org.apache.hadoop.mapreduce.lib.join.WrappedRecordReader
 
EMPTY_ARRAY - Static variable in class org.apache.hadoop.mapred.TaskCompletionEvent
 
EMPTY_ARRAY - Static variable in class org.apache.hadoop.mapreduce.TaskCompletionEvent
 
emptyText - Static variable in class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionHelper
 
endMap(String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
 
endMap(TreeMap, String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
 
endRecord(String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
 
endRecord(Record, String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
 
endVector(String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
 
endVector(ArrayList, String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
 
env_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
Environment - Class in org.apache.hadoop.streaming
This is a class used to get the current environment on the host machines running the map/reduce.
Environment() - Constructor for class org.apache.hadoop.streaming.Environment
 
equals(Object) - Method in class org.apache.hadoop.contrib.index.mapred.Shard
 
equals(Object) - Method in class org.apache.hadoop.examples.MultiFileWordCount.WordOffset
 
equals(Object) - Method in class org.apache.hadoop.examples.pi.math.ArithmeticProgression
equals(Object) - Method in class org.apache.hadoop.examples.pi.math.Summation
equals(Object) - Method in class org.apache.hadoop.examples.pi.SummationWritable
equals(Object) - Method in class org.apache.hadoop.examples.pi.TaskResult
equals(Object) - Method in class org.apache.hadoop.examples.SecondarySort.IntPair
 
equals(Object) - Method in class org.apache.hadoop.mapred.Counters
 
equals(Object) - Method in class org.apache.hadoop.mapred.Counters.Group
Checks for (content) equality of Groups
equals(Object) - Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
Return true iff compareTo(other) retn true.
equals(Object) - Method in class org.apache.hadoop.mapreduce.Counter
 
equals(Object) - Method in class org.apache.hadoop.mapreduce.CounterGroup
 
equals(Object) - Method in class org.apache.hadoop.mapreduce.Counters
 
equals(Object) - Method in class org.apache.hadoop.mapreduce.ID
 
equals(Object) - Method in class org.apache.hadoop.mapreduce.JobID
 
equals(Object) - Method in class org.apache.hadoop.mapreduce.lib.join.TupleWritable
equals(Object) - Method in class org.apache.hadoop.mapreduce.lib.join.WrappedRecordReader
Return true iff compareTo(other) retn true.
equals(Object) - Method in class org.apache.hadoop.mapreduce.TaskAttemptID
 
equals(Object) - Method in class org.apache.hadoop.mapreduce.TaskCompletionEvent
 
equals(Object) - Method in class org.apache.hadoop.mapreduce.TaskID
 
err - Static variable in class org.apache.hadoop.examples.pi.Util
Error stream
estimatePi(int, long, Configuration) - Static method in class org.apache.hadoop.examples.QuasiMonteCarlo
Run a map/reduce job for estimating Pi.
ExampleDriver - Class in org.apache.hadoop.examples
A description of an example program based on its class and a human-readable description.
ExampleDriver() - Constructor for class org.apache.hadoop.examples.ExampleDriver
 
EXECUTABLE - Static variable in class org.apache.hadoop.mapred.pipes.Submitter
 
execute(int, List<E>) - Static method in class org.apache.hadoop.examples.pi.Util
Execute the callables by a number of threads
executeQuery(String) - Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
 
executeQuery(String) - Method in class org.apache.hadoop.mapreduce.lib.db.MySQLDataDrivenDBRecordReader
 
executeQuery(String) - Method in class org.apache.hadoop.mapreduce.lib.db.MySQLDBRecordReader
 
exitUsage(boolean) - Method in class org.apache.hadoop.streaming.StreamJob
 
extractOutputKeyValue(String, String, String, List<Integer>, List<Integer>, int, boolean, boolean) - Method in class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionHelper
 

F

fail(String) - Method in class org.apache.hadoop.streaming.StreamJob
 
FAILED - Static variable in class org.apache.hadoop.mapred.jobcontrol.Job
 
FAILED - Static variable in class org.apache.hadoop.mapred.JobStatus
 
failTask(TaskAttemptID) - Method in class org.apache.hadoop.mapreduce.Job
Fail indicated task attempt.
FieldSelectionHelper - Class in org.apache.hadoop.mapreduce.lib.fieldsel
This class implements a mapper/reducer class that can be used to perform field selections in a manner similar to unix cut.
FieldSelectionHelper() - Constructor for class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionHelper
 
FieldSelectionHelper(Text, Text) - Constructor for class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionHelper
 
FieldSelectionMapper<K,V> - Class in org.apache.hadoop.mapreduce.lib.fieldsel
This class implements a mapper class that can be used to perform field selections in a manner similar to unix cut.
FieldSelectionMapper() - Constructor for class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionMapper
 
FieldSelectionMapReduce<K,V> - Class in org.apache.hadoop.mapred.lib
This class implements a mapper/reducer class that can be used to perform field selections in a manner similar to unix cut.
FieldSelectionMapReduce() - Constructor for class org.apache.hadoop.mapred.lib.FieldSelectionMapReduce
 
FieldSelectionReducer<K,V> - Class in org.apache.hadoop.mapreduce.lib.fieldsel
This class implements a reducer class that can be used to perform field selections in a manner similar to unix cut.
FieldSelectionReducer() - Constructor for class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionReducer
 
FileAlreadyExistsException - Exception in org.apache.hadoop.mapred
Used when target file already exists for any operation and is not configured to be overwritten.
FileAlreadyExistsException() - Constructor for exception org.apache.hadoop.mapred.FileAlreadyExistsException
 
FileAlreadyExistsException(String) - Constructor for exception org.apache.hadoop.mapred.FileAlreadyExistsException
 
fileExists(String) - Method in class org.apache.hadoop.contrib.index.lucene.FileSystemDirectory
 
fileExtension(String) - Method in class org.apache.hadoop.streaming.JarBuilder
 
FileInputFormat<K,V> - Class in org.apache.hadoop.mapred
A base class for file-based InputFormat.
FileInputFormat() - Constructor for class org.apache.hadoop.mapred.FileInputFormat
 
FileInputFormat<K,V> - Class in org.apache.hadoop.mapreduce.lib.input
A base class for file-based InputFormats.
FileInputFormat() - Constructor for class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
 
fileLength(String) - Method in class org.apache.hadoop.contrib.index.lucene.FileSystemDirectory
 
fileModified(String) - Method in class org.apache.hadoop.contrib.index.lucene.FileSystemDirectory
 
FileOutputCommitter - Class in org.apache.hadoop.mapred
An OutputCommitter that commits files specified in job output directory i.e.
FileOutputCommitter() - Constructor for class org.apache.hadoop.mapred.FileOutputCommitter
 
FileOutputCommitter - Class in org.apache.hadoop.mapreduce.lib.output
An OutputCommitter that commits files specified in job output directory i.e.
FileOutputCommitter(Path, TaskAttemptContext) - Constructor for class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
Create a file output committer
FileOutputFormat<K,V> - Class in org.apache.hadoop.mapred
A base class for OutputFormat.
FileOutputFormat() - Constructor for class org.apache.hadoop.mapred.FileOutputFormat
 
FileOutputFormat<K,V> - Class in org.apache.hadoop.mapreduce.lib.output
A base class for OutputFormats that read from FileSystems.
FileOutputFormat() - Constructor for class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
 
FileSplit - Class in org.apache.hadoop.mapred
A section of an input file.
FileSplit() - Constructor for class org.apache.hadoop.mapred.FileSplit
 
FileSplit(Path, long, long, JobConf) - Constructor for class org.apache.hadoop.mapred.FileSplit
Deprecated.  
FileSplit(Path, long, long, String[]) - Constructor for class org.apache.hadoop.mapred.FileSplit
Constructs a split with host information
FileSplit(FileSplit) - Constructor for class org.apache.hadoop.mapred.FileSplit
 
FileSplit - Class in org.apache.hadoop.mapreduce.lib.input
A section of an input file.
FileSplit() - Constructor for class org.apache.hadoop.mapreduce.lib.input.FileSplit
 
FileSplit(Path, long, long, String[]) - Constructor for class org.apache.hadoop.mapreduce.lib.input.FileSplit
Constructs a split with host information
FileSystemDirectory - Class in org.apache.hadoop.contrib.index.lucene
This class implements a Lucene Directory on top of a general FileSystem.
FileSystemDirectory(FileSystem, Path, boolean, Configuration) - Constructor for class org.apache.hadoop.contrib.index.lucene.FileSystemDirectory
Constructor
fileURIs - Variable in class org.apache.hadoop.streaming.StreamJob
 
fillJoinCollector(K) - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
For all child RRs offering the key provided, obtain an iterator at that position in the JoinCollector.
fillJoinCollector(K) - Method in class org.apache.hadoop.mapred.join.OverrideRecordReader
Instead of filling the JoinCollector with iterators from all data sources, fill only the rightmost for this key.
fillJoinCollector(K) - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
For all child RRs offering the key provided, obtain an iterator at that position in the JoinCollector.
fillJoinCollector(K) - Method in class org.apache.hadoop.mapreduce.lib.join.OverrideRecordReader
Instead of filling the JoinCollector with iterators from all data sources, fill only the rightmost for this key.
FILTER_CLASS - Static variable in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter
 
FILTER_FREQUENCY - Static variable in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter
 
FILTER_REGEX - Static variable in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter
 
FilterOutputFormat<K,V> - Class in org.apache.hadoop.mapred.lib
FilterOutputFormat is a convenience class that wraps OutputFormat.
FilterOutputFormat() - Constructor for class org.apache.hadoop.mapred.lib.FilterOutputFormat
 
FilterOutputFormat(OutputFormat<K, V>) - Constructor for class org.apache.hadoop.mapred.lib.FilterOutputFormat
Create a FilterOutputFormat based on the supplied output format.
FilterOutputFormat<K,V> - Class in org.apache.hadoop.mapreduce.lib.output
FilterOutputFormat is a convenience class that wraps OutputFormat.
FilterOutputFormat() - Constructor for class org.apache.hadoop.mapreduce.lib.output.FilterOutputFormat
 
FilterOutputFormat(OutputFormat<K, V>) - Constructor for class org.apache.hadoop.mapreduce.lib.output.FilterOutputFormat
Create a FilterOutputFormat based on the underlying output format.
FilterOutputFormat.FilterRecordWriter<K,V> - Class in org.apache.hadoop.mapred.lib
FilterRecordWriter is a convenience wrapper class that implements RecordWriter.
FilterOutputFormat.FilterRecordWriter() - Constructor for class org.apache.hadoop.mapred.lib.FilterOutputFormat.FilterRecordWriter
 
FilterOutputFormat.FilterRecordWriter(RecordWriter<K, V>) - Constructor for class org.apache.hadoop.mapred.lib.FilterOutputFormat.FilterRecordWriter
 
FilterOutputFormat.FilterRecordWriter<K,V> - Class in org.apache.hadoop.mapreduce.lib.output
FilterRecordWriter is a convenience wrapper class that extends the RecordWriter.
FilterOutputFormat.FilterRecordWriter() - Constructor for class org.apache.hadoop.mapreduce.lib.output.FilterOutputFormat.FilterRecordWriter
 
FilterOutputFormat.FilterRecordWriter(RecordWriter<K, V>) - Constructor for class org.apache.hadoop.mapreduce.lib.output.FilterOutputFormat.FilterRecordWriter
 
findByte(byte[], int, int, byte) - Static method in class org.apache.hadoop.streaming.UTF8ByteArrayUtils
Deprecated. use UTF8ByteArrayUtils.findByte(byte[], int, int, byte)
findBytes(byte[], int, int, byte[]) - Static method in class org.apache.hadoop.streaming.UTF8ByteArrayUtils
Deprecated. use UTF8ByteArrayUtils.findBytes(byte[], int, int, byte[])
findCounter(Enum) - Method in class org.apache.hadoop.mapred.Counters
Find the counter for the given enum.
findCounter(String, String) - Method in class org.apache.hadoop.mapred.Counters
Find a counter given the group and the name.
findCounter(String, int, String) - Method in class org.apache.hadoop.mapred.Counters
Deprecated.  
findCounter(String, String) - Method in class org.apache.hadoop.mapreduce.CounterGroup
Find a counter in a group.
findCounter(String) - Method in class org.apache.hadoop.mapreduce.CounterGroup
 
findCounter(String, String) - Method in class org.apache.hadoop.mapreduce.Counters
 
findCounter(Enum<?>) - Method in class org.apache.hadoop.mapreduce.Counters
Find the counter for the given enum.
findInClasspath(String) - Static method in class org.apache.hadoop.streaming.StreamUtil
 
findInClasspath(String, ClassLoader) - Static method in class org.apache.hadoop.streaming.StreamUtil
 
findNthByte(byte[], int, int, byte, int) - Static method in class org.apache.hadoop.streaming.UTF8ByteArrayUtils
Deprecated. use UTF8ByteArrayUtils.findNthByte(byte[], int, int, byte, int)
findNthByte(byte[], byte, int) - Static method in class org.apache.hadoop.streaming.UTF8ByteArrayUtils
Deprecated. use UTF8ByteArrayUtils.findNthByte(byte[], byte, int)
findSeparator(byte[], int, int, byte) - Static method in class org.apache.hadoop.mapred.KeyValueLineRecordReader
 
findSeparator(byte[], int, int, byte) - Static method in class org.apache.hadoop.mapreduce.lib.input.KeyValueLineRecordReader
 
findTab(byte[], int, int) - Static method in class org.apache.hadoop.streaming.StreamKeyValUtil
Find the first occured tab in a UTF-8 encoded string
findTab(byte[]) - Static method in class org.apache.hadoop.streaming.StreamKeyValUtil
Find the first occured tab in a UTF-8 encoded string
findTab(byte[], int, int) - Static method in class org.apache.hadoop.streaming.UTF8ByteArrayUtils
Deprecated. use StreamKeyValUtil.findTab(byte[], int, int)
findTab(byte[]) - Static method in class org.apache.hadoop.streaming.UTF8ByteArrayUtils
Deprecated. use StreamKeyValUtil.findTab(byte[])
FloatSplitter - Class in org.apache.hadoop.mapreduce.lib.db
Implement DBSplitter over floating-point values.
FloatSplitter() - Constructor for class org.apache.hadoop.mapreduce.lib.db.FloatSplitter
 
flush() - Method in class org.apache.hadoop.mapred.TaskLogAppender
 
flush(TupleWritable) - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader.JoinCollector
Write the next value into key, value as accepted by the operation associated with this set of RecordReaders.
formatBytes(long) - Static method in class org.apache.hadoop.streaming.StreamUtil
 
formatBytes2(long) - Static method in class org.apache.hadoop.streaming.StreamUtil
 
forName(String) - Static method in class org.apache.hadoop.mapred.JobID
Construct a JobId object from given string
forName(String) - Static method in class org.apache.hadoop.mapred.TaskAttemptID
Construct a TaskAttemptID object from given string
forName(String) - Static method in class org.apache.hadoop.mapred.TaskID
 
forName(String) - Static method in class org.apache.hadoop.mapreduce.JobID
Construct a JobId object from given string
forName(String) - Static method in class org.apache.hadoop.mapreduce.TaskAttemptID
Construct a TaskAttemptID object from given string
forName(String) - Static method in class org.apache.hadoop.mapreduce.TaskID
Construct a TaskID object from given string
fourRotations - Static variable in class org.apache.hadoop.examples.dancing.Pentomino
Are all 4 rotations unique?
fromEscapedCompactString(String) - Static method in class org.apache.hadoop.mapred.Counters
Convert a stringified counter representation into a counter object.
fs - Variable in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
 
fs - Variable in class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader
 

G

generateActualKey(K, V) - Method in class org.apache.hadoop.mapred.lib.MultipleOutputFormat
Generate the actual key from the given key/value.
generateActualValue(K, V) - Method in class org.apache.hadoop.mapred.lib.MultipleOutputFormat
Generate the actual value from the given key and value.
generateEntry(String, String, Text) - Static method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor
 
generateEntry(String, String, Text) - Static method in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorBaseDescriptor
 
generateFileNameForKeyValue(K, V, String) - Method in class org.apache.hadoop.mapred.lib.MultipleOutputFormat
Generate the file output file name based on the given key and the leaf file name.
generateGroupKey(TaggedMapOutput) - Method in class org.apache.hadoop.contrib.utils.join.DataJoinMapperBase
Generate a map output key.
generateInputTag(String) - Method in class org.apache.hadoop.contrib.utils.join.DataJoinMapperBase
Determine the source tag based on the input file name.
generateKeyValPairs(Object, Object) - Method in class org.apache.hadoop.examples.AggregateWordCount.WordCountPlugInClass
 
generateKeyValPairs(Object, Object) - Method in class org.apache.hadoop.examples.AggregateWordHistogram.AggregateWordHistogramPlugin
Parse the given value, generate an aggregation-id/value pair per word.
generateKeyValPairs(Object, Object) - Method in class org.apache.hadoop.mapreduce.lib.aggregate.UserDefinedValueAggregatorDescriptor
Generate a list of aggregation-id/value pairs for the given key/value pairs by delegating the invocation to the real object.
generateKeyValPairs(Object, Object) - Method in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorBaseDescriptor
Generate 1 or 2 aggregation-id/value pairs for the given key/value pair.
generateKeyValPairs(Object, Object) - Method in interface org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorDescriptor
Generate a list of aggregation-id/value pairs for the given key/value pair.
generateLeafFileName(String) - Method in class org.apache.hadoop.mapred.lib.MultipleOutputFormat
Generate the leaf name for the output file name.
generateTaggedMapOutput(Object) - Method in class org.apache.hadoop.contrib.utils.join.DataJoinMapperBase
Generate a tagged map output value.
generateValueAggregator(String) - Static method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor
 
generateValueAggregator(String, long) - Static method in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorBaseDescriptor
 
generationFromSegmentsFileName(String) - Static method in class org.apache.hadoop.contrib.index.lucene.LuceneUtil
Parse the generation off the segments file name and return it.
GenSort - Class in org.apache.hadoop.examples.terasort
A single process data generator for the terasort data.
GenSort() - Constructor for class org.apache.hadoop.examples.terasort.GenSort
 
get(String) - Static method in enum org.apache.hadoop.examples.pi.math.Bellard.Parameter
Get the Parameter represented by the String
get(int) - Method in class org.apache.hadoop.mapred.join.CompositeInputSplit
Get ith child InputSplit.
get(int) - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeInputSplit
Get ith child InputSplit.
get(int) - Method in class org.apache.hadoop.mapreduce.lib.join.TupleWritable
Get ith Writable from Tuple.
get(DataInput) - Static method in class org.apache.hadoop.typedbytes.TypedBytesInput
Get a thread-local typed bytes input for the supplied DataInput.
get(DataOutput) - Static method in class org.apache.hadoop.typedbytes.TypedBytesOutput
Get a thread-local typed bytes output for the supplied DataOutput.
get(TypedBytesInput) - Static method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
Get a thread-local typed bytes record input for the supplied TypedBytesInput.
get(DataInput) - Static method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
Get a thread-local typed bytes record input for the supplied DataInput.
get(TypedBytesOutput) - Static method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
Get a thread-local typed bytes record input for the supplied TypedBytesOutput.
get(DataOutput) - Static method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
Get a thread-local typed bytes record output for the supplied DataOutput.
get(TypedBytesInput) - Static method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
Get a thread-local typed bytes writable input for the supplied TypedBytesInput.
get(DataInput) - Static method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
Get a thread-local typed bytes writable input for the supplied DataInput.
get(TypedBytesOutput) - Static method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
Get a thread-local typed bytes writable input for the supplied TypedBytesOutput.
get(DataOutput) - Static method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
Get a thread-local typed bytes writable output for the supplied DataOutput.
getAbsolutePath(String) - Method in class org.apache.hadoop.streaming.PathFinder
Returns the full path name of this file if it is listed in the path
getActiveTaskTrackers() - Method in class org.apache.hadoop.mapreduce.Cluster
Get all active trackers in the cluster.
getActiveTrackerNames() - Method in class org.apache.hadoop.mapred.ClusterStatus
Get the names of task trackers in the cluster.
getAggregatorDescriptors(Configuration) - Static method in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorJobBase
 
getAllJobs() - Method in class org.apache.hadoop.mapred.JobClient
Get the jobs that are submitted.
getAllJobs() - Method in class org.apache.hadoop.mapreduce.Cluster
Get all the jobs in cluster.
getAllTaskAttempts() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskInfo
Get the map of all attempts for this task
getAllTasks() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.JobInfo
Get the map of all tasks in this job
getAllTaskTypes() - Static method in class org.apache.hadoop.mapreduce.TaskID
 
getAllTokens() - Static method in class org.apache.hadoop.mapreduce.security.TokenCache
auxiliary method
getArchiveClassPaths() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get the archive entries in classpath as an array of Path
getArchiveClassPaths() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getArchiveClassPaths() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getArchiveTimestamps() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get the timestamps of the archives.
getArchiveTimestamps() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getArchiveTimestamps() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getAssignedJobID() - Method in class org.apache.hadoop.mapred.jobcontrol.Job
 
getAttemptId() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskAttemptInfo
Get the attempt Id
getAttemptsToStartSkipping(Configuration) - Static method in class org.apache.hadoop.mapred.SkipBadRecords
Get the number of Task attempts AFTER which skip mode will be kicked off.
getAutoIncrMapperProcCount(Configuration) - Static method in class org.apache.hadoop.mapred.SkipBadRecords
Get the flag which if set to true, SkipBadRecords.COUNTER_MAP_PROCESSED_RECORDS is incremented by MapRunner after invoking the map function.
getAutoIncrReducerProcCount(Configuration) - Static method in class org.apache.hadoop.mapred.SkipBadRecords
Get the flag which if set to true, SkipBadRecords.COUNTER_REDUCE_PROCESSED_GROUPS is incremented by framework after invoking the reduce function.
getAvgMapTime() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.AnalyzedJob
Get the average map time
getAvgReduceTime() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.AnalyzedJob
Get the average reduce time
getAvgShuffleTime() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.AnalyzedJob
Get the average shuffle time
getBasePathInJarOut(String) - Method in class org.apache.hadoop.streaming.JarBuilder
 
getBaseRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.mapred.lib.MultipleOutputFormat
 
getBaseRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.mapred.lib.MultipleSequenceFileOutputFormat
 
getBaseRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.mapred.lib.MultipleTextOutputFormat
 
getBlackListedTaskTrackerCount() - Method in class org.apache.hadoop.mapreduce.ClusterMetrics
Get the number of blacklisted trackers in the cluster.
getBlackListedTaskTrackers() - Method in class org.apache.hadoop.mapreduce.Cluster
Get blacklisted trackers.
getBlacklistedTrackerNames() - Method in class org.apache.hadoop.mapred.ClusterStatus
Get the names of task trackers in the cluster.
getBlacklistedTrackers() - Method in class org.apache.hadoop.mapred.ClusterStatus
Get the number of blacklisted task trackers in the cluster.
getBlackListedTrackersInfo() - Method in class org.apache.hadoop.mapred.ClusterStatus
Gets the list of blacklisted trackers along with reasons for blacklisting.
getBlackListReport() - Method in class org.apache.hadoop.mapred.ClusterStatus.BlackListInfo
Gets a descriptive report about why the tasktracker was blacklisted.
getBlacklistReport() - Method in class org.apache.hadoop.mapreduce.TaskTrackerInfo
Gets a descriptive report about why the tasktracker was blacklisted.
getBlockIndex(BlockLocation[], long) - Method in class org.apache.hadoop.mapred.FileInputFormat
 
getBlockIndex(BlockLocation[], long) - Method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
 
getBoundAntProperty(String, String) - Static method in class org.apache.hadoop.streaming.StreamUtil
 
getBoundingValsQuery() - Method in class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat
 
getCacheArchives() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get cache archives set in the Configuration
getCacheArchives() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getCacheArchives() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getCacheFiles() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get cache files set in the Configuration
getCacheFiles() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getCacheFiles() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getCategory(List<List<Pentomino.ColumnName>>) - Method in class org.apache.hadoop.examples.dancing.Pentomino
Find whether the solution has the x in the upper left quadrant, the x-midline, the y-midline or in the center.
getChildQueues(String) - Method in class org.apache.hadoop.mapred.JobClient
Returns an array of queue information objects about immediate children of queue queueName.
getChildQueues(String) - Method in class org.apache.hadoop.mapreduce.Cluster
Returns immediate children of queueName.
getChildren() - Method in class org.apache.hadoop.mapred.JobQueueInfo
 
getClassByName(String) - Static method in class org.apache.hadoop.contrib.utils.join.DataJoinJob
 
getCleanupFinished() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.SummarizedJob
Get number of cleanup tasks that finished
getCleanupProgress() - Method in class org.apache.hadoop.mapreduce.JobStatus
 
getCleanupStarted() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.SummarizedJob
Get number of cleanup tasks started
getCleanupTaskReports(JobID) - Method in class org.apache.hadoop.mapred.JobClient
Get the information of the current state of the cleanup tasks of a job.
getClientInput() - Method in class org.apache.hadoop.streaming.PipeMapRed
Returns the DataInput from which the client output is read.
getClientOutput() - Method in class org.apache.hadoop.streaming.PipeMapRed
Returns the DataOutput to which the client input is written.
getClusterHandle() - Method in class org.apache.hadoop.mapred.JobClient
Get a handle to the Cluster
getClusterNick() - Method in class org.apache.hadoop.streaming.StreamJob
Deprecated. 
getClusterStatus() - Method in class org.apache.hadoop.mapred.JobClient
Get status information about the Map-Reduce cluster.
getClusterStatus(boolean) - Method in class org.apache.hadoop.mapred.JobClient
Get status information about the Map-Reduce cluster.
getClusterStatus() - Method in class org.apache.hadoop.mapreduce.Cluster
Get current cluster status.
getCollector(String, Reporter) - Method in class org.apache.hadoop.mapred.lib.MultipleOutputs
Gets the output collector for a named output.
getCollector(String, String, Reporter) - Method in class org.apache.hadoop.mapred.lib.MultipleOutputs
Gets the output collector for a multi named output.
getColumnName(int) - Method in class org.apache.hadoop.examples.dancing.DancingLinks
Get the name of a given column as a string
getCombinerClass() - Method in class org.apache.hadoop.mapred.JobConf
Get the user-defined combiner class used to combine map-outputs before being sent to the reducers.
getCombinerClass() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get the combiner class for the job.
getCombinerClass() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getCombinerClass() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getCombinerOutput() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.DoubleValueSum
 
getCombinerOutput() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.LongValueMax
 
getCombinerOutput() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.LongValueMin
 
getCombinerOutput() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.LongValueSum
 
getCombinerOutput() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.StringValueMax
 
getCombinerOutput() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.StringValueMin
 
getCombinerOutput() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.UniqValueCount
 
getCombinerOutput() - Method in interface org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregator
 
getCombinerOutput() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.ValueHistogram
 
getComparator() - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
Return comparator defining the ordering for RecordReaders in this composite.
getComparator() - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
Return comparator defining the ordering for RecordReaders in this composite.
getCompletionPollInterval(Configuration) - Static method in class org.apache.hadoop.mapreduce.Job
The interval at which waitForCompletion() should check.
getCompressMapOutput() - Method in class org.apache.hadoop.mapred.JobConf
Are the outputs of the maps be compressed?
getCompressOutput(JobConf) - Static method in class org.apache.hadoop.mapred.FileOutputFormat
Is the job output compressed?
getCompressOutput(JobContext) - Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
Is the job output compressed?
getConditions() - Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
 
getConf() - Method in class org.apache.hadoop.examples.terasort.TeraSort.SimplePartitioner
 
getConf() - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
getConf() - Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
 
getConf() - Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat
 
getConf() - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.FilterBase
 
getConf() - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
getConf() - Method in class org.apache.hadoop.mapreduce.lib.partition.BinaryPartitioner
 
getConf() - Method in class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedComparator
 
getConf() - Method in class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedPartitioner
 
getConf() - Method in class org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner
 
getConf() - Method in class org.apache.hadoop.streaming.DumpTypedBytes
 
getConf() - Method in class org.apache.hadoop.streaming.LoadTypedBytes
 
getConf() - Method in class org.apache.hadoop.streaming.StreamJob
 
getConf() - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
getConfiguration() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Get the underlying configuration object.
getConfiguration() - Method in interface org.apache.hadoop.mapred.RunningJob
Get the underlying job configuration
getConfiguration() - Method in interface org.apache.hadoop.mapreduce.JobContext
Return the configuration for the job.
getConfiguration() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getConfiguration() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getConfiguration() - Method in class org.apache.hadoop.streaming.PipeMapRed
Returns the Configuration.
getConnection() - Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
Returns a connection object o the DB
getConnection() - Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat
 
getConnection() - Method in class org.apache.hadoop.mapreduce.lib.db.DBOutputFormat.DBRecordWriter
 
getConnection() - Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
 
getContext() - Method in class org.apache.hadoop.streaming.PipeMapRed
 
getCounter() - Method in class org.apache.hadoop.mapred.Counters.Counter
What is the current value of this counter?
getCounter(Enum) - Method in class org.apache.hadoop.mapred.Counters
Returns current value of the specified counter, or 0 if the counter does not exist.
getCounter(String) - Method in class org.apache.hadoop.mapred.Counters.Group
Returns the value of the specified counter, or 0 if the counter does not exist.
getCounter(int, String) - Method in class org.apache.hadoop.mapred.Counters.Group
Deprecated. use Counters.Group.getCounter(String) instead
getCounter(Counters, String, String) - Method in class org.apache.hadoop.mapred.JobClient
 
getCounter(Enum<?>) - Method in interface org.apache.hadoop.mapred.Reporter
Get the Counters.Counter of the given group with the given name.
getCounter(String, String) - Method in interface org.apache.hadoop.mapred.Reporter
Get the Counters.Counter of the given group with the given name.
getCounter(Enum<?>) - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getCounter(String, String) - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getCounter(Enum) - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getCounter(String, String) - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getCounter(Enum<?>) - Method in interface org.apache.hadoop.mapreduce.TaskInputOutputContext
Get the Counter for the given counterName.
getCounter(String, String) - Method in interface org.apache.hadoop.mapreduce.TaskInputOutputContext
Get the Counter for the given groupName and counterName.
getCounter(Counters, String, String) - Method in class org.apache.hadoop.mapreduce.tools.CLI
 
getCounterForName(String) - Method in class org.apache.hadoop.mapred.Counters.Group
Get the counter for the given name and create it if it doesn't exist.
getCounters() - Method in interface org.apache.hadoop.mapred.RunningJob
Gets the counters for this job.
getCounters() - Method in class org.apache.hadoop.mapred.TaskReport
 
getCounters() - Method in class org.apache.hadoop.mapreduce.Job
Gets the counters for this job.
getCounters() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskAttemptInfo
Get the counters for the attempt
getCounters() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskInfo
Get the counters for this task
getCountersEnabled(JobConf) - Static method in class org.apache.hadoop.mapred.lib.MultipleOutputs
Returns if the counters for the named outputs are enabled or not.
getCountersEnabled(JobContext) - Static method in class org.apache.hadoop.mapreduce.lib.output.MultipleOutputs
Returns if the counters for the named outputs are enabled or not.
getCountQuery() - Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat
Returns the query for getting the total number of rows, subclasses can override this for custom behaviour.
getCurrentKey() - Method in class org.apache.hadoop.examples.MultiFileWordCount.CombineFileLineRecordReader
 
getCurrentKey() - Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
Get the current key
getCurrentKey() - Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader
 
getCurrentKey() - Method in class org.apache.hadoop.mapreduce.lib.input.KeyValueLineRecordReader
 
getCurrentKey() - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
 
getCurrentKey() - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsTextRecordReader
 
getCurrentKey() - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileRecordReader
 
getCurrentKey() - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
 
getCurrentKey() - Method in class org.apache.hadoop.mapreduce.lib.join.WrappedRecordReader
Get current key
getCurrentKey() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getCurrentKey() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getCurrentKey() - Method in class org.apache.hadoop.mapreduce.RecordReader
Get the current key
getCurrentKey() - Method in interface org.apache.hadoop.mapreduce.TaskInputOutputContext
Get the current key.
getCurrentKey() - Method in class org.apache.hadoop.streaming.io.OutputReader
Returns the current key.
getCurrentKey() - Method in class org.apache.hadoop.streaming.io.RawBytesOutputReader
 
getCurrentKey() - Method in class org.apache.hadoop.streaming.io.TextOutputReader
 
getCurrentKey() - Method in class org.apache.hadoop.streaming.io.TypedBytesOutputReader
 
getCurrentSegmentGeneration(Directory) - Static method in class org.apache.hadoop.contrib.index.lucene.LuceneUtil
Get the generation (N) of the current segments_N file in the directory.
getCurrentSegmentGeneration(String[]) - Static method in class org.apache.hadoop.contrib.index.lucene.LuceneUtil
Get the generation (N) of the current segments_N file from a list of files.
getCurrentSplit(JobConf) - Static method in class org.apache.hadoop.streaming.StreamUtil
 
getCurrentValue() - Method in class org.apache.hadoop.examples.MultiFileWordCount.CombineFileLineRecordReader
 
getCurrentValue(V) - Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
 
getCurrentValue() - Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
Get the current value.
getCurrentValue() - Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader
 
getCurrentValue() - Method in class org.apache.hadoop.mapreduce.lib.input.KeyValueLineRecordReader
 
getCurrentValue() - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
 
getCurrentValue() - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsTextRecordReader
 
getCurrentValue() - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileRecordReader
 
getCurrentValue() - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
Get the current value.
getCurrentValue() - Method in class org.apache.hadoop.mapreduce.lib.join.WrappedRecordReader
Get current value
getCurrentValue() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getCurrentValue() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getCurrentValue() - Method in class org.apache.hadoop.mapreduce.RecordReader
Get the current value.
getCurrentValue() - Method in interface org.apache.hadoop.mapreduce.TaskInputOutputContext
Get the current value.
getCurrentValue() - Method in class org.apache.hadoop.streaming.io.OutputReader
Returns the current value.
getCurrentValue() - Method in class org.apache.hadoop.streaming.io.RawBytesOutputReader
 
getCurrentValue() - Method in class org.apache.hadoop.streaming.io.TextOutputReader
 
getCurrentValue() - Method in class org.apache.hadoop.streaming.io.TypedBytesOutputReader
 
getData() - Method in class org.apache.hadoop.contrib.utils.join.TaggedMapOutput
 
getDBConf() - Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat
 
getDBConf() - Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
 
getDBProductName() - Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat
 
getDecommissionedTaskTrackerCount() - Method in class org.apache.hadoop.mapreduce.ClusterMetrics
Get the number of decommissioned trackers in the cluster.
getDefaultMaps() - Method in class org.apache.hadoop.mapred.JobClient
Get status information about the max available Maps in the cluster.
getDefaultReduces() - Method in class org.apache.hadoop.mapred.JobClient
Get status information about the max available Reduces in the cluster.
getDefaultWorkFile(TaskAttemptContext, String) - Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
Get the default path and filename for the output format.
getDelegate() - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
Obtain an iterator over the child RRs apropos of the value type ultimately emitted from this join.
getDelegate() - Method in class org.apache.hadoop.mapred.join.JoinRecordReader
Return an iterator wrapping the JoinCollector.
getDelegate() - Method in class org.apache.hadoop.mapred.join.MultiFilterRecordReader
Return an iterator returning a single value from the tuple.
getDelegate() - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
Obtain an iterator over the child RRs apropos of the value type ultimately emitted from this join.
getDelegate() - Method in class org.apache.hadoop.mapreduce.lib.join.JoinRecordReader
Return an iterator wrapping the JoinCollector.
getDelegate() - Method in class org.apache.hadoop.mapreduce.lib.join.MultiFilterRecordReader
Return an iterator returning a single value from the tuple.
getDelegationToken(Text) - Method in class org.apache.hadoop.mapred.JobClient
Get a delegation token for the user from the JobTracker.
getDelegationToken(Text) - Method in class org.apache.hadoop.mapreduce.Cluster
Get a delegation token for the user from the JobTracker.
getDelegationToken(String) - Static method in class org.apache.hadoop.mapreduce.security.TokenCache
 
getDependentJobs() - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
 
getDependingJobs() - Method in class org.apache.hadoop.mapred.jobcontrol.Job
 
getDirectory() - Method in class org.apache.hadoop.contrib.index.mapred.IntermediateForm
Get the ram directory of the intermediate form.
getDirectory() - Method in class org.apache.hadoop.contrib.index.mapred.Shard
Get the directory where this shard resides.
getDisplayName() - Method in class org.apache.hadoop.mapred.Counters.Group
Returns localized name of the group.
getDisplayName() - Method in class org.apache.hadoop.mapreduce.Counter
Get the name of the counter.
getDisplayName() - Method in class org.apache.hadoop.mapreduce.CounterGroup
Get the display name of the group.
getDistributionPolicyClass() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Get the distribution policy class.
getDocument() - Method in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
Get the document.
getDocumentAnalyzerClass() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Get the analyzer class.
getDoubleValue(Object) - Method in class org.apache.hadoop.contrib.utils.join.JobBase
 
getElement() - Method in interface org.apache.hadoop.examples.pi.Container
 
getElement() - Method in class org.apache.hadoop.examples.pi.DistSum.Machine.SummationSplit
getElement() - Method in class org.apache.hadoop.examples.pi.math.Bellard.Sum
getElement() - Method in class org.apache.hadoop.examples.pi.math.Summation
getElement() - Method in class org.apache.hadoop.examples.pi.SummationWritable
getElement() - Method in class org.apache.hadoop.examples.pi.TaskResult
getEnd() - Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit
 
getEntry(MapFile.Reader[], Partitioner<K, V>, K, V) - Static method in class org.apache.hadoop.mapred.MapFileOutputFormat
Get an entry from output generated by this class.
getEntry(MapFile.Reader[], Partitioner<K, V>, K, V) - Static method in class org.apache.hadoop.mapreduce.lib.output.MapFileOutputFormat
Get an entry from output generated by this class.
getError() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskAttemptInfo
Get the error string
getError() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskInfo
Get the error
getEventId() - Method in class org.apache.hadoop.mapreduce.TaskCompletionEvent
Returns event Id.
getExecutable(JobConf) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
Get the URI of the application's executable.
getFailedDueToAttemptId() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskInfo
Get the attempt Id that caused this task to fail
getFailedJobList() - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl
 
getFailedJobs() - Method in class org.apache.hadoop.mapred.jobcontrol.JobControl
 
getFailedMaps() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.JobInfo
Get the total number of failed maps
getFailedReduces() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.JobInfo
Get the number of failed reduces
getFieldNames() - Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
 
getFieldSeparator() - Method in class org.apache.hadoop.streaming.PipeMapper
 
getFieldSeparator() - Method in class org.apache.hadoop.streaming.PipeMapRed
Returns the field separator to be used.
getFieldSeparator() - Method in class org.apache.hadoop.streaming.PipeReducer
 
getFileClassPaths() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get the file entries in classpath as an array of Path
getFileClassPaths() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getFileClassPaths() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getFileSystem() - Method in class org.apache.hadoop.mapreduce.Cluster
Get the file system where job-specific files are stored
getFileTimestamps() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get the timestamps of the files.
getFileTimestamps() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getFileTimestamps() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getFilter() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.FilteredJob
Get the current filter
getFilteredMap() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.FilteredJob
Get the map of the filtered tasks
getFinalSync(JobContext) - Static method in class org.apache.hadoop.examples.terasort.TeraOutputFormat
Does the user want a final sync at close?
getFinishedMaps() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.JobInfo
Get the number of finished maps
getFinishedReduces() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.JobInfo
Get the number of finished reduces
getFinishTime() - Method in class org.apache.hadoop.mapreduce.Job
Get finish time of the job.
getFinishTime() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.JobInfo
Get the job finish time
getFinishTime() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskAttemptInfo
Get the finish time of the attempt
getFinishTime() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskInfo
Get the finish time of this task
getFinishTime() - Method in class org.apache.hadoop.mapreduce.JobStatus
Get the finish time of the job.
getFirst() - Method in class org.apache.hadoop.examples.SecondarySort.IntPair
 
getFlippable() - Method in class org.apache.hadoop.examples.dancing.Pentomino.Piece
 
getFormatMinSplitSize() - Method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
Get the lower bound on split size imposed by the format.
getFormatMinSplitSize() - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFormat
 
getFs() - Method in class org.apache.hadoop.mapred.JobClient
Get a filesystem handle.
getGeneration() - Method in class org.apache.hadoop.contrib.index.mapred.Shard
Get the generation of the Lucene instance.
getGroup(String) - Method in class org.apache.hadoop.mapred.Counters
Returns the named counter group, or an empty group if there is none with the specified name.
getGroup(String) - Method in class org.apache.hadoop.mapreduce.Counters
Returns the named counter group, or an empty group if there is none with the specified name.
getGroupingComparator() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get the user defined RawComparator comparator for grouping keys of inputs to the reduce.
getGroupingComparator() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getGroupingComparator() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getGroupNames() - Method in class org.apache.hadoop.mapred.Counters
Returns the names of all counter classes.
getGroupNames() - Method in class org.apache.hadoop.mapreduce.Counters
Returns the names of all counter classes.
getHadoopClientHome() - Method in class org.apache.hadoop.streaming.StreamJob
 
getHistoryFile() - Method in class org.apache.hadoop.mapreduce.JobStatus
 
getHistoryUrl() - Method in interface org.apache.hadoop.mapred.RunningJob
Get the url where history file is archived.
getHistoryUrl() - Method in class org.apache.hadoop.mapreduce.Job
 
getHost() - Method in class org.apache.hadoop.streaming.Environment
 
getHostname() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskAttemptInfo
Get the host name
getHttpPort() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskAttemptInfo
Get the HTTP port for the tracker
getID() - Method in interface org.apache.hadoop.mapred.RunningJob
Get the job identifier.
getId() - Method in class org.apache.hadoop.mapreduce.ID
returns the int which represents the identifier
getIndexInputFormatClass() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Get the index input format class.
getIndexMaxFieldLength() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Get the max field length for a Lucene instance.
getIndexMaxNumSegments() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Get the max number of segments for a Lucene instance.
getIndexShards() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Get the string representation of a number of shards.
getIndexShards(IndexUpdateConfiguration) - Static method in class org.apache.hadoop.contrib.index.mapred.Shard
 
getIndexUpdaterClass() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Get the index updater class.
getIndexUseCompoundFile() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Check whether to use the compound file format for a Lucene instance.
getInputBoundingQuery() - Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
 
getInputClass() - Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
 
getInputConditions() - Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
 
getInputCountQuery() - Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
 
getInputDataLength() - Method in class org.apache.hadoop.mapreduce.split.JobSplit.SplitMetaInfo
 
getInputDataLength() - Method in class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitMetaInfo
 
getInputFieldNames() - Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
 
getInputFileBasedOutputFileName(JobConf, String) - Method in class org.apache.hadoop.mapred.lib.MultipleOutputFormat
Generate the outfile name based on a given anme and the input file name.
getInputFormat() - Method in class org.apache.hadoop.mapred.JobConf
Get the InputFormat implementation for the map-reduce job, defaults to TextInputFormat if not specified explicity.
getInputFormatClass() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get the InputFormat class for the job.
getInputFormatClass() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getInputFormatClass() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getInputOrderBy() - Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
 
getInputPathFilter(JobConf) - Static method in class org.apache.hadoop.mapred.FileInputFormat
Get a PathFilter instance of the filter set for the input paths.
getInputPathFilter(JobContext) - Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
Get a PathFilter instance of the filter set for the input paths.
getInputPaths(JobConf) - Static method in class org.apache.hadoop.mapred.FileInputFormat
Get the list of input Paths for the map-reduce job.
getInputPaths(JobContext) - Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
Get the list of input Paths for the map-reduce job.
getInputQuery() - Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
 
getInputSeparator() - Method in class org.apache.hadoop.streaming.PipeMapper
 
getInputSeparator() - Method in class org.apache.hadoop.streaming.PipeMapRed
Returns the input separator to be used.
getInputSeparator() - Method in class org.apache.hadoop.streaming.PipeReducer
 
getInputSplit() - Method in interface org.apache.hadoop.mapred.Reporter
Get the InputSplit object for a map.
getInputSplit() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
Get the input split for this map.
getInputSplit() - Method in interface org.apache.hadoop.mapreduce.MapContext
Get the input split for this map.
getInputTableName() - Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
 
getInputWriterClass() - Method in class org.apache.hadoop.streaming.io.IdentifierResolver
Returns the resolved InputWriter class.
getInstance(Cluster) - Static method in class org.apache.hadoop.mapreduce.Job
 
getInstance(Cluster, Configuration) - Static method in class org.apache.hadoop.mapreduce.Job
 
getInstance(Cluster, JobStatus, Configuration) - Static method in class org.apache.hadoop.mapreduce.Job
 
getIOSortMB() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Get the IO sort space in MB.
getIsCleanup() - Method in class org.apache.hadoop.mapred.TaskLogAppender
Get whether task is cleanup attempt or not.
getIsJavaMapper(JobConf) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
Check whether the job is using a Java Mapper.
getIsJavaRecordReader(JobConf) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
Check whether the job is using a Java RecordReader
getIsJavaRecordWriter(JobConf) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
Will the reduce use a Java RecordWriter?
getIsJavaReducer(JobConf) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
Check whether the job is using a Java Reducer.
getJar() - Method in class org.apache.hadoop.mapred.JobConf
Get the user jar for the map-reduce job.
getJar() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get the pathname of the job's jar.
getJar() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getJar() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getJarUnpackPattern() - Method in class org.apache.hadoop.mapred.JobConf
Get the pattern for jar contents to unpack on the tasktracker
getJob(JobID) - Method in class org.apache.hadoop.mapred.JobClient
Get an RunningJob object to track an ongoing job.
getJob(String) - Method in class org.apache.hadoop.mapred.JobClient
Deprecated. Applications should rather use JobClient.getJob(JobID).
getJob() - Method in class org.apache.hadoop.mapred.lib.CombineFileSplit
 
getJob(JobID) - Method in class org.apache.hadoop.mapreduce.Cluster
Get job corresponding to jobid.
getJob() - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
 
getJobACLs() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.JobInfo
 
getJobACLs() - Method in class org.apache.hadoop.mapreduce.JobStatus
 
getJobClient() - Method in class org.apache.hadoop.mapred.jobcontrol.Job
 
getJobConf() - Method in interface org.apache.hadoop.mapred.JobContext
Get the job Configuration
getJobConf() - Method in class org.apache.hadoop.mapred.jobcontrol.Job
 
getJobConf() - Method in interface org.apache.hadoop.mapred.TaskAttemptContext
 
getJobConfPath() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.JobInfo
Get the path for the job configuration file
getJobEndNotificationURI() - Method in class org.apache.hadoop.mapred.JobConf
Get the uri to be invoked in-order to send a notification after the job has completed (success/failure).
getJobFile() - Method in interface org.apache.hadoop.mapred.RunningJob
Get the path of the submitted job configuration.
getJobFile() - Method in class org.apache.hadoop.mapreduce.Job
Get the path of the submitted job configuration.
getJobFile() - Method in class org.apache.hadoop.mapreduce.JobStatus
Get the configuration file for the job.
getJobHistoryUrl(JobID) - Method in class org.apache.hadoop.mapreduce.Cluster
Get the job history file path for a given job id.
getJobId() - Method in class org.apache.hadoop.mapred.JobStatus
Deprecated. use getJobID instead
getJobID() - Method in class org.apache.hadoop.mapred.JobStatus
 
getJobID() - Method in interface org.apache.hadoop.mapred.RunningJob
Deprecated. This method is deprecated and will be removed. Applications should rather use RunningJob.getID().
getJobID() - Method in class org.apache.hadoop.mapred.TaskAttemptID
 
getJobID() - Method in class org.apache.hadoop.mapred.TaskID
 
getJobID() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get the unique ID for the job.
getJobId() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.JobInfo
Get the job id
getJobID() - Method in class org.apache.hadoop.mapreduce.JobStatus
 
getJobID() - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
 
getJobID() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getJobID() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getJobID() - Method in class org.apache.hadoop.mapreduce.TaskAttemptID
Returns the JobID object that this task attempt belongs to
getJobID() - Method in class org.apache.hadoop.mapreduce.TaskID
Returns the JobID object that this tip belongs to
getJobIDsPattern(String, Integer) - Static method in class org.apache.hadoop.mapred.JobID
Deprecated. 
getJobLocalDir() - Method in class org.apache.hadoop.mapred.JobConf
Get job-specific shared directory for use as scratch space
getJobName() - Method in class org.apache.hadoop.mapred.JobConf
Get the user-specified job name.
getJobName() - Method in interface org.apache.hadoop.mapred.RunningJob
Get the name of the job.
getJobName() - Method in class org.apache.hadoop.mapreduce.Job
The user-specified job name.
getJobName() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get the user-specified job name.
getJobname() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.JobInfo
Get the job name
getJobName() - Method in class org.apache.hadoop.mapreduce.JobStatus
Get the user-specified job name.
getJobName() - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
 
getJobName() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getJobName() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getJobPriority() - Method in class org.apache.hadoop.mapred.JobConf
Get the JobPriority for this job.
getJobPriority() - Method in class org.apache.hadoop.mapred.JobStatus
Return the priority of the job
getJobRunState(int) - Static method in class org.apache.hadoop.mapred.JobStatus
Helper method to get human-readable state of the job.
getJobSetupCleanupNeeded() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get whether job-setup and job-cleanup is needed for the job
getJobSetupCleanupNeeded() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getJobSetupCleanupNeeded() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getJobsFromQueue(String) - Method in class org.apache.hadoop.mapred.JobClient
Gets all the jobs which were added to particular Job Queue
getJobState() - Method in interface org.apache.hadoop.mapred.RunningJob
Returns the current state of the Job.
getJobState() - Method in class org.apache.hadoop.mapreduce.Job
Returns the current state of the Job.
getJobState() - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
 
getJobStatus() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.JobInfo
Get the job status
getJobStatuses() - Method in class org.apache.hadoop.mapreduce.QueueInfo
Get the jobs submitted to queue
getJobToken(TokenStorage) - Static method in class org.apache.hadoop.mapreduce.security.TokenCache
 
getJobTrackerHostPort() - Method in class org.apache.hadoop.streaming.StreamJob
 
getJobTrackerState() - Method in class org.apache.hadoop.mapred.ClusterStatus
Get the current state of the JobTracker, as JobTracker.State
getJobTrackerState() - Method in class org.apache.hadoop.mapreduce.Cluster
Get JobTracker's state
getJtIdentifier() - Method in class org.apache.hadoop.mapreduce.JobID
 
getKeepCommandFile(JobConf) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
Does the user want to keep the command file for debugging? If this is true, pipes will write a copy of the command data to a file in the task directory named "downlink.data", which may be used to run the C++ program under the debugger.
getKeepFailedTaskFiles() - Method in class org.apache.hadoop.mapred.JobConf
Should the temporary files for failed tasks be kept?
getKeepTaskFilesPattern() - Method in class org.apache.hadoop.mapred.JobConf
Get the regular expression that is matched against the task names to see if we need to keep the files.
getKey() - Method in class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionHelper
 
getKeyClass() - Method in class org.apache.hadoop.mapred.KeyValueLineRecordReader
 
getKeyClass() - Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
The class of key that must be passed to SequenceFileRecordReader.next(Object, Object)..
getKeyClassName() - Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
Retrieve the name of the key class for this SequenceFile.
getKeyClassName() - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
Retrieve the name of the key class for this SequenceFile.
getKeyFieldComparatorOption() - Method in class org.apache.hadoop.mapred.JobConf
Get the KeyFieldBasedComparator options
getKeyFieldComparatorOption(JobContext) - Static method in class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedComparator
Get the KeyFieldBasedComparator options
getKeyFieldPartitionerOption() - Method in class org.apache.hadoop.mapred.JobConf
Get the KeyFieldBasedPartitioner options
getKeyFieldPartitionerOption(JobContext) - Method in class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedPartitioner
Get the KeyFieldBasedPartitioner options
getLastOutput() - Method in class org.apache.hadoop.streaming.io.OutputReader
Returns the last output from the client as a String.
getLastOutput() - Method in class org.apache.hadoop.streaming.io.RawBytesOutputReader
 
getLastOutput() - Method in class org.apache.hadoop.streaming.io.TextOutputReader
 
getLastOutput() - Method in class org.apache.hadoop.streaming.io.TypedBytesOutputReader
 
getLaunchTime() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.JobInfo
Get the job launch time
getLength() - Method in class org.apache.hadoop.examples.BaileyBorweinPlouffe.BbpSplit
Get the size of the split, so that the input splits can be sorted by size.
getLength() - Method in class org.apache.hadoop.examples.pi.DistSum.Machine.SummationSplit
Get the size of the split, so that the input splits can be sorted by size.
getLength() - Method in class org.apache.hadoop.mapred.FileSplit
The number of bytes in the file to process.
getLength() - Method in interface org.apache.hadoop.mapred.InputSplit
Get the total number of bytes in the data of the InputSplit.
getLength() - Method in class org.apache.hadoop.mapred.join.CompositeInputSplit
Return the aggregate length of all child InputSplits currently added.
getLength(int) - Method in class org.apache.hadoop.mapred.join.CompositeInputSplit
Get the length of ith child InputSplit.
getLength() - Method in class org.apache.hadoop.mapreduce.InputSplit
Get the size of the split, so that the input splits can be sorted by size.
getLength() - Method in class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat.DataDrivenDBInputSplit
 
getLength() - Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit
 
getLength() - Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
 
getLength(int) - Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
Returns the length of the ith Path
getLength() - Method in class org.apache.hadoop.mapreduce.lib.input.FileSplit
The number of bytes in the file to process.
getLength() - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeInputSplit
Return the aggregate length of all child InputSplits currently added.
getLength(int) - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeInputSplit
Get the length of ith child InputSplit.
getLengths() - Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
Returns an array containing the lengths of the files in the split
getLocalAnalysisClass() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Get the local analysis class.
getLocalCacheArchives() - Method in interface org.apache.hadoop.mapreduce.JobContext
Return the path array of the localized caches
getLocalCacheArchives() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getLocalCacheArchives() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getLocalCacheFiles() - Method in interface org.apache.hadoop.mapreduce.JobContext
Return the path array of the localized files
getLocalCacheFiles() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getLocalCacheFiles() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getLocalDirs() - Method in class org.apache.hadoop.mapred.JobConf
 
getLocalPath(String) - Method in class org.apache.hadoop.mapred.JobConf
Constructs a local file name.
getLocation(int) - Method in class org.apache.hadoop.mapred.join.CompositeInputSplit
getLocations from ith InputSplit.
getLocation(int) - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeInputSplit
getLocations from ith InputSplit.
getLocations() - Method in class org.apache.hadoop.examples.BaileyBorweinPlouffe.BbpSplit
No location is needed.
getLocations() - Method in class org.apache.hadoop.examples.pi.DistSum.Machine.SummationSplit
Get the list of nodes by name where the data for the split would be local.
getLocations() - Method in class org.apache.hadoop.mapred.FileSplit
 
getLocations() - Method in interface org.apache.hadoop.mapred.InputSplit
Get the list of hostnames where the input split is located.
getLocations() - Method in class org.apache.hadoop.mapred.join.CompositeInputSplit
Collect a set of hosts from all child InputSplits.
getLocations() - Method in class org.apache.hadoop.mapred.MultiFileSplit
 
getLocations() - Method in class org.apache.hadoop.mapreduce.InputSplit
Get the list of nodes by name where the data for the split would be local.
getLocations() - Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit
Get the list of nodes by name where the data for the split would be local.
getLocations() - Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
Returns all the Paths where this input-split resides
getLocations() - Method in class org.apache.hadoop.mapreduce.lib.input.FileSplit
 
getLocations() - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeInputSplit
Collect a set of hosts from all child InputSplits.
getLocations() - Method in class org.apache.hadoop.mapreduce.split.JobSplit.SplitMetaInfo
 
getLocations() - Method in class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitMetaInfo
 
getLongValue(Object) - Method in class org.apache.hadoop.contrib.utils.join.JobBase
 
getLowerClause() - Method in class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat.DataDrivenDBInputSplit
 
getMapContext(MapContext<KEYIN, VALUEIN, KEYOUT, VALUEOUT>) - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper
Get a wrapped WrappedMapper.Context for custom implementations.
getMapCounters() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.JobInfo
Get the map counters for the job
getMapDebugScript() - Method in class org.apache.hadoop.mapred.JobConf
Get the map task's debug script.
getMapFinished() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.SummarizedJob
Get number of maps that finished
getMapFinishTime() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskAttemptInfo
Get the map finish time.
getMapOutputCompressorClass(Class<? extends CompressionCodec>) - Method in class org.apache.hadoop.mapred.JobConf
Get the CompressionCodec for compressing the map outputs.
getMapOutputKeyClass() - Static method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateMapper
Get the map output key class.
getMapOutputKeyClass() - Method in class org.apache.hadoop.mapred.JobConf
Get the key class for the map output data.
getMapOutputKeyClass() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get the key class for the map output data.
getMapOutputKeyClass() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getMapOutputKeyClass() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getMapOutputValueClass() - Static method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateMapper
Get the map output value class.
getMapOutputValueClass() - Method in class org.apache.hadoop.mapred.JobConf
Get the value class for the map output data.
getMapOutputValueClass() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get the value class for the map output data.
getMapOutputValueClass() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getMapOutputValueClass() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getMapper() - Method in class org.apache.hadoop.mapred.MapRunner
 
getMapperClass() - Method in class org.apache.hadoop.mapred.JobConf
Get the Mapper class for the job.
getMapperClass() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get the Mapper class for the job.
getMapperClass(JobContext) - Static method in class org.apache.hadoop.mapreduce.lib.map.MultithreadedMapper
Get the application's mapper class.
getMapperClass() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getMapperClass() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getMapperMaxSkipRecords(Configuration) - Static method in class org.apache.hadoop.mapred.SkipBadRecords
Get the number of acceptable skip records surrounding the bad record PER bad record in mapper.
getMapProgress() - Method in class org.apache.hadoop.mapreduce.JobStatus
 
getMapredJobID() - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
 
getMapredTempDir() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Get the Map/Reduce temp directory.
getMapRunnerClass() - Method in class org.apache.hadoop.mapred.JobConf
Get the MapRunnable class for the job.
getMapSlotCapacity() - Method in class org.apache.hadoop.mapreduce.ClusterMetrics
Get the total number of map slots in the cluster.
getMapSpeculativeExecution() - Method in class org.apache.hadoop.mapred.JobConf
Should speculative execution be used for this job for map tasks? Defaults to true.
getMapStarted() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.SummarizedJob
Get number of maps that were started
getMapTaskReports(JobID) - Method in class org.apache.hadoop.mapred.JobClient
Get the information of the current state of the map tasks of a job.
getMapTaskReports(String) - Method in class org.apache.hadoop.mapred.JobClient
Deprecated. Applications should rather use JobClient.getMapTaskReports(JobID)
getMapTasks() - Method in class org.apache.hadoop.mapred.ClusterStatus
Get the number of currently running map tasks in the cluster.
getMapTasks() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.AnalyzedJob
Get the map tasks list
getMaxMapAttempts() - Method in class org.apache.hadoop.mapred.JobConf
Get the configured number of maximum attempts that will be made to run a map task, as specified by the mapreduce.map.maxattempts property.
getMaxMapAttempts() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get the configured number of maximum attempts that will be made to run a map task, as specified by the mapred.map.max.attempts property.
getMaxMapAttempts() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getMaxMapAttempts() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getMaxMapTaskFailuresPercent() - Method in class org.apache.hadoop.mapred.JobConf
Get the maximum percentage of map tasks that can fail without the job being aborted.
getMaxMapTasks() - Method in class org.apache.hadoop.mapred.ClusterStatus
Get the maximum capacity for running map tasks in the cluster.
getMaxMemory() - Method in class org.apache.hadoop.mapred.ClusterStatus
Get the maximum configured heap memory that can be used by the JobTracker
getMaxPhysicalMemoryForTask() - Method in class org.apache.hadoop.mapred.JobConf
Deprecated. this variable is deprecated and nolonger in use.
getMaxRAMSizeInBytes() - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Get the max ram index size in bytes.
getMaxReduceAttempts() - Method in class org.apache.hadoop.mapred.JobConf
Get the configured number of maximum attempts that will be made to run a reduce task, as specified by the mapreduce.reduce.maxattempts property.
getMaxReduceAttempts() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get the configured number of maximum attempts that will be made to run a reduce task, as specified by the mapred.reduce.max.attempts property.
getMaxReduceAttempts() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getMaxReduceAttempts() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getMaxReduceTaskFailuresPercent() - Method in class org.apache.hadoop.mapred.JobConf
Get the maximum percentage of reduce tasks that can fail without the job being aborted.
getMaxReduceTasks() - Method in class org.apache.hadoop.mapred.ClusterStatus
Get the maximum capacity for running reduce tasks in the cluster.
getMaxSplitSize(JobContext) - Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
Get the maximum split size.
getMaxTaskFailuresPerTracker() - Method in class org.apache.hadoop.mapred.JobConf
Expert: Get the maximum no.
getMaxVirtualMemoryForTask() - Method in class org.apache.hadoop.mapred.JobConf
Deprecated. Use JobConf.getMemoryForMapTask() and JobConf.getMemoryForReduceTask()
getMemoryForMapTask() - Method in class org.apache.hadoop.mapred.JobConf
Get memory required to run a map task of the job, in MB.
getMemoryForReduceTask() - Method in class org.apache.hadoop.mapred.JobConf
Get memory required to run a reduce task of the job, in MB.
getMessage() - Method in exception org.apache.hadoop.mapred.InvalidInputException
Get a summary message of the problems found.
getMessage() - Method in exception org.apache.hadoop.mapreduce.lib.input.InvalidInputException
Get a summary message of the problems found.
getMessage() - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
 
getMinSplitSize(JobContext) - Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
Get the minimum split size
getName() - Method in class org.apache.hadoop.examples.dancing.Pentomino.Piece
 
getName() - Method in class org.apache.hadoop.mapred.Counters.Group
Returns raw name of the group.
getName() - Method in class org.apache.hadoop.mapreduce.Counter
 
getName() - Method in class org.apache.hadoop.mapreduce.CounterGroup
Get the internal name of the group
getNamedOutputFormatClass(JobConf, String) - Static method in class org.apache.hadoop.mapred.lib.MultipleOutputs
Returns the named output OutputFormat.
getNamedOutputKeyClass(JobConf, String) - Static method in class org.apache.hadoop.mapred.lib.MultipleOutputs
Returns the key class for a named output.
getNamedOutputs() - Method in class org.apache.hadoop.mapred.lib.MultipleOutputs
Returns iterator with the defined name outputs.
getNamedOutputsList(JobConf) - Static method in class org.apache.hadoop.mapred.lib.MultipleOutputs
Returns list of channel names.
getNamedOutputValueClass(JobConf, String) - Static method in class org.apache.hadoop.mapred.lib.MultipleOutputs
Returns the value class for a named output.
getNode() - Method in class org.apache.hadoop.mapred.join.Parser.NodeToken
 
getNode() - Method in class org.apache.hadoop.mapred.join.Parser.Token
 
getNode() - Method in class org.apache.hadoop.mapreduce.lib.join.Parser.NodeToken
 
getNode() - Method in class org.apache.hadoop.mapreduce.lib.join.Parser.Token
 
getNum() - Method in class org.apache.hadoop.mapred.join.Parser.NumToken
 
getNum() - Method in class org.apache.hadoop.mapred.join.Parser.Token
 
getNum() - Method in class org.apache.hadoop.mapreduce.lib.join.Parser.NumToken
 
getNum() - Method in class org.apache.hadoop.mapreduce.lib.join.Parser.Token
 
getNumberColumns() - Method in class org.apache.hadoop.examples.dancing.DancingLinks
Get the number of columns.
getNumberOfThreads(JobContext) - Static method in class org.apache.hadoop.mapreduce.lib.map.MultithreadedMapper
The number of threads in the thread pool that will run the map function.
getNumExcludedNodes() - Method in class org.apache.hadoop.mapred.ClusterStatus
Get the number of excluded hosts in the cluster.
getNumFailedCleanups() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.SummarizedJob
Get number of failed cleanup tasks
getNumFailedMaps() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.SummarizedJob
Get number of failed maps
getNumFailedReduces() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.SummarizedJob
Get number of failed reduces
getNumFailedSetups() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.SummarizedJob
Get number of failed set up tasks
getNumFinishedCleanups() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.SummarizedJob
Get number of cleanup tasks that finished
getNumFinishedSetups() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.SummarizedJob
Get number of finished set up tasks
getNumKilledCleanups() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.SummarizedJob
Get number of killed cleanup tasks
getNumKilledMaps() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.SummarizedJob
Get number of killed maps
getNumKilledReduces() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.SummarizedJob
Get number of killed reduces
getNumKilledSetups() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.SummarizedJob
Get number of killed set up tasks
getNumLinesPerSplit(JobContext) - Static method in class org.apache.hadoop.mapreduce.lib.input.NLineInputFormat
Get the number of lines per split
getNumMapTasks() - Method in class org.apache.hadoop.mapred.JobConf
Get configured the number of reduce tasks for this job.
getNumOfKeyFields() - Method in class org.apache.hadoop.streaming.PipeMapper
 
getNumOfKeyFields() - Method in class org.apache.hadoop.streaming.PipeMapRed
Returns the number of key fields.
getNumOfKeyFields() - Method in class org.apache.hadoop.streaming.PipeReducer
 
getNumPaths() - Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
Returns the number of Paths in the split
getNumReduceTasks() - Method in class org.apache.hadoop.mapred.JobConf
Get configured the number of reduce tasks for this job.
getNumReduceTasks() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get configured the number of reduce tasks for this job.
getNumReduceTasks() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getNumReduceTasks() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getNumTasksToExecutePerJvm() - Method in class org.apache.hadoop.mapred.JobConf
Get the number of tasks that a spawned JVM should execute
getOccupiedMapSlots() - Method in class org.apache.hadoop.mapreduce.ClusterMetrics
Get number of occupied map slots in the cluster.
getOccupiedReduceSlots() - Method in class org.apache.hadoop.mapreduce.ClusterMetrics
Get the number of occupied reduce slots in the cluster.
getOffset(int) - Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
Returns the start offset of the ith Path
getOp() - Method in class org.apache.hadoop.contrib.index.example.LineDocTextAndOp
Get the type of the operation.
getOp() - Method in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
Get the type of operation.
getOperations() - Method in class org.apache.hadoop.mapreduce.QueueAclsInfo
Get opearations allowed on queue.
getOutputCommitter(TaskAttemptContext) - Method in class org.apache.hadoop.examples.terasort.TeraOutputFormat
 
getOutputCommitter() - Method in class org.apache.hadoop.mapred.JobConf
Get the OutputCommitter implementation for the map-reduce job, defaults to FileOutputCommitter if not specified explicitly.
getOutputCommitter(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.db.DBOutputFormat
 
getOutputCommitter() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getOutputCommitter(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
 
getOutputCommitter(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.output.FilterOutputFormat
 
getOutputCommitter(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.output.LazyOutputFormat
 
getOutputCommitter(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.output.NullOutputFormat
 
getOutputCommitter() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getOutputCommitter(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.OutputFormat
Get the output committer for this output format.
getOutputCommitter() - Method in interface org.apache.hadoop.mapreduce.TaskInputOutputContext
Get the OutputCommitter for the task-attempt.
getOutputCompressionType(JobConf) - Static method in class org.apache.hadoop.mapred.SequenceFileOutputFormat
Get the SequenceFile.CompressionType for the output SequenceFile.
getOutputCompressionType(JobContext) - Static method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat
Get the SequenceFile.CompressionType for the output SequenceFile.
getOutputCompressorClass(JobConf, Class<? extends CompressionCodec>) - Static method in class org.apache.hadoop.mapred.FileOutputFormat
Get the CompressionCodec for compressing the job outputs.
getOutputCompressorClass(JobContext, Class<? extends CompressionCodec>) - Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
Get the CompressionCodec for compressing the job outputs.
getOutputFieldCount() - Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
 
getOutputFieldNames() - Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
 
getOutputFormat() - Method in class org.apache.hadoop.mapred.JobConf
Get the OutputFormat implementation for the map-reduce job, defaults to TextOutputFormat if not specified explicity.
getOutputFormatClass() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get the OutputFormat class for the job.
getOutputFormatClass() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getOutputFormatClass() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getOutputKeyClass() - Static method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateReducer
Get the reduce output key class.
getOutputKeyClass() - Method in class org.apache.hadoop.mapred.JobConf
Get the key class for the job output data.
getOutputKeyClass() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get the key class for the job output data.
getOutputKeyClass() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getOutputKeyClass() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getOutputKeyClass() - Method in class org.apache.hadoop.streaming.io.IdentifierResolver
Returns the resolved output key class.
getOutputKeyComparator() - Method in class org.apache.hadoop.mapred.JobConf
Get the RawComparator comparator used to compare keys.
getOutputName(JobContext) - Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
Get the base output name for the output file.
getOutputPath(JobConf) - Static method in class org.apache.hadoop.mapred.FileOutputFormat
Get the Path to the output directory for the map-reduce job.
getOutputPath(JobContext) - Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
Get the Path to the output directory for the map-reduce job.
getOutputReaderClass() - Method in class org.apache.hadoop.streaming.io.IdentifierResolver
Returns the resolved OutputReader class.
getOutputReplication(JobContext) - Static method in class org.apache.hadoop.examples.terasort.TeraSort
 
getOutputTableName() - Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
 
getOutputValueClass() - Static method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateReducer
Get the reduce output value class.
getOutputValueClass() - Method in class org.apache.hadoop.mapred.JobConf
Get the value class for job outputs.
getOutputValueClass() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get the value class for job outputs.
getOutputValueClass() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getOutputValueClass() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getOutputValueClass() - Method in class org.apache.hadoop.streaming.io.IdentifierResolver
Returns the resolved output value class.
getOutputValueGroupingComparator() - Method in class org.apache.hadoop.mapred.JobConf
Get the user defined WritableComparable comparator for grouping keys of inputs to the reduce.
getPartition(Shard, IntermediateForm, int) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdatePartitioner
 
getPartition(IntWritable, SummationWritable, int) - Method in class org.apache.hadoop.examples.pi.DistSum.ReduceSide.IndexPartitioner
Return the index as the partition.
getPartition(SecondarySort.IntPair, IntWritable, int) - Method in class org.apache.hadoop.examples.SecondarySort.FirstPartitioner
 
getPartition(Text, Text, int) - Method in class org.apache.hadoop.examples.terasort.TeraSort.SimplePartitioner
 
getPartition(K2, V2, int) - Method in class org.apache.hadoop.mapred.lib.HashPartitioner
Use Object.hashCode() to partition.
getPartition(K2, V2, int) - Method in interface org.apache.hadoop.mapred.Partitioner
Get the paritition number for a given key (hence record) given the total number of partitions i.e.
getPartition(BinaryComparable, V, int) - Method in class org.apache.hadoop.mapreduce.lib.partition.BinaryPartitioner
Use (the specified slice of the array returned by) BinaryComparable.getBytes() to partition.
getPartition(K, V, int) - Method in class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner
Use Object.hashCode() to partition.
getPartition(K2, V2, int) - Method in class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedPartitioner
 
getPartition(int, int) - Method in class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedPartitioner
 
getPartition(K, V, int) - Method in class org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner
 
getPartition(KEY, VALUE, int) - Method in class org.apache.hadoop.mapreduce.Partitioner
Get the partition number for a given key (hence record) given the total number of partitions i.e.
getPartitionerClass() - Method in class org.apache.hadoop.mapred.JobConf
Get the Partitioner used to partition Mapper-outputs to be sent to the Reducers.
getPartitionerClass() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get the Partitioner class for the job.
getPartitionerClass() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getPartitionerClass() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getPartitionFile(Configuration) - Static method in class org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner
Get the path to the SequenceFile storing the sorted partition keyset.
getPath() - Method in class org.apache.hadoop.mapred.FileSplit
The file containing this split's data.
getPath(int) - Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
Returns the ith Path
getPath() - Method in class org.apache.hadoop.mapreduce.lib.input.FileSplit
The file containing this split's data.
getPathForCustomFile(JobConf, String) - Static method in class org.apache.hadoop.mapred.FileOutputFormat
Helper function to generate a Path for a file that is unique for the task within the job output directory.
getPathForWorkFile(TaskInputOutputContext<?, ?, ?, ?>, String, String) - Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
Helper function to generate a Path for a file that is unique for the task within the job output directory.
getPaths() - Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
Returns all the Paths in the split
getPos() - Method in class org.apache.hadoop.contrib.index.example.LineDocRecordReader
 
getPos() - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
Unsupported (returns zero in all cases).
getPos() - Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
Request position from proxied RR.
getPos() - Method in class org.apache.hadoop.mapred.KeyValueLineRecordReader
 
getPos() - Method in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
return the amount of data processed
getPos() - Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBRecordReader
 
getPos() - Method in interface org.apache.hadoop.mapred.RecordReader
Returns the current position in the input.
getPos() - Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
 
getPos() - Method in class org.apache.hadoop.mapred.SequenceFileAsTextRecordReader
 
getPos() - Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
 
getPos() - Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
Deprecated.  
getPos() - Method in class org.apache.hadoop.streaming.StreamBaseRecordReader
Returns the current position in the input.
getPriority() - Method in class org.apache.hadoop.mapreduce.Job
Get scheduling info of the job.
getPriority() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.JobInfo
Get the priority of this job
getPriority() - Method in class org.apache.hadoop.mapreduce.JobStatus
Return the priority of the job
getProblems() - Method in exception org.apache.hadoop.mapred.InvalidInputException
Get the complete list of the problems reported.
getProblems() - Method in exception org.apache.hadoop.mapreduce.lib.input.InvalidInputException
Get the complete list of the problems reported.
getProfileEnabled() - Method in class org.apache.hadoop.mapred.JobConf
Get whether the task profiling is enabled.
getProfileEnabled() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get whether the task profiling is enabled.
getProfileEnabled() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getProfileEnabled() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getProfileParams() - Method in class org.apache.hadoop.mapred.JobConf
Get the profiler configuration arguments.
getProfileParams() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get the profiler configuration arguments.
getProfileParams() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getProfileParams() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getProfileTaskRange(boolean) - Method in class org.apache.hadoop.mapred.JobConf
Get the range of maps or reduces to profile.
getProfileTaskRange(boolean) - Method in interface org.apache.hadoop.mapreduce.JobContext
Get the range of maps or reduces to profile.
getProfileTaskRange(boolean) - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getProfileTaskRange(boolean) - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getProgress() - Method in class org.apache.hadoop.contrib.index.example.LineDocRecordReader
 
getProgress() - Method in class org.apache.hadoop.examples.MultiFileWordCount.CombineFileLineRecordReader
 
getProgress() - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
Report progress as the minimum of all child RR progress.
getProgress() - Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
Request progress from proxied RR.
getProgress() - Method in class org.apache.hadoop.mapred.KeyValueLineRecordReader
 
getProgress() - Method in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
return progress based on the amount of data processed so far.
getProgress() - Method in interface org.apache.hadoop.mapred.RecordReader
How much of the input has the RecordReader consumed i.e.
getProgress() - Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
Return the progress within the input split
getProgress() - Method in class org.apache.hadoop.mapred.SequenceFileAsTextRecordReader
 
getProgress() - Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
Return the progress within the input split
getProgress() - Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
The current progress of the record reader through its data.
getProgress() - Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader
return progress based on the amount of data processed so far.
getProgress() - Method in class org.apache.hadoop.mapreduce.lib.input.KeyValueLineRecordReader
 
getProgress() - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
Return the progress within the input split
getProgress() - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsTextRecordReader
 
getProgress() - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileRecordReader
Return the progress within the input split
getProgress() - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
Report progress as the minimum of all child RR progress.
getProgress() - Method in class org.apache.hadoop.mapreduce.lib.join.WrappedRecordReader
Request progress from proxied RR.
getProgress() - Method in class org.apache.hadoop.mapreduce.RecordReader
The current progress of the record reader through its data.
getProgress() - Method in class org.apache.hadoop.streaming.StreamBaseRecordReader
 
getProgressible() - Method in interface org.apache.hadoop.mapred.JobContext
Get the progress mechanism for reporting progress.
getProgressible() - Method in interface org.apache.hadoop.mapred.TaskAttemptContext
 
getProgressPollInterval(Configuration) - Static method in class org.apache.hadoop.mapreduce.Job
The interval at which monitorAndPrintJob() prints status
getProperties() - Method in class org.apache.hadoop.mapreduce.QueueInfo
Get properties.
getQueue(String) - Method in class org.apache.hadoop.mapreduce.Cluster
Get queue information for the specified name.
getQueueAclsForCurrentUser() - Method in class org.apache.hadoop.mapred.JobClient
Gets the Queue ACLs for current user
getQueueAclsForCurrentUser() - Method in class org.apache.hadoop.mapreduce.Cluster
Gets the Queue ACLs for current user
getQueueChildren() - Method in class org.apache.hadoop.mapreduce.QueueInfo
Get immediate children.
getQueueInfo(String) - Method in class org.apache.hadoop.mapred.JobClient
Gets the queue information associated to a particular Job Queue
getQueueName() - Method in class org.apache.hadoop.mapred.JobConf
Return the name of the queue to which this job is submitted.
getQueueName() - Method in class org.apache.hadoop.mapreduce.QueueAclsInfo
Get queue name.
getQueueName() - Method in class org.apache.hadoop.mapreduce.QueueInfo
Get the queue name from JobQueueInfo
getQueues() - Method in class org.apache.hadoop.mapred.JobClient
Return an array of queue information objects about all the Job Queues configured.
getQueues() - Method in class org.apache.hadoop.mapreduce.Cluster
Get all the queues in cluster.
getReaders(FileSystem, Path, Configuration) - Static method in class org.apache.hadoop.mapred.MapFileOutputFormat
Open the output generated by this format.
getReaders(Configuration, Path) - Static method in class org.apache.hadoop.mapred.SequenceFileOutputFormat
Open the output generated by this format.
getReaders(Path, Configuration) - Static method in class org.apache.hadoop.mapreduce.lib.output.MapFileOutputFormat
Open the output generated by this format.
getReadyJobs() - Method in class org.apache.hadoop.mapred.jobcontrol.JobControl
 
getReadyJobsList() - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl
 
getReasonForBlacklist() - Method in class org.apache.hadoop.mapreduce.TaskTrackerInfo
Gets the reason for which the tasktracker was blacklisted.
getReasonForBlackListing() - Method in class org.apache.hadoop.mapred.ClusterStatus.BlackListInfo
Gets the reason for which the tasktracker was blacklisted.
getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.contrib.index.example.LineDocInputFormat
 
getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.mapred.FileInputFormat
 
getRecordReader(InputSplit, JobConf, Reporter) - Method in interface org.apache.hadoop.mapred.InputFormat
Get the RecordReader for the given InputSplit.
getRecordReader(InputSplit, JobConf, Reporter) - Method in interface org.apache.hadoop.mapred.join.ComposableInputFormat
 
getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.mapred.join.CompositeInputFormat
Construct a CompositeRecordReader for the children of this InputFormat as defined in the init expression.
getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.mapred.KeyValueTextInputFormat
 
getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.mapred.lib.CombineFileInputFormat
This is not implemented yet.
getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat
Get the RecordReader for the given InputSplit.
getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.mapred.lib.NLineInputFormat
 
getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.mapred.MultiFileInputFormat
 
getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat
 
getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.mapred.SequenceFileAsTextInputFormat
 
getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.mapred.SequenceFileInputFilter
Create a record reader for the given split
getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.mapred.SequenceFileInputFormat
 
getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.mapred.TextInputFormat
 
getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.streaming.AutoInputFormat
 
getRecordReader(InputSplit, JobConf, Reporter) - Method in class org.apache.hadoop.streaming.StreamInputFormat
 
getRecordReaderQueue() - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
Return sorted list of RecordReaders for this composite.
getRecordReaderQueue() - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
Return sorted list of RecordReaders for this composite.
getRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateOutputFormat
 
getRecordWriter(TaskAttemptContext) - Method in class org.apache.hadoop.examples.terasort.TeraOutputFormat
 
getRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.mapred.FileOutputFormat
 
getRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.mapred.lib.db.DBOutputFormat
Get the RecordWriter for the given job.
getRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.mapred.lib.FilterOutputFormat
 
getRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.mapred.lib.LazyOutputFormat
 
getRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.mapred.lib.MultipleOutputFormat
Create a composite record writer that can write key/value data to different output files
getRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.mapred.lib.NullOutputFormat
 
getRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.mapred.MapFileOutputFormat
 
getRecordWriter(FileSystem, JobConf, String, Progressable) - Method in interface org.apache.hadoop.mapred.OutputFormat
Get the RecordWriter for the given job.
getRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat
 
getRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.mapred.SequenceFileOutputFormat
 
getRecordWriter(FileSystem, JobConf, String, Progressable) - Method in class org.apache.hadoop.mapred.TextOutputFormat
 
getRecordWriter(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.db.DBOutputFormat
Get the RecordWriter for the given task.
getRecordWriter(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
 
getRecordWriter(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.output.FilterOutputFormat
 
getRecordWriter(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.output.LazyOutputFormat
 
getRecordWriter(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.output.MapFileOutputFormat
 
getRecordWriter(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.output.NullOutputFormat
 
getRecordWriter(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat
 
getRecordWriter(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat
 
getRecordWriter(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat
 
getRecordWriter(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.OutputFormat
Get the RecordWriter for the given task.
getReduceCounters() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.JobInfo
Get the reduce counters for the job
getReduceDebugScript() - Method in class org.apache.hadoop.mapred.JobConf
Get the reduce task's debug Script
getReduceFinished() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.SummarizedJob
Get number of reducers that finished
getReduceProgress() - Method in class org.apache.hadoop.mapreduce.JobStatus
 
getReducerClass() - Method in class org.apache.hadoop.mapred.JobConf
Get the Reducer class for the job.
getReducerClass() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get the Reducer class for the job.
getReducerClass() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getReducerClass() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getReducerContext(ReduceContext<KEYIN, VALUEIN, KEYOUT, VALUEOUT>) - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer
A a wrapped WrappedReducer.Context for custom implementations.
getReducerMaxSkipGroups(Configuration) - Static method in class org.apache.hadoop.mapred.SkipBadRecords
Get the number of acceptable skip groups surrounding the bad group PER bad group in reducer.
getReduceSlotCapacity() - Method in class org.apache.hadoop.mapreduce.ClusterMetrics
Get the total number of reduce slots in the cluster.
getReduceSpeculativeExecution() - Method in class org.apache.hadoop.mapred.JobConf
Should speculative execution be used for this job for reduce tasks? Defaults to true.
getReduceStarted() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.SummarizedJob
Get number of Reducers that were started
getReduceTaskReports(JobID) - Method in class org.apache.hadoop.mapred.JobClient
Get the information of the current state of the reduce tasks of a job.
getReduceTaskReports(String) - Method in class org.apache.hadoop.mapred.JobClient
Deprecated. Applications should rather use JobClient.getReduceTaskReports(JobID)
getReduceTasks() - Method in class org.apache.hadoop.mapred.ClusterStatus
Get the number of currently running reduce tasks in the cluster.
getReduceTasks() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.AnalyzedJob
Get the reduce tasks list
getReport() - Method in class org.apache.hadoop.contrib.utils.join.JobBase
log the counters
getReport() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.DoubleValueSum
 
getReport() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.LongValueMax
 
getReport() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.LongValueMin
 
getReport() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.LongValueSum
 
getReport() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.StringValueMax
 
getReport() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.StringValueMin
 
getReport() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.UniqValueCount
 
getReport() - Method in interface org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregator
 
getReport() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.ValueHistogram
 
getReportDetails() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.ValueHistogram
 
getReportItems() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.ValueHistogram
 
getRepresentingCharacter(TaskType) - Static method in class org.apache.hadoop.mapreduce.TaskID
Gets the character representing the TaskType
getReservedMapSlots() - Method in class org.apache.hadoop.mapreduce.ClusterMetrics
Get number of reserved map slots in the cluster.
getReservedReduceSlots() - Method in class org.apache.hadoop.mapreduce.ClusterMetrics
Get the number of reserved reduce slots in the cluster.
getResult() - Method in class org.apache.hadoop.examples.Sort
Get the last job that was run using this instance.
getRootQueues() - Method in class org.apache.hadoop.mapred.JobClient
Returns an array of queue information objects about root level queues configured
getRootQueues() - Method in class org.apache.hadoop.mapreduce.Cluster
Gets the root level queues.
getRotations() - Method in class org.apache.hadoop.examples.dancing.Pentomino.Piece
 
getRunningJobList() - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl
 
getRunningJobs() - Method in class org.apache.hadoop.mapred.jobcontrol.JobControl
 
getRunningMaps() - Method in class org.apache.hadoop.mapreduce.ClusterMetrics
Get the number of running map tasks in the cluster.
getRunningReduces() - Method in class org.apache.hadoop.mapreduce.ClusterMetrics
Get the number of running reduce tasks in the cluster.
getRunningTaskAttempts() - Method in class org.apache.hadoop.mapred.TaskReport
Get the running task attempt IDs for this task
getRunState() - Method in class org.apache.hadoop.mapred.JobStatus
 
getSample(InputFormat<K, V>, Job) - Method in class org.apache.hadoop.mapreduce.lib.partition.InputSampler.IntervalSampler
For each split sampled, emit when the ratio of the number of records retained to the total record count is less than the specified frequency.
getSample(InputFormat<K, V>, Job) - Method in class org.apache.hadoop.mapreduce.lib.partition.InputSampler.RandomSampler
Randomize the split order, then take the specified number of keys from each split sampled, where each key is selected with the specified probability and possibly replaced by a subsequently selected key when the quota of keys from that split is satisfied.
getSample(InputFormat<K, V>, Job) - Method in interface org.apache.hadoop.mapreduce.lib.partition.InputSampler.Sampler
For a given job, collect and return a subset of the keys from the input data.
getSample(InputFormat<K, V>, Job) - Method in class org.apache.hadoop.mapreduce.lib.partition.InputSampler.SplitSampler
From each split sampled, take the first numSamples / numSplits records.
getSchedulingInfo() - Method in class org.apache.hadoop.mapreduce.Job
Get scheduling info of the job.
getSchedulingInfo() - Method in class org.apache.hadoop.mapreduce.JobStatus
Gets the Scheduling information associated to a particular Job.
getSchedulingInfo() - Method in class org.apache.hadoop.mapreduce.QueueInfo
Gets the scheduling information associated to particular job queue.
getSecond() - Method in class org.apache.hadoop.examples.SecondarySort.IntPair
 
getSecretKey(Text) - Static method in class org.apache.hadoop.mapreduce.security.TokenCache
auxiliary method to get user's secret keys..
getSelectQuery() - Method in class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBRecordReader
Returns the query for selecting the records, subclasses can override this for custom behaviour.
getSelectQuery() - Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
Returns the query for selecting the records, subclasses can override this for custom behaviour.
getSelectQuery() - Method in class org.apache.hadoop.mapreduce.lib.db.OracleDBRecordReader
Returns the query for selecting the records from an Oracle DB.
getSequenceFileOutputKeyClass(JobConf) - Static method in class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat
Get the key class for the SequenceFile
getSequenceFileOutputKeyClass(JobContext) - Static method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat
Get the key class for the SequenceFile
getSequenceFileOutputValueClass(JobConf) - Static method in class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat
Get the value class for the SequenceFile
getSequenceFileOutputValueClass(JobContext) - Static method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat
Get the value class for the SequenceFile
getSequenceWriter(TaskAttemptContext, Class<?>, Class<?>) - Method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat
 
getSessionId() - Method in class org.apache.hadoop.mapred.JobConf
Deprecated. 
getSetupFinished() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.SummarizedJob
Get number of setup tasks that finished
getSetupProgress() - Method in class org.apache.hadoop.mapreduce.JobStatus
 
getSetupStarted() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.SummarizedJob
Get number of setup tasks that started
getSetupTaskReports(JobID) - Method in class org.apache.hadoop.mapred.JobClient
Get the information of the current state of the setup tasks of a job.
getShape(boolean, int) - Method in class org.apache.hadoop.examples.dancing.Pentomino.Piece
 
getShuffleFinishTime() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskAttemptInfo
Get the shuffle finish time.
getSize() - Method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat.WritableValueBytes
 
getSkipOutputPath(Configuration) - Static method in class org.apache.hadoop.mapred.SkipBadRecords
Get the directory to which skipped records are written.
getSortComparator() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get the RawComparator comparator used to compare keys.
getSortComparator() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getSortComparator() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getSortFinishTime() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskAttemptInfo
Get the sort finish time.
getSpace(int) - Static method in class org.apache.hadoop.streaming.StreamUtil
 
getSpeculativeExecution() - Method in class org.apache.hadoop.mapred.JobConf
Should speculative execution be used for this job? Defaults to true.
getSplit() - Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
 
getSplitHosts(BlockLocation[], long, long, NetworkTopology) - Method in class org.apache.hadoop.mapred.FileInputFormat
This function identifies and returns the hosts that contribute most for a given split.
getSplitIndex() - Method in class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitMetaInfo
 
getSplitLocation() - Method in class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitIndex
 
getSplitLocation() - Method in class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitMetaInfo
 
getSplitLocations() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskInfo
Get the split locations
getSplits(JobContext) - Method in class org.apache.hadoop.examples.BaileyBorweinPlouffe.BbpInputFormat
Logically split the set of input files for the job.
getSplits(int) - Method in class org.apache.hadoop.examples.dancing.Pentomino
Generate a list of prefixes to a given depth
getSplits(JobContext) - Method in class org.apache.hadoop.examples.pi.DistSum.MapSide.PartitionInputFormat
Partitions the summation into parts and then return them as splits
getSplits(JobContext) - Method in class org.apache.hadoop.examples.pi.DistSum.ReduceSide.SummationInputFormat
 
getSplits(JobContext) - Method in class org.apache.hadoop.examples.terasort.TeraInputFormat
 
getSplits(JobConf, int) - Method in class org.apache.hadoop.mapred.FileInputFormat
Splits files returned by FileInputFormat.listStatus(JobConf) when they're too big.
getSplits(JobConf, int) - Method in interface org.apache.hadoop.mapred.InputFormat
Logically split the set of input files for the job.
getSplits(JobConf, int) - Method in class org.apache.hadoop.mapred.join.CompositeInputFormat
Build a CompositeInputSplit from the child InputFormats by assigning the ith split from each child to the ith composite split.
getSplits(JobConf, int) - Method in class org.apache.hadoop.mapred.lib.CombineFileInputFormat
 
getSplits(JobConf, int) - Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat
Logically split the set of input files for the job.
getSplits(JobConf, int) - Method in class org.apache.hadoop.mapred.lib.NLineInputFormat
Logically splits the set of input files for the job, splits N lines of the input as one split.
getSplits(JobConf, int) - Method in class org.apache.hadoop.mapred.MultiFileInputFormat
 
getSplits(JobContext) - Method in class org.apache.hadoop.mapreduce.InputFormat
Logically split the set of input files for the job.
getSplits(JobContext) - Method in class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat
Logically split the set of input files for the job.
getSplits(JobContext) - Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat
Logically split the set of input files for the job.
getSplits(JobContext) - Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat
 
getSplits(JobContext) - Method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
Generate the list of files and make them into FileSplits.
getSplits(JobContext) - Method in class org.apache.hadoop.mapreduce.lib.input.NLineInputFormat
Logically splits the set of input files for the job, splits N lines of the input as one split.
getSplits(JobContext) - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeInputFormat
Build a CompositeInputSplit from the child InputFormats by assigning the ith split from each child to the ith composite split.
getSplitsForFile(FileStatus, Configuration, int) - Static method in class org.apache.hadoop.mapreduce.lib.input.NLineInputFormat
 
getSplitter(int) - Method in class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat
 
getSplitter(int) - Method in class org.apache.hadoop.mapreduce.lib.db.OracleDataDrivenDBInputFormat
 
getStagingAreaDir() - Method in class org.apache.hadoop.mapreduce.Cluster
Grab the jobtracker's view of the staging directory path where job-specific files will be placed.
getStart() - Method in class org.apache.hadoop.mapred.FileSplit
The position of the first byte in the file to process.
getStart() - Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit
 
getStart() - Method in class org.apache.hadoop.mapreduce.lib.input.FileSplit
The position of the first byte in the file to process.
getStartOffset() - Method in class org.apache.hadoop.mapreduce.split.JobSplit.SplitMetaInfo
 
getStartOffset() - Method in class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitIndex
 
getStartOffset() - Method in class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitMetaInfo
 
getStartOffsets() - Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
Returns an array containing the start offsets of the files in the split
getStartTime() - Method in class org.apache.hadoop.mapreduce.Job
Get start time of the job.
getStartTime() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskAttemptInfo
Get the start time of the attempt
getStartTime() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskInfo
Get the start time of this task
getStartTime() - Method in class org.apache.hadoop.mapreduce.JobStatus
 
getState() - Method in class org.apache.hadoop.mapred.jobcontrol.Job
 
getState() - Method in class org.apache.hadoop.mapred.jobcontrol.JobControl
 
getState() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskAttemptInfo
Get the state
getState() - Method in class org.apache.hadoop.mapreduce.JobStatus
 
getState() - Method in class org.apache.hadoop.mapreduce.QueueInfo
Return the queue state
getState(String) - Static method in enum org.apache.hadoop.mapreduce.QueueState
 
getStatement() - Method in class org.apache.hadoop.mapreduce.lib.db.DBOutputFormat.DBRecordWriter
 
getStatement() - Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
 
getStateName() - Method in enum org.apache.hadoop.mapreduce.QueueState
 
getStatus() - Method in class org.apache.hadoop.mapreduce.Job
 
getStatus() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getStatus() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getStatus() - Method in interface org.apache.hadoop.mapreduce.TaskAttemptContext
Get the last set status message.
getStatus() - Method in class org.apache.hadoop.mapreduce.TaskCompletionEvent
Returns enum Status.SUCESS or Status.FAILURE.
getSteps() - Method in class org.apache.hadoop.examples.pi.math.ArithmeticProgression
Get the number of steps
getStr() - Method in class org.apache.hadoop.mapred.join.Parser.StrToken
 
getStr() - Method in class org.apache.hadoop.mapred.join.Parser.Token
 
getStr() - Method in class org.apache.hadoop.mapreduce.lib.join.Parser.StrToken
 
getStr() - Method in class org.apache.hadoop.mapreduce.lib.join.Parser.Token
 
getSubmitTime() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.JobInfo
Get the job submit time
getSuccessfulJobList() - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl
 
getSuccessfulJobs() - Method in class org.apache.hadoop.mapred.jobcontrol.JobControl
 
getSuccessfulTaskAttempt() - Method in class org.apache.hadoop.mapred.TaskReport
Get the attempt ID that took this task to completion
getSum() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.DoubleValueSum
 
getSum() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.LongValueSum
 
getSums(long, int, Map<Bellard.Parameter, List<T>>) - Static method in class org.apache.hadoop.examples.pi.math.Bellard
Get the sums for the Bellard formula.
getSymlink() - Method in interface org.apache.hadoop.mapreduce.JobContext
This method checks to see if symlinks are to be create for the localized cache files in the current working directory
getSymlink() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getSymlink() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getSystemDir() - Method in class org.apache.hadoop.mapred.JobClient
Grab the jobtracker system directory path where job-specific files are to be placed.
getSystemDir() - Method in class org.apache.hadoop.mapreduce.Cluster
Grab the jobtracker system directory path where job-specific files will be placed.
getTableName() - Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
 
getTag() - Method in class org.apache.hadoop.contrib.utils.join.TaggedMapOutput
 
getTaskAttemptID() - Method in interface org.apache.hadoop.mapred.TaskAttemptContext
 
getTaskAttemptId() - Method in class org.apache.hadoop.mapred.TaskCompletionEvent
Returns task id.
getTaskAttemptID() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getTaskAttemptID() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getTaskAttemptID() - Method in interface org.apache.hadoop.mapreduce.TaskAttemptContext
Get the unique name for this task attempt.
getTaskAttemptId() - Method in class org.apache.hadoop.mapreduce.TaskCompletionEvent
Returns task id.
getTaskAttemptIDsPattern(String, Integer, Boolean, Integer, Integer) - Static method in class org.apache.hadoop.mapred.TaskAttemptID
Deprecated. 
getTaskAttemptIDsPattern(String, Integer, TaskType, Integer, Integer) - Static method in class org.apache.hadoop.mapred.TaskAttemptID
Deprecated. 
getTaskCompletionEvents(int) - Method in interface org.apache.hadoop.mapred.RunningJob
Get events indicating completion (success/failure) of component tasks.
getTaskCompletionEvents(int, int) - Method in class org.apache.hadoop.mapreduce.Job
Get events indicating completion (success/failure) of component tasks.
getTaskDiagnostics(TaskAttemptID) - Method in interface org.apache.hadoop.mapred.RunningJob
Gets the diagnostic messages for a given task attempt.
getTaskDiagnostics(TaskAttemptID) - Method in class org.apache.hadoop.mapreduce.Job
Gets the diagnostic messages for a given task attempt.
getTaskID() - Method in class org.apache.hadoop.mapred.TaskAttemptID
 
getTaskId() - Method in class org.apache.hadoop.mapred.TaskCompletionEvent
Deprecated. use TaskCompletionEvent.getTaskAttemptId() instead.
getTaskId() - Method in class org.apache.hadoop.mapred.TaskLogAppender
Getter/Setter methods for log4j.
getTaskID() - Method in class org.apache.hadoop.mapred.TaskReport
The id of the task.
getTaskId() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskInfo
Get the Task ID
getTaskID() - Method in class org.apache.hadoop.mapreduce.TaskAttemptID
Returns the TaskID object that this task attempt belongs to
getTaskIDsPattern(String, Integer, Boolean, Integer) - Static method in class org.apache.hadoop.mapred.TaskID
Deprecated. Use TaskID.getTaskIDsPattern(String, Integer, TaskType, Integer)
getTaskIDsPattern(String, Integer, TaskType, Integer) - Static method in class org.apache.hadoop.mapred.TaskID
Deprecated. 
getTaskInfo(JobConf) - Static method in class org.apache.hadoop.streaming.StreamUtil
 
getTaskLogURL(TaskAttemptID, String) - Static method in class org.apache.hadoop.mapreduce.tools.CLI
 
getTaskOutputFilter(JobConf) - Static method in class org.apache.hadoop.mapred.JobClient
Get the task output filter out of the JobConf.
getTaskOutputFilter() - Method in class org.apache.hadoop.mapred.JobClient
Deprecated. 
getTaskOutputFilter(Configuration) - Static method in class org.apache.hadoop.mapreduce.Job
Get the task output filter.
getTaskOutputPath(JobConf, String) - Static method in class org.apache.hadoop.mapred.FileOutputFormat
Helper function to create the task's temporary output directory and return the path to the task's output file.
getTaskReports(TaskType) - Method in class org.apache.hadoop.mapreduce.Job
Get the information of the current state of the tasks of a job.
getTaskRunTime() - Method in class org.apache.hadoop.mapreduce.TaskCompletionEvent
Returns time (in millisec) the task took to complete.
getTaskStatus() - Method in class org.apache.hadoop.mapred.TaskCompletionEvent
Returns enum Status.SUCESS or Status.FAILURE.
getTaskStatus() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskAttemptInfo
Get the task status
getTaskStatus() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskInfo
Get the task status
getTaskTrackerCount() - Method in class org.apache.hadoop.mapreduce.ClusterMetrics
Get the number of active trackers in the cluster.
getTaskTrackerExpiryInterval() - Method in class org.apache.hadoop.mapreduce.Cluster
Get the tasktracker expiry interval for the cluster
getTaskTrackerHttp() - Method in class org.apache.hadoop.mapreduce.TaskCompletionEvent
http location of the tasktracker where this task ran.
getTaskTrackerName() - Method in class org.apache.hadoop.mapreduce.TaskTrackerInfo
Gets the tasktracker's name.
getTaskTrackers() - Method in class org.apache.hadoop.mapred.ClusterStatus
Get the number of task trackers in the cluster.
getTaskType() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskAttemptInfo
Get the task type
getTaskType() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskInfo
Get the task type
getTaskType() - Method in class org.apache.hadoop.mapreduce.TaskAttemptID
Returns the TaskType of the TaskAttemptID
getTaskType() - Method in class org.apache.hadoop.mapreduce.TaskID
Get the type of the task
getTaskType(char) - Static method in class org.apache.hadoop.mapreduce.TaskID
Gets the TaskType corresponding to the character
getTerm() - Method in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
Get the term.
getText() - Method in class org.apache.hadoop.contrib.index.example.LineDocTextAndOp
Get the text that represents a document.
getText() - Method in class org.apache.hadoop.contrib.index.mapred.DocumentID
The text of the document id.
getThreadState() - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl
 
getTokenStorage() - Static method in class org.apache.hadoop.mapreduce.security.TokenCache
 
getTotalCleanups() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.SummarizedJob
Get number of clean up tasks
getTotalCounters() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.JobInfo
Get the counters for the job
getTotalJobSubmissions() - Method in class org.apache.hadoop.mapreduce.ClusterMetrics
Get the total number of job submissions in the cluster.
getTotalLogFileSize() - Method in class org.apache.hadoop.mapred.TaskLogAppender
 
getTotalMaps() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.SummarizedJob
Get total maps
getTotalMaps() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.JobInfo
Get the total number of maps
getTotalReduces() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.SummarizedJob
Get total reduces
getTotalReduces() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.JobInfo
Get the total number of reduces
getTotalSetups() - Method in class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.SummarizedJob
Get number of set up tasks
getTrackerName() - Method in class org.apache.hadoop.mapred.ClusterStatus.BlackListInfo
Gets the blacklisted tasktracker's name.
getTrackerName() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskAttemptInfo
Get the tracker name where the attempt executed
getTrackingURL() - Method in interface org.apache.hadoop.mapred.RunningJob
Get the URL where some job progress information will be displayed.
getTrackingURL() - Method in class org.apache.hadoop.mapreduce.Job
Get the URL where some job progress information will be displayed.
getTrackingUrl() - Method in class org.apache.hadoop.mapreduce.JobStatus
Get the link to the web-ui for details of the job.
getTTExpiryInterval() - Method in class org.apache.hadoop.mapred.ClusterStatus
Get the tasktracker expiry interval for the cluster
getType() - Method in class org.apache.hadoop.mapred.join.Parser.Token
 
getType() - Method in class org.apache.hadoop.mapreduce.lib.join.Parser.Token
 
getType() - Method in class org.apache.hadoop.typedbytes.TypedBytesWritable
Get the type code embedded in the first byte.
getUniqueFile(TaskAttemptContext, String, String) - Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
Generate a unique filename, based on the task id, name, and extension
getUniqueItems() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.UniqValueCount
 
getUniqueName(JobConf, String) - Static method in class org.apache.hadoop.mapred.FileOutputFormat
Helper function to generate a name that is unique for the task.
getUpperClause() - Method in class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat.DataDrivenDBInputSplit
 
getURIs(String, String) - Method in class org.apache.hadoop.streaming.StreamJob
get the uris of all the files/caches
getUsedMemory() - Method in class org.apache.hadoop.mapred.ClusterStatus
Get the total heap memory used by the JobTracker
getUseNewMapper() - Method in class org.apache.hadoop.mapred.JobConf
Should the framework use the new context-object code for running the mapper?
getUseNewReducer() - Method in class org.apache.hadoop.mapred.JobConf
Should the framework use the new context-object code for running the reducer?
getUser() - Method in class org.apache.hadoop.mapred.JobConf
Get the reported username for this job.
getUser() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get the reported username for this job.
getUser() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getUser() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getUsername() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.JobInfo
Get the user name
getUsername() - Method in class org.apache.hadoop.mapreduce.JobStatus
 
getUseSimplePartitioner(JobContext) - Static method in class org.apache.hadoop.examples.terasort.TeraSort
 
getVal() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.LongValueMax
 
getVal() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.LongValueMin
 
getVal() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.StringValueMax
 
getVal() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.StringValueMin
 
getValue() - Method in class org.apache.hadoop.examples.pi.math.Bellard.Sum
get the value of sigma
getValue() - Method in class org.apache.hadoop.examples.pi.math.Summation
Return the value of this summation
getValue() - Method in class org.apache.hadoop.mapreduce.Counter
What is the current value of this counter?
getValue() - Method in enum org.apache.hadoop.mapreduce.JobStatus.State
 
getValue() - Method in class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionHelper
 
getValue() - Method in class org.apache.hadoop.typedbytes.TypedBytesWritable
Get the typed bytes as a Java object.
getValueAggregatorDescriptor(String, Configuration) - Static method in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorJobBase
 
getValueClass() - Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
The class of value that must be passed to SequenceFileRecordReader.next(Object, Object)..
getValueClassName() - Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
Retrieve the name of the value class for this SequenceFile.
getValueClassName() - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
Retrieve the name of the value class for this SequenceFile.
getValues() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getValues() - Method in interface org.apache.hadoop.mapreduce.ReduceContext
Iterate through the values for the current key, reusing the same value object, which is stored in the context.
getVersion() - Method in class org.apache.hadoop.contrib.index.mapred.Shard
Get the version number of the entire index.
getWaitingJobList() - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl
 
getWaitingJobs() - Method in class org.apache.hadoop.mapred.jobcontrol.JobControl
 
getWorkingDirectory() - Method in class org.apache.hadoop.mapred.JobConf
Get the current working directory for the default file system.
getWorkingDirectory() - Method in interface org.apache.hadoop.mapreduce.JobContext
Get the current working directory for the default file system.
getWorkingDirectory() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
getWorkingDirectory() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
getWorkOutputPath(JobConf) - Static method in class org.apache.hadoop.mapred.FileOutputFormat
Get the Path to the task's temporary output directory for the map-reduce job

Tasks' Side-Effect Files
getWorkOutputPath(TaskInputOutputContext<?, ?, ?, ?>) - Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
Get the Path to the task's temporary output directory for the map-reduce job

Tasks' Side-Effect Files
getWorkPath() - Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
Get the directory that the task should write results into
go() - Method in class org.apache.hadoop.streaming.StreamJob
Deprecated. use StreamJob.run(String[]) instead.
goodClassOrNull(Configuration, String, String) - Static method in class org.apache.hadoop.streaming.StreamUtil
It may seem strange to silently switch behaviour when a String is not a classname; the reason is simplified Usage:
Grep - Class in org.apache.hadoop.examples
 
Gridmix - Class in org.apache.hadoop.mapred.gridmix
Driver class for the Gridmix3 benchmark.
Gridmix() - Constructor for class org.apache.hadoop.mapred.gridmix.Gridmix
 
GRIDMIX_OUT_DIR - Static variable in class org.apache.hadoop.mapred.gridmix.Gridmix
Output (scratch) directory for submitted jobs.
GRIDMIX_QUE_DEP - Static variable in class org.apache.hadoop.mapred.gridmix.Gridmix
The depth of the queue of job descriptions.
GRIDMIX_SUB_MUL - Static variable in class org.apache.hadoop.mapred.gridmix.Gridmix
Multiplier to accelerate or decelerate job submission.
GRIDMIX_SUB_THR - Static variable in class org.apache.hadoop.mapred.gridmix.Gridmix
Number of submitting threads at the client and upper bound for in-memory split data.
GROUP - Static variable in class org.apache.hadoop.mapreduce.lib.map.RegexMapper
 


H

HadoopStreaming - Class in org.apache.hadoop.streaming
The main entrypoint.
HadoopStreaming() - Constructor for class org.apache.hadoop.streaming.HadoopStreaming
 
has(int) - Method in class org.apache.hadoop.mapreduce.lib.join.TupleWritable
Return true if tuple has an element at the position provided.
hashCode() - Method in class org.apache.hadoop.contrib.index.mapred.DocumentID
 
hashCode() - Method in class org.apache.hadoop.contrib.index.mapred.Shard
 
hashCode() - Method in class org.apache.hadoop.examples.MultiFileWordCount.WordOffset
 
hashCode() - Method in class org.apache.hadoop.examples.pi.math.ArithmeticProgression
Not supported
hashCode() - Method in class org.apache.hadoop.examples.pi.math.Summation
Not supported
hashCode() - Method in class org.apache.hadoop.examples.pi.SummationWritable
Not supported
hashCode() - Method in class org.apache.hadoop.examples.pi.TaskResult
Not supported
hashCode() - Method in class org.apache.hadoop.examples.SecondarySort.IntPair
 
hashCode() - Method in class org.apache.hadoop.mapred.Counters.Group
 
hashCode() - Method in class org.apache.hadoop.mapred.Counters
 
hashCode() - Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
 
hashCode() - Method in class org.apache.hadoop.mapreduce.Counter
 
hashCode() - Method in class org.apache.hadoop.mapreduce.CounterGroup
 
hashCode() - Method in class org.apache.hadoop.mapreduce.Counters
 
hashCode() - Method in class org.apache.hadoop.mapreduce.ID
 
hashCode() - Method in class org.apache.hadoop.mapreduce.JobID
 
hashCode() - Method in class org.apache.hadoop.mapreduce.lib.join.TupleWritable
 
hashCode() - Method in class org.apache.hadoop.mapreduce.lib.join.WrappedRecordReader
 
hashCode(byte[], int, int, int) - Method in class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedPartitioner
 
hashCode() - Method in class org.apache.hadoop.mapreduce.TaskAttemptID
 
hashCode() - Method in class org.apache.hadoop.mapreduce.TaskCompletionEvent
 
hashCode() - Method in class org.apache.hadoop.mapreduce.TaskID
 
HashingDistributionPolicy - Class in org.apache.hadoop.contrib.index.example
Choose a shard for each insert or delete based on document id hashing.
HashingDistributionPolicy() - Constructor for class org.apache.hadoop.contrib.index.example.HashingDistributionPolicy
 
HashPartitioner<K2,V2> - Class in org.apache.hadoop.mapred.lib
Partition keys by their Object.hashCode().
HashPartitioner() - Constructor for class org.apache.hadoop.mapred.lib.HashPartitioner
 
HashPartitioner<K,V> - Class in org.apache.hadoop.mapreduce.lib.partition
Partition keys by their Object.hashCode().
HashPartitioner() - Constructor for class org.apache.hadoop.mapreduce.lib.partition.HashPartitioner
 
hasNext() - Method in class org.apache.hadoop.contrib.utils.join.ArrayListBackedIterator
 
hasNext() - Method in interface org.apache.hadoop.mapred.join.ComposableRecordReader
Returns true if the stream is not empty, but provides no guarantee that a call to next(K,V) will succeed.
hasNext() - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
Return true if it is possible that this could emit more values.
hasNext() - Method in class org.apache.hadoop.mapred.join.JoinRecordReader.JoinDelegationIterator
 
hasNext() - Method in class org.apache.hadoop.mapred.join.MultiFilterRecordReader.MultiFilterDelegationIterator
 
hasNext() - Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
Return true if the RR- including the k,v pair stored in this object- is exhausted.
hasNext() - Method in class org.apache.hadoop.mapreduce.lib.join.ArrayListBackedIterator
 
hasNext() - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
Return true if it is possible that this could emit more values.
hasNext() - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader.JoinCollector
Returns false if exhausted or if reset(K) has not been called.
hasNext() - Method in class org.apache.hadoop.mapreduce.lib.join.JoinRecordReader.JoinDelegationIterator
 
hasNext() - Method in class org.apache.hadoop.mapreduce.lib.join.MultiFilterRecordReader.MultiFilterDelegationIterator
 
hasNext() - Method in class org.apache.hadoop.mapreduce.lib.join.ResetableIterator.EMPTY
 
hasNext() - Method in interface org.apache.hadoop.mapreduce.lib.join.ResetableIterator
True if a call to next may return a value.
hasNext() - Method in class org.apache.hadoop.mapreduce.lib.join.StreamBackedIterator
 
hasNext() - Method in class org.apache.hadoop.mapreduce.lib.join.WrappedRecordReader
Return true if the RR- including the k,v pair stored in this object- is exhausted.
hasNext() - Method in class org.apache.hadoop.mapreduce.MarkableIterator
 
hasNext() - Method in class org.apache.hadoop.mapreduce.task.ReduceContextImpl.ValueIterator
 
hasSimpleInputSpecs_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
HEIGHT - Static variable in class org.apache.hadoop.examples.dancing.Pentomino
 
height - Variable in class org.apache.hadoop.examples.dancing.Pentomino
 
HistoryViewer.AnalyzedJob - Class in org.apache.hadoop.mapreduce.jobhistory
Utility class used while analyzing the job.
HistoryViewer.AnalyzedJob(JobHistoryParser.JobInfo) - Constructor for class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.AnalyzedJob
Generate analysis information for the parsed job
HistoryViewer.FilteredJob - Class in org.apache.hadoop.mapreduce.jobhistory
Utility to filter out events based on the task status
HistoryViewer.FilteredJob(JobHistoryParser.JobInfo, String) - Constructor for class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.FilteredJob
Apply the filter (status) on the parsed job and generate summary
HistoryViewer.SummarizedJob - Class in org.apache.hadoop.mapreduce.jobhistory
Utility class used the summarize the job.
HistoryViewer.SummarizedJob(JobHistoryParser.JobInfo) - Constructor for class org.apache.hadoop.mapreduce.jobhistory.HistoryViewer.SummarizedJob
Create summary information for the parsed job

I

ID - Class in org.apache.hadoop.mapred
A general identifier, which internally stores the id as an integer.
ID(int) - Constructor for class org.apache.hadoop.mapred.ID
constructs an ID object from the given int
ID() - Constructor for class org.apache.hadoop.mapred.ID
 
id() - Method in interface org.apache.hadoop.mapred.join.ComposableRecordReader
Return the position in the collector this class occupies.
id() - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
Return the position in the collector this class occupies.
id - Variable in class org.apache.hadoop.mapred.join.Parser.Node
 
id() - Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
Return the position in the collector this class occupies.
ID - Class in org.apache.hadoop.mapreduce
A general identifier, which internally stores the id as an integer.
ID(int) - Constructor for class org.apache.hadoop.mapreduce.ID
constructs an ID object from the given int
ID() - Constructor for class org.apache.hadoop.mapreduce.ID
 
id - Variable in class org.apache.hadoop.mapreduce.ID
 
id() - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
Return the position in the collector this class occupies.
id - Variable in class org.apache.hadoop.mapreduce.lib.join.Parser.Node
 
id() - Method in class org.apache.hadoop.mapreduce.lib.join.WrappedRecordReader
Return the position in the collector this class occupies.
ident - Variable in class org.apache.hadoop.mapred.join.Parser.Node
 
ident - Variable in class org.apache.hadoop.mapreduce.lib.join.Parser.Node
 
IdentifierResolver - Class in org.apache.hadoop.streaming.io
This class is used to resolve a string identifier into the required IO classes.
IdentifierResolver() - Constructor for class org.apache.hadoop.streaming.io.IdentifierResolver
 
IdentityLocalAnalysis - Class in org.apache.hadoop.contrib.index.example
Identity local analysis maps inputs directly into outputs.
IdentityLocalAnalysis() - Constructor for class org.apache.hadoop.contrib.index.example.IdentityLocalAnalysis
 
IdentityMapper<K,V> - Class in org.apache.hadoop.mapred.lib
Implements the identity function, mapping inputs directly to outputs.
IdentityMapper() - Constructor for class org.apache.hadoop.mapred.lib.IdentityMapper
 
IdentityReducer<K,V> - Class in org.apache.hadoop.mapred.lib
Performs no reduction, writing all input values directly to the output.
IdentityReducer() - Constructor for class org.apache.hadoop.mapred.lib.IdentityReducer
 
idFormat - Static variable in class org.apache.hadoop.mapreduce.JobID
 
idFormat - Static variable in class org.apache.hadoop.mapreduce.TaskID
 
IDistributionPolicy - Interface in org.apache.hadoop.contrib.index.mapred
A distribution policy decides, given a document with a document id, which one shard the request should be sent to if the request is an insert, and which shard(s) the request should be sent to if the request is a delete.
idWithinJob() - Method in class org.apache.hadoop.mapreduce.TaskCompletionEvent
 
idx - Variable in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
 
idx - Variable in class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader
 
ifmt(double) - Static method in class org.apache.hadoop.streaming.StreamUtil
 
IIndexUpdater - Interface in org.apache.hadoop.contrib.index.mapred
A class implements an index updater interface should create a Map/Reduce job configuration and run the Map/Reduce job to analyze documents and update Lucene instances in parallel.
ILocalAnalysis<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> - Interface in org.apache.hadoop.contrib.index.mapred
Application specific local analysis.
incrAllCounters(Counters) - Method in class org.apache.hadoop.mapred.Counters
Increments multiple counters by their amounts in another Counters instance.
incrAllCounters(CounterGroup) - Method in class org.apache.hadoop.mapreduce.CounterGroup
 
incrAllCounters(Counters) - Method in class org.apache.hadoop.mapreduce.Counters
Increments multiple counters by their amounts in another Counters instance.
incrCounter(Enum, long) - Method in class org.apache.hadoop.mapred.Counters
Increments the specified counter by the specified amount, creating it if it didn't already exist.
incrCounter(String, String, long) - Method in class org.apache.hadoop.mapred.Counters
Increments the specified counter by the specified amount, creating it if it didn't already exist.
incrCounter(Enum<?>, long) - Method in interface org.apache.hadoop.mapred.Reporter
Increments the counter identified by the key, which can be of any Enum type, by the specified amount.
incrCounter(String, String, long) - Method in interface org.apache.hadoop.mapred.Reporter
Increments the counter identified by the group and counter name by the specified amount.
increment(long) - Method in class org.apache.hadoop.mapreduce.Counter
Increment this counter by the given value
IndexUpdateCombiner - Class in org.apache.hadoop.contrib.index.mapred
This combiner combines multiple intermediate forms into one intermediate form.
IndexUpdateCombiner() - Constructor for class org.apache.hadoop.contrib.index.mapred.IndexUpdateCombiner
 
IndexUpdateConfiguration - Class in org.apache.hadoop.contrib.index.mapred
This class provides the getters and the setters to a number of parameters.
IndexUpdateConfiguration(Configuration) - Constructor for class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Constructor
IndexUpdateMapper<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> - Class in org.apache.hadoop.contrib.index.mapred
This class applies local analysis on a key-value pair and then convert the result docid-operation pair to a shard-and-intermediate form pair.
IndexUpdateMapper() - Constructor for class org.apache.hadoop.contrib.index.mapred.IndexUpdateMapper
 
IndexUpdateOutputFormat - Class in org.apache.hadoop.contrib.index.mapred
The record writer of this output format simply puts a message in an output path when a shard update is done.
IndexUpdateOutputFormat() - Constructor for class org.apache.hadoop.contrib.index.mapred.IndexUpdateOutputFormat
 
IndexUpdatePartitioner - Class in org.apache.hadoop.contrib.index.mapred
This partitioner class puts the values of the same key - in this case the same shard - in the same partition.
IndexUpdatePartitioner() - Constructor for class org.apache.hadoop.contrib.index.mapred.IndexUpdatePartitioner
 
IndexUpdater - Class in org.apache.hadoop.contrib.index.mapred
An implementation of an index updater interface which creates a Map/Reduce job configuration and run the Map/Reduce job to analyze documents and update Lucene instances in parallel.
IndexUpdater() - Constructor for class org.apache.hadoop.contrib.index.mapred.IndexUpdater
 
IndexUpdateReducer - Class in org.apache.hadoop.contrib.index.mapred
This reducer applies to a shard the changes for it.
IndexUpdateReducer() - Constructor for class org.apache.hadoop.contrib.index.mapred.IndexUpdateReducer
 
init(Shard[]) - Method in class org.apache.hadoop.contrib.index.example.HashingDistributionPolicy
 
init(Shard[]) - Method in class org.apache.hadoop.contrib.index.example.RoundRobinDistributionPolicy
 
init(Shard[]) - Method in interface org.apache.hadoop.contrib.index.mapred.IDistributionPolicy
Initialization.
init(Job) - Method in class org.apache.hadoop.examples.pi.DistSum.MapSide
Initialize a job
init(Job) - Method in class org.apache.hadoop.examples.pi.DistSum.MixMachine
Initialize a job
init(Job) - Method in class org.apache.hadoop.examples.pi.DistSum.ReduceSide
Initialize a job
init(JobConf) - Method in class org.apache.hadoop.mapred.JobClient
Connect to the default JobTracker.
init() - Method in class org.apache.hadoop.streaming.StreamJob
 
init() - Method in class org.apache.hadoop.streaming.StreamXmlRecordReader
 
initialize(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.examples.MultiFileWordCount.CombineFileLineRecordReader
 
initialize(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
 
initialize(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader
 
initialize(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.input.KeyValueLineRecordReader
 
initialize(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
 
initialize(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsTextRecordReader
 
initialize(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileRecordReader
 
initialize(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
 
initialize(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.join.MultiFilterRecordReader
 
initialize(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.join.WrappedRecordReader
 
initialize(InputSplit, TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.RecordReader
Called once at initialization.
initialize(PipeMapRed) - Method in class org.apache.hadoop.streaming.io.InputWriter
Initializes the InputWriter.
initialize(PipeMapRed) - Method in class org.apache.hadoop.streaming.io.OutputReader
Initializes the OutputReader.
initialize(PipeMapRed) - Method in class org.apache.hadoop.streaming.io.RawBytesInputWriter
 
initialize(PipeMapRed) - Method in class org.apache.hadoop.streaming.io.RawBytesOutputReader
 
initialize(PipeMapRed) - Method in class org.apache.hadoop.streaming.io.TextInputWriter
 
initialize(PipeMapRed) - Method in class org.apache.hadoop.streaming.io.TextOutputReader
 
initialize(PipeMapRed) - Method in class org.apache.hadoop.streaming.io.TypedBytesInputWriter
 
initialize(PipeMapRed) - Method in class org.apache.hadoop.streaming.io.TypedBytesOutputReader
 
initializePieces() - Method in class org.apache.hadoop.examples.dancing.OneSidedPentomino
Define the one sided pieces.
initializePieces() - Method in class org.apache.hadoop.examples.dancing.Pentomino
Fill in the pieces list.
initNextRecordReader() - Method in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
Get the record reader for the next chunk in this CombineFileSplit.
initNextRecordReader() - Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader
Get the record reader for the next chunk in this CombineFileSplit.
InnerJoinRecordReader<K extends org.apache.hadoop.io.WritableComparable> - Class in org.apache.hadoop.mapred.join
Full inner join.
InnerJoinRecordReader<K extends org.apache.hadoop.io.WritableComparable<?>> - Class in org.apache.hadoop.mapreduce.lib.join
Full inner join.
INPUT_BOUNDING_QUERY - Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
Input query to get the max and min values of the jdbc.input.query
INPUT_CLASS_PROPERTY - Static variable in class org.apache.hadoop.mapred.lib.db.DBConfiguration
Class name implementing DBWritable which will hold input tuples
INPUT_CLASS_PROPERTY - Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
Class name implementing DBWritable which will hold input tuples
INPUT_CONDITIONS_PROPERTY - Static variable in class org.apache.hadoop.mapred.lib.db.DBConfiguration
WHERE clause in the input SELECT statement
INPUT_CONDITIONS_PROPERTY - Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
WHERE clause in the input SELECT statement
INPUT_COUNT_QUERY - Static variable in class org.apache.hadoop.mapred.lib.db.DBConfiguration
Input query to get the count of records
INPUT_COUNT_QUERY - Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
Input query to get the count of records
INPUT_DIR - Static variable in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
 
INPUT_FIELD_NAMES_PROPERTY - Static variable in class org.apache.hadoop.mapred.lib.db.DBConfiguration
Field names in the Input table
INPUT_FIELD_NAMES_PROPERTY - Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
Field names in the Input table
INPUT_FORMAT - Static variable in class org.apache.hadoop.mapred.pipes.Submitter
 
INPUT_ORDER_BY_PROPERTY - Static variable in class org.apache.hadoop.mapred.lib.db.DBConfiguration
ORDER BY clause in the input SELECT statement
INPUT_ORDER_BY_PROPERTY - Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
ORDER BY clause in the input SELECT statement
INPUT_QUERY - Static variable in class org.apache.hadoop.mapred.lib.db.DBConfiguration
Whole input query, exluding LIMIT...OFFSET
INPUT_QUERY - Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
Whole input query, exluding LIMIT...OFFSET
INPUT_TABLE_NAME_PROPERTY - Static variable in class org.apache.hadoop.mapred.lib.db.DBConfiguration
Input table name
INPUT_TABLE_NAME_PROPERTY - Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
Input table name
inputFile - Variable in class org.apache.hadoop.contrib.utils.join.DataJoinMapperBase
 
inputFile - Variable in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorBaseDescriptor
 
InputFormat<K,V> - Interface in org.apache.hadoop.mapred
InputFormat describes the input-specification for a Map-Reduce job.
InputFormat<K,V> - Class in org.apache.hadoop.mapreduce
InputFormat describes the input-specification for a Map-Reduce job.
InputFormat() - Constructor for class org.apache.hadoop.mapreduce.InputFormat
 
inputFormatSpec_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
InputSampler<K,V> - Class in org.apache.hadoop.mapred.lib
 
InputSampler(JobConf) - Constructor for class org.apache.hadoop.mapred.lib.InputSampler
 
InputSampler<K,V> - Class in org.apache.hadoop.mapreduce.lib.partition
Utility for collecting samples and writing a partition file for TotalOrderPartitioner.
InputSampler(Configuration) - Constructor for class org.apache.hadoop.mapreduce.lib.partition.InputSampler
 
InputSampler.IntervalSampler<K,V> - Class in org.apache.hadoop.mapreduce.lib.partition
Sample from s splits at regular intervals.
InputSampler.IntervalSampler(double) - Constructor for class org.apache.hadoop.mapreduce.lib.partition.InputSampler.IntervalSampler
Create a new IntervalSampler sampling all splits.
InputSampler.IntervalSampler(double, int) - Constructor for class org.apache.hadoop.mapreduce.lib.partition.InputSampler.IntervalSampler
Create a new IntervalSampler.
InputSampler.RandomSampler<K,V> - Class in org.apache.hadoop.mapreduce.lib.partition
Sample from random points in the input.
InputSampler.RandomSampler(double, int) - Constructor for class org.apache.hadoop.mapreduce.lib.partition.InputSampler.RandomSampler
Create a new RandomSampler sampling all splits.
InputSampler.RandomSampler(double, int, int) - Constructor for class org.apache.hadoop.mapreduce.lib.partition.InputSampler.RandomSampler
Create a new RandomSampler.
InputSampler.Sampler<K,V> - Interface in org.apache.hadoop.mapreduce.lib.partition
Interface to sample using an InputFormat.
InputSampler.SplitSampler<K,V> - Class in org.apache.hadoop.mapreduce.lib.partition
Samples the first n records from s splits.
InputSampler.SplitSampler(int) - Constructor for class org.apache.hadoop.mapreduce.lib.partition.InputSampler.SplitSampler
Create a SplitSampler sampling all splits.
InputSampler.SplitSampler(int, int) - Constructor for class org.apache.hadoop.mapreduce.lib.partition.InputSampler.SplitSampler
Create a new SplitSampler.
inputSpecs_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
InputSplit - Interface in org.apache.hadoop.mapred
InputSplit represents the data to be processed by an individual Mapper.
InputSplit - Class in org.apache.hadoop.mapreduce
InputSplit represents the data to be processed by an individual Mapper.
InputSplit() - Constructor for class org.apache.hadoop.mapreduce.InputSplit
 
inputTag - Variable in class org.apache.hadoop.contrib.utils.join.DataJoinMapperBase
 
InputWriter<K,V> - Class in org.apache.hadoop.streaming.io
Abstract base for classes that write the client's input.
InputWriter() - Constructor for class org.apache.hadoop.streaming.io.InputWriter
 
inReaderSpec_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
INSERT - Static variable in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp.Op
 
IntegerSplitter - Class in org.apache.hadoop.mapreduce.lib.db
Implement DBSplitter over integer values.
IntegerSplitter() - Constructor for class org.apache.hadoop.mapreduce.lib.db.IntegerSplitter
 
IntermediateForm - Class in org.apache.hadoop.contrib.index.mapred
An intermediate form for one or more parsed Lucene documents and/or delete terms.
IntermediateForm() - Constructor for class org.apache.hadoop.contrib.index.mapred.IntermediateForm
Constructor
INTERPRETOR - Static variable in class org.apache.hadoop.mapred.pipes.Submitter
 
IntSumReducer<Key> - Class in org.apache.hadoop.mapreduce.lib.reduce
 
IntSumReducer() - Constructor for class org.apache.hadoop.mapreduce.lib.reduce.IntSumReducer
 
InvalidFileTypeException - Exception in org.apache.hadoop.mapred
Used when file type differs from the desired file type.
InvalidFileTypeException() - Constructor for exception org.apache.hadoop.mapred.InvalidFileTypeException
 
InvalidFileTypeException(String) - Constructor for exception org.apache.hadoop.mapred.InvalidFileTypeException
 
InvalidInputException - Exception in org.apache.hadoop.mapred
This class wraps a list of problems with the input, so that the user can get a list of problems together instead of finding and fixing them one by one.
InvalidInputException(List<IOException>) - Constructor for exception org.apache.hadoop.mapred.InvalidInputException
Create the exception with the given list.
InvalidInputException - Exception in org.apache.hadoop.mapreduce.lib.input
This class wraps a list of problems with the input, so that the user can get a list of problems together instead of finding and fixing them one by one.
InvalidInputException(List<IOException>) - Constructor for exception org.apache.hadoop.mapreduce.lib.input.InvalidInputException
Create the exception with the given list.
InvalidJobConfException - Exception in org.apache.hadoop.mapred
This exception is thrown when jobconf misses some mendatory attributes or value of some attributes is invalid.
InvalidJobConfException() - Constructor for exception org.apache.hadoop.mapred.InvalidJobConfException
 
InvalidJobConfException(String) - Constructor for exception org.apache.hadoop.mapred.InvalidJobConfException
 
InverseMapper<K,V> - Class in org.apache.hadoop.mapred.lib
A Mapper that swaps keys and values.
InverseMapper() - Constructor for class org.apache.hadoop.mapred.lib.InverseMapper
 
InverseMapper<K,V> - Class in org.apache.hadoop.mapreduce.lib.map
A Mapper that swaps keys and values.
InverseMapper() - Constructor for class org.apache.hadoop.mapreduce.lib.map.InverseMapper
 
ioSpec_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
IS_JAVA_MAP - Static variable in class org.apache.hadoop.mapred.pipes.Submitter
 
IS_JAVA_REDUCE - Static variable in class org.apache.hadoop.mapred.pipes.Submitter
 
IS_JAVA_RR - Static variable in class org.apache.hadoop.mapred.pipes.Submitter
 
IS_JAVA_RW - Static variable in class org.apache.hadoop.mapred.pipes.Submitter
 
isBlacklisted() - Method in class org.apache.hadoop.mapreduce.TaskTrackerInfo
Whether tracker is blacklisted
isComplete() - Method in interface org.apache.hadoop.mapred.RunningJob
Check if the job is finished or not.
isComplete() - Method in class org.apache.hadoop.mapreduce.Job
Check if the job is finished or not.
isCompleted() - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
 
isCygwin() - Static method in class org.apache.hadoop.streaming.StreamUtil
 
isJobComplete() - Method in class org.apache.hadoop.mapreduce.JobStatus
Returns true if the status is for a completed job.
isLocalHadoop() - Method in class org.apache.hadoop.streaming.StreamJob
 
isLocalJobTracker(JobConf) - Static method in class org.apache.hadoop.streaming.StreamUtil
 
isMapTask() - Method in class org.apache.hadoop.mapreduce.TaskCompletionEvent
 
isMultiNamedOutput(JobConf, String) - Static method in class org.apache.hadoop.mapred.lib.MultipleOutputs
Returns if a named output is multiple.
IsolationRunner - Class in org.apache.hadoop.mapred
IsolationRunner is intended to facilitate debugging by re-running a specific task, given left-over task files for a (typically failed) past job.
IsolationRunner() - Constructor for class org.apache.hadoop.mapred.IsolationRunner
 
isReady() - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
 
isRetired() - Method in interface org.apache.hadoop.mapred.RunningJob
Check whether the job has been removed from JobTracker memory and retired.
isRetired() - Method in class org.apache.hadoop.mapreduce.Job
 
isRetired() - Method in class org.apache.hadoop.mapreduce.JobStatus
Check whether the job has retired.
isSegmentsFile(String) - Static method in class org.apache.hadoop.contrib.index.lucene.LuceneUtil
Check if the file is a segments_N file
isSegmentsGenFile(String) - Static method in class org.apache.hadoop.contrib.index.lucene.LuceneUtil
Check if the file is the segments.gen file
isSplitable(FileSystem, Path) - Method in class org.apache.hadoop.mapred.FileInputFormat
Is the given filename splitable? Usually, true, but if the file is stream compressed, it will not be.
isSplitable(FileSystem, Path) - Method in class org.apache.hadoop.mapred.KeyValueTextInputFormat
 
isSplitable(FileSystem, Path) - Method in class org.apache.hadoop.mapred.TextInputFormat
 
isSplitable(JobContext, Path) - Method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
Is the given filename splitable? Usually, true, but if the file is stream compressed, it will not be.
isSplitable(JobContext, Path) - Method in class org.apache.hadoop.mapreduce.lib.input.KeyValueTextInputFormat
 
isSplitable(JobContext, Path) - Method in class org.apache.hadoop.mapreduce.lib.input.TextInputFormat
 
isSuccessful() - Method in interface org.apache.hadoop.mapred.RunningJob
Check if the job completed successfully.
isSuccessful() - Method in class org.apache.hadoop.mapreduce.Job
Check if the job completed successfully.
iterator() - Method in class org.apache.hadoop.examples.pi.math.Bellard.Sum
iterator() - Method in class org.apache.hadoop.mapred.Counters.Group
 
iterator() - Method in class org.apache.hadoop.mapred.Counters
 
iterator() - Method in class org.apache.hadoop.mapreduce.CounterGroup
 
iterator() - Method in class org.apache.hadoop.mapreduce.Counters
 
iterator() - Method in class org.apache.hadoop.mapreduce.lib.join.TupleWritable
Return an iterator over the elements in this tuple.
iterator() - Method in class org.apache.hadoop.mapreduce.task.ReduceContextImpl.ValueIterable
 

J

jar_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
JarBuilder - Class in org.apache.hadoop.streaming
This class is the main class for generating job.jar for Hadoop Streaming jobs.
JarBuilder() - Constructor for class org.apache.hadoop.streaming.JarBuilder
 
jc - Variable in class org.apache.hadoop.mapred.join.CompositeRecordReader
 
jc - Variable in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
 
jc - Variable in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
 
jc_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
job - Variable in class org.apache.hadoop.contrib.utils.join.DataJoinMapperBase
 
job - Variable in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
 
Job - Class in org.apache.hadoop.mapred.jobcontrol
 
Job(JobConf, ArrayList<?>) - Constructor for class org.apache.hadoop.mapred.jobcontrol.Job
Construct a job.
Job(JobConf) - Constructor for class org.apache.hadoop.mapred.jobcontrol.Job
 
Job - Class in org.apache.hadoop.mapreduce
The job submitter's view of the Job.
Job() - Constructor for class org.apache.hadoop.mapreduce.Job
Deprecated. 
Job(Configuration) - Constructor for class org.apache.hadoop.mapreduce.Job
Deprecated. 
Job(Configuration, String) - Constructor for class org.apache.hadoop.mapreduce.Job
Deprecated. 
JOB - Static variable in class org.apache.hadoop.mapreduce.JobID
 
Job.JobState - Enum in org.apache.hadoop.mapreduce
 
Job.TaskStatusFilter - Enum in org.apache.hadoop.mapreduce
 
JOB_TOKEN_HDFS_FILE - Static variable in class org.apache.hadoop.mapreduce.security.TokenCache
file name used on HDFS for generated job token
JOB_TOKENS_FILENAME - Static variable in class org.apache.hadoop.mapreduce.security.TokenCache
conf setting for job tokens cache file name
JobBase - Class in org.apache.hadoop.contrib.utils.join
A common base implementing some statics collecting mechanisms that are commonly used in a typical map/reduce job.
JobBase() - Constructor for class org.apache.hadoop.contrib.utils.join.JobBase
 
JobClient - Class in org.apache.hadoop.mapred
JobClient is the primary interface for the user-job to interact with the JobTracker.
JobClient() - Constructor for class org.apache.hadoop.mapred.JobClient
Create a job client.
JobClient(JobConf) - Constructor for class org.apache.hadoop.mapred.JobClient
Build a job client with the given JobConf, and connect to the default JobTracker.
JobClient(Configuration) - Constructor for class org.apache.hadoop.mapred.JobClient
Build a job client with the given Configuration, and connect to the default JobTracker.
JobClient(InetSocketAddress, Configuration) - Constructor for class org.apache.hadoop.mapred.JobClient
Build a job client, connect to the indicated job tracker.
JobClient.TaskStatusFilter - Enum in org.apache.hadoop.mapred
 
JobConf - Class in org.apache.hadoop.mapred
A map/reduce job configuration.
JobConf() - Constructor for class org.apache.hadoop.mapred.JobConf
Construct a map/reduce job configuration.
JobConf(Class) - Constructor for class org.apache.hadoop.mapred.JobConf
Construct a map/reduce job configuration.
JobConf(Configuration) - Constructor for class org.apache.hadoop.mapred.JobConf
Construct a map/reduce job configuration.
JobConf(Configuration, Class) - Constructor for class org.apache.hadoop.mapred.JobConf
Construct a map/reduce job configuration.
JobConf(String) - Constructor for class org.apache.hadoop.mapred.JobConf
Construct a map/reduce configuration.
JobConf(Path) - Constructor for class org.apache.hadoop.mapred.JobConf
Construct a map/reduce configuration.
JobConf(boolean) - Constructor for class org.apache.hadoop.mapred.JobConf
A new map/reduce configuration where the behavior of reading from the default resources can be turned off.
jobConf_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
JobConfigurable - Interface in org.apache.hadoop.mapred
That what may be configured.
JobContext - Interface in org.apache.hadoop.mapred
 
JobContext - Interface in org.apache.hadoop.mapreduce
A read-only view of the job that is provided to the tasks while they are running.
JobControl - Class in org.apache.hadoop.mapred.jobcontrol
 
JobControl(String) - Constructor for class org.apache.hadoop.mapred.jobcontrol.JobControl
Construct a job control for a group of jobs.
JobControl - Class in org.apache.hadoop.mapreduce.lib.jobcontrol
This class encapsulates a set of MapReduce jobs and its dependency.
JobControl(String) - Constructor for class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl
Construct a job control for a group of jobs.
JobControl.ThreadState - Enum in org.apache.hadoop.mapreduce.lib.jobcontrol
 
JobCounter - Enum in org.apache.hadoop.mapreduce
 
JobHistoryParser.JobInfo - Class in org.apache.hadoop.mapreduce.jobhistory
The class where job information is aggregated into after parsing
JobHistoryParser.JobInfo() - Constructor for class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.JobInfo
Create a job info object where job information will be stored after a parse
JobHistoryParser.TaskAttemptInfo - Class in org.apache.hadoop.mapreduce.jobhistory
Task Attempt Information is aggregated in this class after parsing
JobHistoryParser.TaskAttemptInfo() - Constructor for class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskAttemptInfo
Create a Task Attempt Info which will store attempt level information on a history parse.
JobHistoryParser.TaskInfo - Class in org.apache.hadoop.mapreduce.jobhistory
TaskInformation is aggregated in this class after parsing
JobHistoryParser.TaskInfo() - Constructor for class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskInfo
 
JobID - Class in org.apache.hadoop.mapred
JobID represents the immutable and unique identifier for the job.
JobID(String, int) - Constructor for class org.apache.hadoop.mapred.JobID
Constructs a JobID object
JobID() - Constructor for class org.apache.hadoop.mapred.JobID
 
JobID - Class in org.apache.hadoop.mapreduce
JobID represents the immutable and unique identifier for the job.
JobID(String, int) - Constructor for class org.apache.hadoop.mapreduce.JobID
Constructs a JobID object
JobID() - Constructor for class org.apache.hadoop.mapreduce.JobID
 
jobId_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
jobInfo() - Method in class org.apache.hadoop.streaming.StreamJob
 
JobPriority - Enum in org.apache.hadoop.mapred
Used to describe the priority of the running job.
JobPriority - Enum in org.apache.hadoop.mapreduce
Used to describe the priority of the running job.
JobQueueInfo - Class in org.apache.hadoop.mapred
Class that contains the information regarding the Job Queues which are maintained by the Hadoop Map/Reduce framework.
JobQueueInfo() - Constructor for class org.apache.hadoop.mapred.JobQueueInfo
Default constructor for Job Queue Info.
JobQueueInfo(String, String) - Constructor for class org.apache.hadoop.mapred.JobQueueInfo
Construct a new JobQueueInfo object using the queue name and the scheduling information passed.
JobSplit.SplitMetaInfo - Class in org.apache.hadoop.mapreduce.split
This represents the meta information about the task split.
JobSplit.SplitMetaInfo() - Constructor for class org.apache.hadoop.mapreduce.split.JobSplit.SplitMetaInfo
 
JobSplit.SplitMetaInfo(String[], long, long) - Constructor for class org.apache.hadoop.mapreduce.split.JobSplit.SplitMetaInfo
 
JobSplit.SplitMetaInfo(InputSplit, long) - Constructor for class org.apache.hadoop.mapreduce.split.JobSplit.SplitMetaInfo
 
JobSplit.TaskSplitIndex - Class in org.apache.hadoop.mapreduce.split
This represents the meta information about the task split that the task gets
JobSplit.TaskSplitIndex() - Constructor for class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitIndex
 
JobSplit.TaskSplitIndex(String, long) - Constructor for class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitIndex
 
JobSplit.TaskSplitMetaInfo - Class in org.apache.hadoop.mapreduce.split
This represents the meta information about the task split that the JobTracker creates
JobSplit.TaskSplitMetaInfo() - Constructor for class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitMetaInfo
 
JobSplit.TaskSplitMetaInfo(JobSplit.TaskSplitIndex, String[], long) - Constructor for class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitMetaInfo
 
JobSplit.TaskSplitMetaInfo(InputSplit, long) - Constructor for class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitMetaInfo
 
JobSplit.TaskSplitMetaInfo(String[], long, long) - Constructor for class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitMetaInfo
 
JobStatus - Class in org.apache.hadoop.mapred
Describes the current status of a job.
JobStatus() - Constructor for class org.apache.hadoop.mapred.JobStatus
 
JobStatus(JobID, float, float, float, int, String, String, String, String) - Constructor for class org.apache.hadoop.mapred.JobStatus
Create a job status object for a given jobid.
JobStatus(JobID, float, float, int, String, String, String, String) - Constructor for class org.apache.hadoop.mapred.JobStatus
Create a job status object for a given jobid.
JobStatus(JobID, float, float, float, int, JobPriority, String, String, String, String) - Constructor for class org.apache.hadoop.mapred.JobStatus
Create a job status object for a given jobid.
JobStatus(JobID, float, float, float, float, int, JobPriority, String, String, String, String) - Constructor for class org.apache.hadoop.mapred.JobStatus
Create a job status object for a given jobid.
JobStatus - Class in org.apache.hadoop.mapreduce
Describes the current status of a job.
JobStatus() - Constructor for class org.apache.hadoop.mapreduce.JobStatus
 
JobStatus(JobID, float, float, float, float, JobStatus.State, JobPriority, String, String, String, String) - Constructor for class org.apache.hadoop.mapreduce.JobStatus
Create a job status object for a given jobid.
JobStatus.State - Enum in org.apache.hadoop.mapreduce
Current state of the job
jobsToComplete() - Method in class org.apache.hadoop.mapred.JobClient
Get the jobs that are not completed and not failed.
jobSubmit(JobConf) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
Submit a job to the Map-Reduce framework.
Join - Class in org.apache.hadoop.examples
Given a set of sorted datasets keyed with the same class and yielding equal partitions, it is possible to effect a join of those datasets prior to the map.
Join() - Constructor for class org.apache.hadoop.examples.Join
 
join(long) - Method in class org.apache.hadoop.mapred.gridmix.Statistics
Wait until the service completes.
JOIN_COMPARATOR - Static variable in class org.apache.hadoop.mapreduce.lib.join.CompositeInputFormat
 
JOIN_EXPR - Static variable in class org.apache.hadoop.mapreduce.lib.join.CompositeInputFormat
 
JoinRecordReader<K extends org.apache.hadoop.io.WritableComparable> - Class in org.apache.hadoop.mapred.join
Base class for Composite joins returning Tuples of arbitrary Writables.
JoinRecordReader(int, JobConf, int, Class<? extends WritableComparator>) - Constructor for class org.apache.hadoop.mapred.join.JoinRecordReader
 
JoinRecordReader<K extends org.apache.hadoop.io.WritableComparable<?>> - Class in org.apache.hadoop.mapreduce.lib.join
Base class for Composite joins returning Tuples of arbitrary Writables.
JoinRecordReader(int, Configuration, int, Class<? extends WritableComparator>) - Constructor for class org.apache.hadoop.mapreduce.lib.join.JoinRecordReader
 
JoinRecordReader.JoinDelegationIterator - Class in org.apache.hadoop.mapred.join
Since the JoinCollector is effecting our operation, we need only provide an iterator proxy wrapping its operation.
JoinRecordReader.JoinDelegationIterator() - Constructor for class org.apache.hadoop.mapred.join.JoinRecordReader.JoinDelegationIterator
 
JoinRecordReader.JoinDelegationIterator - Class in org.apache.hadoop.mapreduce.lib.join
Since the JoinCollector is effecting our operation, we need only provide an iterator proxy wrapping its operation.
JoinRecordReader.JoinDelegationIterator() - Constructor for class org.apache.hadoop.mapreduce.lib.join.JoinRecordReader.JoinDelegationIterator
 

K

key() - Method in interface org.apache.hadoop.mapred.join.ComposableRecordReader
Return the key this RecordReader would supply on a call to next(K,V)
key(K) - Method in interface org.apache.hadoop.mapred.join.ComposableRecordReader
Clone the key at the head of this RecordReader into the object provided.
key() - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
Return the key for the current join or the value at the top of the RecordReader heap.
key(K) - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
Clone the key at the top of this RR into the given object.
key() - Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
Return the key at the head of this RR.
key(K) - Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
Clone the key at the head of this RR into the object supplied.
key() - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader.JoinCollector
Return the key associated with this collection.
key - Variable in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
 
key() - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
Return the key for the current join or the value at the top of the RecordReader heap.
key(K) - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
Clone the key at the top of this RR into the given object.
key() - Method in class org.apache.hadoop.mapreduce.lib.join.WrappedRecordReader
Return the key at the head of this RR.
key(K) - Method in class org.apache.hadoop.mapreduce.lib.join.WrappedRecordReader
Clone the key at the head of this RR into the object supplied.
KEY_CLASS - Static variable in class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat
 
KEY_VALUE_SEPERATOR - Static variable in class org.apache.hadoop.mapreduce.lib.input.KeyValueLineRecordReader
 
keyclass - Variable in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
 
KeyFieldBasedComparator<K,V> - Class in org.apache.hadoop.mapred.lib
This comparator implementation provides a subset of the features provided by the Unix/GNU Sort.
KeyFieldBasedComparator() - Constructor for class org.apache.hadoop.mapred.lib.KeyFieldBasedComparator
 
KeyFieldBasedComparator<K,V> - Class in org.apache.hadoop.mapreduce.lib.partition
This comparator implementation provides a subset of the features provided by the Unix/GNU Sort.
KeyFieldBasedComparator() - Constructor for class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedComparator
 
KeyFieldBasedPartitioner<K2,V2> - Class in org.apache.hadoop.mapred.lib
Defines a way to partition keys based on certain key fields (also see KeyFieldBasedComparator.
KeyFieldBasedPartitioner() - Constructor for class org.apache.hadoop.mapred.lib.KeyFieldBasedPartitioner
 
KeyFieldBasedPartitioner<K2,V2> - Class in org.apache.hadoop.mapreduce.lib.partition
Defines a way to partition keys based on certain key fields (also see KeyFieldBasedComparator.
KeyFieldBasedPartitioner() - Constructor for class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedPartitioner
 
KeyValueLineRecordReader - Class in org.apache.hadoop.mapred
This class treats a line in the input as a key/value pair separated by a separator character.
KeyValueLineRecordReader(Configuration, FileSplit) - Constructor for class org.apache.hadoop.mapred.KeyValueLineRecordReader
 
KeyValueLineRecordReader - Class in org.apache.hadoop.mapreduce.lib.input
This class treats a line in the input as a key/value pair separated by a separator character.
KeyValueLineRecordReader(Configuration) - Constructor for class org.apache.hadoop.mapreduce.lib.input.KeyValueLineRecordReader
 
KeyValueTextInputFormat - Class in org.apache.hadoop.mapred
An InputFormat for plain text files.
KeyValueTextInputFormat() - Constructor for class org.apache.hadoop.mapred.KeyValueTextInputFormat
 
KeyValueTextInputFormat - Class in org.apache.hadoop.mapreduce.lib.input
An InputFormat for plain text files.
KeyValueTextInputFormat() - Constructor for class org.apache.hadoop.mapreduce.lib.input.KeyValueTextInputFormat
 
kids - Variable in class org.apache.hadoop.mapred.join.CompositeRecordReader
 
kids - Variable in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
 
KILLED - Static variable in class org.apache.hadoop.mapred.JobStatus
 
killJob() - Method in interface org.apache.hadoop.mapred.RunningJob
Kill the running job.
killJob() - Method in class org.apache.hadoop.mapreduce.Job
Kill the running job.
killJob() - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
 
killTask(TaskAttemptID, boolean) - Method in interface org.apache.hadoop.mapred.RunningJob
Kill indicated task attempt.
killTask(String, boolean) - Method in interface org.apache.hadoop.mapred.RunningJob
Deprecated. Applications should rather use RunningJob.killTask(TaskAttemptID, boolean)
killTask(TaskAttemptID) - Method in class org.apache.hadoop.mapreduce.Job
Kill indicated task attempt.

L

largestNumOfValues - Variable in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
 
lazyOutput_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
LazyOutputFormat<K,V> - Class in org.apache.hadoop.mapred.lib
A Convenience class that creates output lazily.
LazyOutputFormat() - Constructor for class org.apache.hadoop.mapred.lib.LazyOutputFormat
 
LazyOutputFormat<K,V> - Class in org.apache.hadoop.mapreduce.lib.output
A Convenience class that creates output lazily.
LazyOutputFormat() - Constructor for class org.apache.hadoop.mapreduce.lib.output.LazyOutputFormat
 
LEFT_OFFSET_PROPERTY_NAME - Static variable in class org.apache.hadoop.mapreduce.lib.partition.BinaryPartitioner
 
limit - Variable in class org.apache.hadoop.examples.pi.math.ArithmeticProgression
Ending value
LineDocInputFormat - Class in org.apache.hadoop.contrib.index.example
An InputFormat for LineDoc for plain text files where each line is a doc.
LineDocInputFormat() - Constructor for class org.apache.hadoop.contrib.index.example.LineDocInputFormat
 
LineDocLocalAnalysis - Class in org.apache.hadoop.contrib.index.example
Convert LineDocTextAndOp to DocumentAndOp as required by ILocalAnalysis.
LineDocLocalAnalysis() - Constructor for class org.apache.hadoop.contrib.index.example.LineDocLocalAnalysis
 
LineDocRecordReader - Class in org.apache.hadoop.contrib.index.example
A simple RecordReader for LineDoc for plain text files where each line is a doc.
LineDocRecordReader(Configuration, FileSplit) - Constructor for class org.apache.hadoop.contrib.index.example.LineDocRecordReader
Constructor
LineDocTextAndOp - Class in org.apache.hadoop.contrib.index.example
This class represents an operation.
LineDocTextAndOp() - Constructor for class org.apache.hadoop.contrib.index.example.LineDocTextAndOp
Constructor
LineRecordReader.LineReader - Class in org.apache.hadoop.mapred
A class that provides a line reader from an input stream.
LineRecordReader.LineReader(InputStream, Configuration) - Constructor for class org.apache.hadoop.mapred.LineRecordReader.LineReader
 
LINES_PER_MAP - Static variable in class org.apache.hadoop.mapreduce.lib.input.NLineInputFormat
 
LINK_URI - Static variable in class org.apache.hadoop.streaming.StreamJob
 
list() - Method in class org.apache.hadoop.contrib.index.lucene.FileSystemDirectory
 
listJobConfProperties() - Method in class org.apache.hadoop.streaming.StreamJob
Prints out the jobconf properties on stdout when verbose is specified.
listStatus(JobConf) - Method in class org.apache.hadoop.mapred.FileInputFormat
List input directories.
listStatus(JobConf) - Method in class org.apache.hadoop.mapred.SequenceFileInputFormat
 
listStatus(JobContext) - Method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
List input directories.
listStatus(JobContext) - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFormat
 
ljustify(String, int) - Static method in class org.apache.hadoop.streaming.StreamUtil
 
loadTaskTokenStorage(String, JobConf) - Static method in class org.apache.hadoop.mapreduce.security.TokenCache
load token storage and stores it
loadTokens(String, JobConf) - Static method in class org.apache.hadoop.mapreduce.security.TokenCache
load job token from a file
LoadTypedBytes - Class in org.apache.hadoop.streaming
Utility program that reads typed bytes from standard input and stores them in a sequence file for which the path is given as an argument.
LoadTypedBytes(Configuration) - Constructor for class org.apache.hadoop.streaming.LoadTypedBytes
 
LoadTypedBytes() - Constructor for class org.apache.hadoop.streaming.LoadTypedBytes
 
localHadoop_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
localizeBin(String) - Static method in class org.apache.hadoop.streaming.StreamUtil
 
LOG - Static variable in class org.apache.hadoop.contrib.index.main.UpdateIndex
 
LOG - Static variable in class org.apache.hadoop.contrib.index.mapred.IndexUpdater
 
LOG - Static variable in class org.apache.hadoop.contrib.utils.join.JobBase
 
log(Log) - Method in class org.apache.hadoop.mapred.Counters
Logs the current counter values.
LOG - Static variable in class org.apache.hadoop.mapred.FileInputFormat
 
LOG - Static variable in class org.apache.hadoop.mapred.FileOutputCommitter
 
LOG - Static variable in class org.apache.hadoop.mapred.gridmix.Gridmix
 
LOG - Static variable in class org.apache.hadoop.mapred.gridmix.SerialJobFactory
 
LOG - Static variable in class org.apache.hadoop.mapred.gridmix.Statistics
 
LOG - Static variable in class org.apache.hadoop.mapred.gridmix.StressJobFactory
 
LOG - Static variable in class org.apache.hadoop.mapred.lib.FieldSelectionMapReduce
 
LOG - Static variable in class org.apache.hadoop.mapred.pipes.Submitter
 
LOG - Static variable in class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionMapper
 
LOG - Static variable in class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionReducer
 
LOG - Static variable in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter
 
LOG - Static variable in class org.apache.hadoop.streaming.PipeMapRed
 
LOG - Static variable in class org.apache.hadoop.streaming.StreamBaseRecordReader
 
LOG - Static variable in class org.apache.hadoop.streaming.StreamJob
 
logSpec() - Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJobBase
 
logSpec() - Static method in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorJobBase
 
long2string(long) - Static method in class org.apache.hadoop.examples.pi.Util
Covert a long to a String in comma separated number format.
LONG_VALUE_MAX - Static variable in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor
 
LONG_VALUE_MAX - Static variable in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorBaseDescriptor
 
LONG_VALUE_MIN - Static variable in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor
 
LONG_VALUE_MIN - Static variable in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorBaseDescriptor
 
LONG_VALUE_SUM - Static variable in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor
 
LONG_VALUE_SUM - Static variable in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorBaseDescriptor
 
LongSumReducer<K> - Class in org.apache.hadoop.mapred.lib
A Reducer that sums long values.
LongSumReducer() - Constructor for class org.apache.hadoop.mapred.lib.LongSumReducer
 
LongSumReducer<KEY> - Class in org.apache.hadoop.mapreduce.lib.reduce
 
LongSumReducer() - Constructor for class org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer
 
LongValueMax - Class in org.apache.hadoop.mapred.lib.aggregate
This class implements a value aggregator that maintain the maximum of a sequence of long values.
LongValueMax() - Constructor for class org.apache.hadoop.mapred.lib.aggregate.LongValueMax
 
LongValueMax - Class in org.apache.hadoop.mapreduce.lib.aggregate
This class implements a value aggregator that maintain the maximum of a sequence of long values.
LongValueMax() - Constructor for class org.apache.hadoop.mapreduce.lib.aggregate.LongValueMax
the default constructor
LongValueMin - Class in org.apache.hadoop.mapred.lib.aggregate
This class implements a value aggregator that maintain the minimum of a sequence of long values.
LongValueMin() - Constructor for class org.apache.hadoop.mapred.lib.aggregate.LongValueMin
 
LongValueMin - Class in org.apache.hadoop.mapreduce.lib.aggregate
This class implements a value aggregator that maintain the minimum of a sequence of long values.
LongValueMin() - Constructor for class org.apache.hadoop.mapreduce.lib.aggregate.LongValueMin
the default constructor
LongValueSum - Class in org.apache.hadoop.mapred.lib.aggregate
This class implements a value aggregator that sums up a sequence of long values.
LongValueSum() - Constructor for class org.apache.hadoop.mapred.lib.aggregate.LongValueSum
 
LongValueSum - Class in org.apache.hadoop.mapreduce.lib.aggregate
This class implements a value aggregator that sums up a sequence of long values.
LongValueSum() - Constructor for class org.apache.hadoop.mapreduce.lib.aggregate.LongValueSum
the default constructor
LuceneUtil - Class in org.apache.hadoop.contrib.index.lucene
This class copies some methods from Lucene's SegmentInfos since that class is not public.
LuceneUtil() - Constructor for class org.apache.hadoop.contrib.index.lucene.LuceneUtil
 

M

main(String[]) - Static method in class org.apache.hadoop.contrib.index.main.UpdateIndex
The main() method
main(String[]) - Static method in class org.apache.hadoop.contrib.utils.join.DataJoinJob
 
main(String[]) - Static method in class org.apache.hadoop.examples.AggregateWordCount
The main driver for word count map/reduce program.
main(String[]) - Static method in class org.apache.hadoop.examples.AggregateWordHistogram
The main driver for word count map/reduce program.
main(String[]) - Static method in class org.apache.hadoop.examples.BaileyBorweinPlouffe
The main method for running it as a stand alone command.
main(String[]) - Static method in class org.apache.hadoop.examples.dancing.DistributedPentomino
Launch the solver on 9x10 board and the one sided pentominos.
main(String[]) - Static method in class org.apache.hadoop.examples.dancing.OneSidedPentomino
Solve the 3x30 puzzle.
main(String[]) - Static method in class org.apache.hadoop.examples.dancing.Pentomino
Solve the 6x10 pentomino puzzle.
main(String[]) - Static method in class org.apache.hadoop.examples.dancing.Sudoku
Solves a set of sudoku puzzles.
main(String[]) - Static method in class org.apache.hadoop.examples.DBCountPageView
 
main(String[]) - Static method in class org.apache.hadoop.examples.ExampleDriver
 
main(String[]) - Static method in class org.apache.hadoop.examples.Grep
 
main(String[]) - Static method in class org.apache.hadoop.examples.Join
 
main(String[]) - Static method in class org.apache.hadoop.examples.MultiFileWordCount
 
main(String[]) - Static method in class org.apache.hadoop.examples.pi.DistBbp
main
main(String[]) - Static method in class org.apache.hadoop.examples.pi.DistSum
main
main(String[]) - Static method in class org.apache.hadoop.examples.pi.math.Bellard
main
main(String[]) - Static method in class org.apache.hadoop.examples.pi.Parser
main
main(String[]) - Static method in class org.apache.hadoop.examples.QuasiMonteCarlo
main method for running it as a stand alone command.
main(String[]) - Static method in class org.apache.hadoop.examples.RandomTextWriter
 
main(String[]) - Static method in class org.apache.hadoop.examples.RandomWriter
 
main(String[]) - Static method in class org.apache.hadoop.examples.SecondarySort
 
main(String[]) - Static method in class org.apache.hadoop.examples.Sort
 
main(String[]) - Static method in class org.apache.hadoop.examples.terasort.GenSort
 
main(String[]) - Static method in class org.apache.hadoop.examples.terasort.TeraChecksum
 
main(String[]) - Static method in class org.apache.hadoop.examples.terasort.TeraGen
 
main(String[]) - Static method in class org.apache.hadoop.examples.terasort.TeraSort
 
main(String[]) - Static method in class org.apache.hadoop.examples.terasort.TeraValidate
 
main(String[]) - Static method in class org.apache.hadoop.examples.WordCount
 
main(String[]) - Static method in class org.apache.hadoop.mapred.gridmix.Gridmix
 
main(String[]) - Static method in class org.apache.hadoop.mapred.IsolationRunner
Run a single task.
main(String[]) - Static method in class org.apache.hadoop.mapred.JobClient
 
main(String[]) - Static method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJob
create and run an Aggregate based map/reduce job.
main(String[]) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
Submit a pipes job based on the command line arguments.
main(String[]) - Static method in class org.apache.hadoop.mapred.tools.MRAdmin
 
main(String[]) - Static method in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorJob
create and run an Aggregate based map/reduce job.
main(String[]) - Static method in class org.apache.hadoop.mapreduce.lib.partition.InputSampler
 
main(String[]) - Static method in class org.apache.hadoop.mapreduce.tools.CLI
 
main(String[]) - Static method in class org.apache.hadoop.streaming.DumpTypedBytes
 
main(String[]) - Static method in class org.apache.hadoop.streaming.HadoopStreaming
 
main(String[]) - Static method in class org.apache.hadoop.streaming.JarBuilder
Test program
main(String[]) - Static method in class org.apache.hadoop.streaming.LoadTypedBytes
 
main(String[]) - Static method in class org.apache.hadoop.streaming.PathFinder
 
makeCompactString() - Method in class org.apache.hadoop.mapred.Counters
Convert a counters object into a single line that is easy to parse.
makeEscapedCompactString() - Method in class org.apache.hadoop.mapred.Counters.Counter
Returns the compact stringified version of the counter in the format [(actual-name)(display-name)(value)]
makeEscapedCompactString() - Method in class org.apache.hadoop.mapred.Counters.Group
Returns the compact stringified version of the group in the format {(actual-name)(display-name)(value)[][][]} where [] are compact strings for the counters within.
makeEscapedCompactString() - Method in class org.apache.hadoop.mapred.Counters
Represent the counter in a textual format that can be converted back to its object form
makeJavaCommand(Class, String[]) - Static method in class org.apache.hadoop.streaming.StreamUtil
 
makeLock(String) - Method in class org.apache.hadoop.contrib.index.lucene.FileSystemDirectory
 
makeSplit(Path, long, long, String[]) - Method in class org.apache.hadoop.examples.terasort.TeraInputFormat
 
makeSplit(Path, long, long, String[]) - Method in class org.apache.hadoop.mapred.FileInputFormat
A factory that makes the split for this class.
makeSplit(Path, long, long, String[]) - Method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
A factory that makes the split for this class.
map(DocumentID, DocumentAndOp, OutputCollector<DocumentID, DocumentAndOp>, Reporter) - Method in class org.apache.hadoop.contrib.index.example.IdentityLocalAnalysis
 
map(DocumentID, LineDocTextAndOp, OutputCollector<DocumentID, DocumentAndOp>, Reporter) - Method in class org.apache.hadoop.contrib.index.example.LineDocLocalAnalysis
 
map(K, V, OutputCollector<Shard, IntermediateForm>, Reporter) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateMapper
Map a key-value pair to a shard-and-intermediate form pair.
map(Object, Object, OutputCollector, Reporter) - Method in class org.apache.hadoop.contrib.utils.join.DataJoinMapperBase
 
map(Object, Object, OutputCollector, Reporter) - Method in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
 
map(LongWritable, IntWritable, Mapper<LongWritable, IntWritable, LongWritable, BytesWritable>.Context) - Method in class org.apache.hadoop.examples.BaileyBorweinPlouffe.BbpMapper
Compute the (offset+1)th to (offset+length)th digits.
map(WritableComparable<?>, Text, Mapper<WritableComparable<?>, Text, Text, Text>.Context) - Method in class org.apache.hadoop.examples.dancing.DistributedPentomino.PentMap
Break the prefix string into moves (a sequence of integer row ids that will be selected for each column in order).
map(MultiFileWordCount.WordOffset, Text, Mapper<MultiFileWordCount.WordOffset, Text, Text, IntWritable>.Context) - Method in class org.apache.hadoop.examples.MultiFileWordCount.MapClass
 
map(NullWritable, SummationWritable, Mapper<NullWritable, SummationWritable, NullWritable, TaskResult>.Context) - Method in class org.apache.hadoop.examples.pi.DistSum.MapSide.SummingMapper
 
map(NullWritable, SummationWritable, Mapper<NullWritable, SummationWritable, IntWritable, SummationWritable>.Context) - Method in class org.apache.hadoop.examples.pi.DistSum.ReduceSide.PartitionMapper
Partitions sigma into parts
map(LongWritable, LongWritable, Mapper<LongWritable, LongWritable, BooleanWritable, LongWritable>.Context) - Method in class org.apache.hadoop.examples.QuasiMonteCarlo.QmcMapper
Map method.
map(LongWritable, Text, Mapper<LongWritable, Text, SecondarySort.IntPair, IntWritable>.Context) - Method in class org.apache.hadoop.examples.SecondarySort.MapClass
 
map(LongWritable, NullWritable, Mapper<LongWritable, NullWritable, Text, Text>.Context) - Method in class org.apache.hadoop.examples.terasort.TeraGen.SortGenMapper
 
map(Object, Text, Mapper<Object, Text, Text, IntWritable>.Context) - Method in class org.apache.hadoop.examples.WordCount.TokenizerMapper
 
map(K1, V1, OutputCollector<Text, Text>, Reporter) - Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorCombiner
Do nothing.
map(K1, V1, OutputCollector<Text, Text>, Reporter) - Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorMapper
the map function.
map(K1, V1, OutputCollector<Text, Text>, Reporter) - Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorReducer
Do nothing.
map(Object, Object, OutputCollector, Reporter) - Method in class org.apache.hadoop.mapred.lib.ChainMapper
Chains the map(...) methods of the Mappers in the chain.
map(K, V, OutputCollector<Text, Text>, Reporter) - Method in class org.apache.hadoop.mapred.lib.FieldSelectionMapReduce
The identify function.
map(K, V, OutputCollector<K, V>, Reporter) - Method in class org.apache.hadoop.mapred.lib.IdentityMapper
The identify function.
map(K, V, OutputCollector<V, K>, Reporter) - Method in class org.apache.hadoop.mapred.lib.InverseMapper
The inverse function.
map(K, Text, OutputCollector<Text, LongWritable>, Reporter) - Method in class org.apache.hadoop.mapred.lib.RegexMapper
 
map(K, Text, OutputCollector<Text, LongWritable>, Reporter) - Method in class org.apache.hadoop.mapred.lib.TokenCountMapper
 
map(K1, V1, OutputCollector<K2, V2>, Reporter) - Method in interface org.apache.hadoop.mapred.Mapper
Maps a single input key/value pair into an intermediate key/value pair.
map(K1, V1, Mapper<K1, V1, Text, Text>.Context) - Method in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorMapper
the map function.
map(K, V, Mapper<K, V, Text, Text>.Context) - Method in class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionMapper
The identify function.
map(K, V, Mapper<K, V, V, K>.Context) - Method in class org.apache.hadoop.mapreduce.lib.map.InverseMapper
The inverse function.
map(K, Text, Mapper<K, Text, Text, LongWritable>.Context) - Method in class org.apache.hadoop.mapreduce.lib.map.RegexMapper
 
map(Object, Text, Mapper<Object, Text, Text, IntWritable>.Context) - Method in class org.apache.hadoop.mapreduce.lib.map.TokenCounterMapper
 
map(KEYIN, VALUEIN, Mapper<KEYIN, VALUEIN, KEYOUT, VALUEOUT>.Context) - Method in class org.apache.hadoop.mapreduce.Mapper
Called once for each key/value pair in the input split.
map(Object, Object, OutputCollector, Reporter) - Method in class org.apache.hadoop.streaming.PipeMapper
 
MAP_CLASS - Static variable in class org.apache.hadoop.mapreduce.lib.map.MultithreadedMapper
 
MAP_OUTPUT_KEY_VALUE_SPEC - Static variable in class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionHelper
 
mapCmd_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
mapContext - Variable in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
MapContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT> - Interface in org.apache.hadoop.mapreduce
The context that is given to the Mapper.
mapDebugSpec_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
MapFileOutputFormat - Class in org.apache.hadoop.mapred
An OutputFormat that writes MapFiles.
MapFileOutputFormat() - Constructor for class org.apache.hadoop.mapred.MapFileOutputFormat
 
MapFileOutputFormat - Class in org.apache.hadoop.mapreduce.lib.output
An OutputFormat that writes MapFiles.
MapFileOutputFormat() - Constructor for class org.apache.hadoop.mapreduce.lib.output.MapFileOutputFormat
 
Mapper<K1,V1,K2,V2> - Interface in org.apache.hadoop.mapred
Maps input key/value pairs to a set of intermediate key/value pairs.
Mapper<KEYIN,VALUEIN,KEYOUT,VALUEOUT> - Class in org.apache.hadoop.mapreduce
Maps input key/value pairs to a set of intermediate key/value pairs.
Mapper() - Constructor for class org.apache.hadoop.mapreduce.Mapper
 
Mapper.Context - Class in org.apache.hadoop.mapreduce
The Context passed on to the Mapper implementations.
Mapper.Context() - Constructor for class org.apache.hadoop.mapreduce.Mapper.Context
 
mapProgress() - Method in class org.apache.hadoop.mapred.JobStatus
 
mapProgress() - Method in interface org.apache.hadoop.mapred.RunningJob
Get the progress of the job's map-tasks, as a float between 0.0 and 1.0.
mapProgress() - Method in class org.apache.hadoop.mapreduce.Job
Get the progress of the job's map-tasks, as a float between 0.0 and 1.0.
MAPRED_LOCAL_DIR_PROPERTY - Static variable in class org.apache.hadoop.mapred.JobConf
Property name for the configuration property mapreduce.cluster.local.dir
MAPRED_MAP_TASK_ENV - Static variable in class org.apache.hadoop.mapred.JobConf
Configuration key to set the maximum virutal memory available to the map tasks.
MAPRED_MAP_TASK_JAVA_OPTS - Static variable in class org.apache.hadoop.mapred.JobConf
Configuration key to set the java command line options for the map tasks.
MAPRED_MAP_TASK_LOG_LEVEL - Static variable in class org.apache.hadoop.mapred.JobConf
Configuration key to set the logging Level for the map task.
MAPRED_MAP_TASK_ULIMIT - Static variable in class org.apache.hadoop.mapred.JobConf
Configuration key to set the maximum virutal memory available to the map tasks (in kilo-bytes).
MAPRED_REDUCE_TASK_ENV - Static variable in class org.apache.hadoop.mapred.JobConf
Configuration key to set the maximum virutal memory available to the reduce tasks.
MAPRED_REDUCE_TASK_JAVA_OPTS - Static variable in class org.apache.hadoop.mapred.JobConf
Configuration key to set the java command line options for the reduce tasks.
MAPRED_REDUCE_TASK_LOG_LEVEL - Static variable in class org.apache.hadoop.mapred.JobConf
Configuration key to set the logging Level for the reduce task.
MAPRED_REDUCE_TASK_ULIMIT - Static variable in class org.apache.hadoop.mapred.JobConf
Configuration key to set the maximum virutal memory available to the reduce tasks (in kilo-bytes).
MAPRED_TASK_DEFAULT_MAXVMEM_PROPERTY - Static variable in class org.apache.hadoop.mapred.JobConf
Deprecated.  
MAPRED_TASK_ENV - Static variable in class org.apache.hadoop.mapred.JobConf
Deprecated. Use JobConf.MAPRED_MAP_TASK_ENV or JobConf.MAPRED_REDUCE_TASK_ENV
MAPRED_TASK_JAVA_OPTS - Static variable in class org.apache.hadoop.mapred.JobConf
Deprecated. Use JobConf.MAPRED_MAP_TASK_JAVA_OPTS or JobConf.MAPRED_REDUCE_TASK_JAVA_OPTS
MAPRED_TASK_MAXPMEM_PROPERTY - Static variable in class org.apache.hadoop.mapred.JobConf
Deprecated.  
MAPRED_TASK_MAXVMEM_PROPERTY - Static variable in class org.apache.hadoop.mapred.JobConf
Deprecated. Use JobConf.MAPRED_JOB_MAP_MEMORY_MB_PROPERTY and JobConf.MAPRED_JOB_REDUCE_MEMORY_MB_PROPERTY
MAPRED_TASK_ULIMIT - Static variable in class org.apache.hadoop.mapred.JobConf
Deprecated. Use JobConf.MAPRED_MAP_TASK_ULIMIT or JobConf.MAPRED_REDUCE_TASK_ULIMIT
mapRedFinished() - Method in class org.apache.hadoop.streaming.PipeMapRed
 
MapReduceBase - Class in org.apache.hadoop.mapred
Base class for Mapper and Reducer implementations.
MapReduceBase() - Constructor for class org.apache.hadoop.mapred.MapReduceBase
 
MapRunnable<K1,V1,K2,V2> - Interface in org.apache.hadoop.mapred
Expert: Generic interface for Mappers.
MapRunner<K1,V1,K2,V2> - Class in org.apache.hadoop.mapred
Default MapRunnable implementation.
MapRunner() - Constructor for class org.apache.hadoop.mapred.MapRunner
 
MAPS_PER_HOST - Static variable in class org.apache.hadoop.examples.RandomTextWriter
 
MAPS_PER_HOST - Static variable in class org.apache.hadoop.examples.RandomWriter
 
mark() - Method in class org.apache.hadoop.mapreduce.MarkableIterator
 
mark() - Method in class org.apache.hadoop.mapreduce.task.ReduceContextImpl.ValueIterator
 
MarkableIterator<VALUE> - Class in org.apache.hadoop.mapreduce
MarkableIterator is a wrapper iterator class that implements the MarkableIteratorInterface.
MarkableIterator(Iterator<VALUE>) - Constructor for class org.apache.hadoop.mapreduce.MarkableIterator
Create a new iterator layered on the input iterator
MAX_KEY - Static variable in class org.apache.hadoop.examples.RandomTextWriter
 
MAX_KEY - Static variable in class org.apache.hadoop.examples.RandomWriter
 
MAX_NUM_UNIQUE_VALUES - Static variable in class org.apache.hadoop.mapreduce.lib.aggregate.UniqValueCount
 
MAX_TRIE_DEPTH - Static variable in class org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner
 
MAX_VALUE - Static variable in class org.apache.hadoop.examples.RandomTextWriter
 
MAX_VALUE - Static variable in class org.apache.hadoop.examples.RandomWriter
 
MD5_LEN - Static variable in class org.apache.hadoop.mapred.SequenceFileInputFilter.MD5Filter
 
MD5_LEN - Static variable in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.MD5Filter
 
merge(List, List, String) - Method in class org.apache.hadoop.streaming.JarBuilder
 
millis2String(long) - Static method in class org.apache.hadoop.examples.pi.Util
Covert milliseconds to a String.
MIN_KEY - Static variable in class org.apache.hadoop.examples.RandomTextWriter
 
MIN_KEY - Static variable in class org.apache.hadoop.examples.RandomWriter
 
MIN_VALUE - Static variable in class org.apache.hadoop.examples.RandomTextWriter
 
MIN_VALUE - Static variable in class org.apache.hadoop.examples.RandomWriter
 
minRecWrittenToEnableSkip_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
mod(long, long) - Static method in class org.apache.hadoop.examples.pi.math.Modular
Compute 2^e mod n
modInverse(long, long) - Static method in class org.apache.hadoop.examples.pi.math.Modular
Given 0 < x < y, return x^(-1) mod y.
Modular - Class in org.apache.hadoop.examples.pi.math
Modular arithmetics
Modular() - Constructor for class org.apache.hadoop.examples.pi.math.Modular
 
monitorAndPrintJob(JobConf, RunningJob) - Method in class org.apache.hadoop.mapred.JobClient
Monitor a job and print status in real-time as progress is made and tasks fail.
monitorAndPrintJob() - Method in class org.apache.hadoop.mapreduce.Job
Monitor a job and print status in real-time as progress is made and tasks fail.
MRAdmin - Class in org.apache.hadoop.mapred.tools
Administrative access to Hadoop Map-Reduce.
MRAdmin() - Constructor for class org.apache.hadoop.mapred.tools.MRAdmin
 
MRAdmin(Configuration) - Constructor for class org.apache.hadoop.mapred.tools.MRAdmin
 
msg(String) - Method in class org.apache.hadoop.streaming.StreamJob
 
MultiFileInputFormat<K,V> - Class in org.apache.hadoop.mapred
An abstract InputFormat that returns MultiFileSplit's in MultiFileInputFormat.getSplits(JobConf, int) method.
MultiFileInputFormat() - Constructor for class org.apache.hadoop.mapred.MultiFileInputFormat
 
MultiFileSplit - Class in org.apache.hadoop.mapred
A sub-collection of input files.
MultiFileSplit(JobConf, Path[], long[]) - Constructor for class org.apache.hadoop.mapred.MultiFileSplit
 
MultiFileWordCount - Class in org.apache.hadoop.examples
MultiFileWordCount is an example to demonstrate the usage of MultiFileInputFormat.
MultiFileWordCount() - Constructor for class org.apache.hadoop.examples.MultiFileWordCount
 
MultiFileWordCount.CombineFileLineRecordReader - Class in org.apache.hadoop.examples
RecordReader is responsible from extracting records from a chunk of the CombineFileSplit.
MultiFileWordCount.CombineFileLineRecordReader(CombineFileSplit, TaskAttemptContext, Integer) - Constructor for class org.apache.hadoop.examples.MultiFileWordCount.CombineFileLineRecordReader
 
MultiFileWordCount.MapClass - Class in org.apache.hadoop.examples
This Mapper is similar to the one in MultiFileWordCount.MapClass.
MultiFileWordCount.MapClass() - Constructor for class org.apache.hadoop.examples.MultiFileWordCount.MapClass
 
MultiFileWordCount.MyInputFormat - Class in org.apache.hadoop.examples
To use CombineFileInputFormat, one should extend it, to return a (custom) RecordReader.
MultiFileWordCount.MyInputFormat() - Constructor for class org.apache.hadoop.examples.MultiFileWordCount.MyInputFormat
 
MultiFileWordCount.WordOffset - Class in org.apache.hadoop.examples
This record keeps <filename,offset> pairs.
MultiFileWordCount.WordOffset() - Constructor for class org.apache.hadoop.examples.MultiFileWordCount.WordOffset
 
MultiFilterRecordReader<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> - Class in org.apache.hadoop.mapred.join
Base class for Composite join returning values derived from multiple sources, but generally not tuples.
MultiFilterRecordReader(int, JobConf, int, Class<? extends WritableComparator>) - Constructor for class org.apache.hadoop.mapred.join.MultiFilterRecordReader
 
MultiFilterRecordReader<K extends org.apache.hadoop.io.WritableComparable<?>,V extends org.apache.hadoop.io.Writable> - Class in org.apache.hadoop.mapreduce.lib.join
Base class for Composite join returning values derived from multiple sources, but generally not tuples.
MultiFilterRecordReader(int, Configuration, int, Class<? extends WritableComparator>) - Constructor for class org.apache.hadoop.mapreduce.lib.join.MultiFilterRecordReader
 
MultiFilterRecordReader.MultiFilterDelegationIterator - Class in org.apache.hadoop.mapred.join
Proxy the JoinCollector, but include callback to emit.
MultiFilterRecordReader.MultiFilterDelegationIterator() - Constructor for class org.apache.hadoop.mapred.join.MultiFilterRecordReader.MultiFilterDelegationIterator
 
MultiFilterRecordReader.MultiFilterDelegationIterator - Class in org.apache.hadoop.mapreduce.lib.join
Proxy the JoinCollector, but include callback to emit.
MultiFilterRecordReader.MultiFilterDelegationIterator() - Constructor for class org.apache.hadoop.mapreduce.lib.join.MultiFilterRecordReader.MultiFilterDelegationIterator
 
MultipleInputs - Class in org.apache.hadoop.mapred.lib
This class supports MapReduce jobs that have multiple input paths with a different InputFormat and Mapper for each path
MultipleInputs() - Constructor for class org.apache.hadoop.mapred.lib.MultipleInputs
 
MultipleInputs - Class in org.apache.hadoop.mapreduce.lib.input
This class supports MapReduce jobs that have multiple input paths with a different InputFormat and Mapper for each path
MultipleInputs() - Constructor for class org.apache.hadoop.mapreduce.lib.input.MultipleInputs
 
MultipleOutputFormat<K,V> - Class in org.apache.hadoop.mapred.lib
This abstract class extends the FileOutputFormat, allowing to write the output data to different output files.
MultipleOutputFormat() - Constructor for class org.apache.hadoop.mapred.lib.MultipleOutputFormat
 
MultipleOutputs - Class in org.apache.hadoop.mapred.lib
The MultipleOutputs class simplifies writting to additional outputs other than the job default output via the OutputCollector passed to the map() and reduce() methods of the Mapper and Reducer implementations.
MultipleOutputs(JobConf) - Constructor for class org.apache.hadoop.mapred.lib.MultipleOutputs
Creates and initializes multiple named outputs support, it should be instantiated in the Mapper/Reducer configure method.
MultipleOutputs<KEYOUT,VALUEOUT> - Class in org.apache.hadoop.mapreduce.lib.output
The MultipleOutputs class simplifies writing output data to multiple outputs
MultipleOutputs(TaskInputOutputContext<?, ?, KEYOUT, VALUEOUT>) - Constructor for class org.apache.hadoop.mapreduce.lib.output.MultipleOutputs
Creates and initializes multiple outputs support, it should be instantiated in the Mapper/Reducer setup method.
MultipleSequenceFileOutputFormat<K,V> - Class in org.apache.hadoop.mapred.lib
This class extends the MultipleOutputFormat, allowing to write the output data to different output files in sequence file output format.
MultipleSequenceFileOutputFormat() - Constructor for class org.apache.hadoop.mapred.lib.MultipleSequenceFileOutputFormat
 
MultipleTextOutputFormat<K,V> - Class in org.apache.hadoop.mapred.lib
This class extends the MultipleOutputFormat, allowing to write the output data to different output files in Text output format.
MultipleTextOutputFormat() - Constructor for class org.apache.hadoop.mapred.lib.MultipleTextOutputFormat
 
MultithreadedMapper<K1,V1,K2,V2> - Class in org.apache.hadoop.mapreduce.lib.map
Multithreaded implementation for @link org.apache.hadoop.mapreduce.Mapper.
MultithreadedMapper() - Constructor for class org.apache.hadoop.mapreduce.lib.map.MultithreadedMapper
 
MultithreadedMapRunner<K1,V1,K2,V2> - Class in org.apache.hadoop.mapred.lib
Multithreaded implementation for @link org.apache.hadoop.mapred.MapRunnable.
MultithreadedMapRunner() - Constructor for class org.apache.hadoop.mapred.lib.MultithreadedMapRunner
 
MySQLDataDrivenDBRecordReader<T extends DBWritable> - Class in org.apache.hadoop.mapreduce.lib.db
A RecordReader that reads records from a MySQL table via DataDrivenDBRecordReader
MySQLDataDrivenDBRecordReader(DBInputFormat.DBInputSplit, Class<T>, Configuration, Connection, DBConfiguration, String, String[], String) - Constructor for class org.apache.hadoop.mapreduce.lib.db.MySQLDataDrivenDBRecordReader
 
MySQLDBRecordReader<T extends DBWritable> - Class in org.apache.hadoop.mapreduce.lib.db
A RecordReader that reads records from a MySQL table.
MySQLDBRecordReader(DBInputFormat.DBInputSplit, Class<T>, Configuration, Connection, DBConfiguration, String, String[], String) - Constructor for class org.apache.hadoop.mapreduce.lib.db.MySQLDBRecordReader
 

N

N - Variable in class org.apache.hadoop.examples.pi.math.Summation
Variable n in the summation.
NATURAL_ORDER - Static variable in class org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner
 
needsTaskCommit(TaskAttemptContext) - Method in class org.apache.hadoop.mapred.FileOutputCommitter
 
needsTaskCommit(TaskAttemptContext) - Method in class org.apache.hadoop.mapred.OutputCommitter
Check whether task needs a commit
needsTaskCommit(TaskAttemptContext) - Method in class org.apache.hadoop.mapred.OutputCommitter
This method implements the new interface by calling the old method.
needsTaskCommit(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
Did this task write any files in the work directory?
needsTaskCommit(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.OutputCommitter
Check whether task needs a commit
next(DocumentID, LineDocTextAndOp) - Method in class org.apache.hadoop.contrib.index.example.LineDocRecordReader
 
next() - Method in class org.apache.hadoop.contrib.utils.join.ArrayListBackedIterator
 
next(TupleWritable) - Method in class org.apache.hadoop.mapred.join.JoinRecordReader.JoinDelegationIterator
 
next(K, TupleWritable) - Method in class org.apache.hadoop.mapred.join.JoinRecordReader
Emit the next set of key, value pairs as defined by the child RecordReaders and operation associated with this composite RR.
next(V) - Method in class org.apache.hadoop.mapred.join.MultiFilterRecordReader.MultiFilterDelegationIterator
 
next(K, V) - Method in class org.apache.hadoop.mapred.join.MultiFilterRecordReader
Reads the next key/value pair from the input for processing.
next() - Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
Read the next k,v pair into the head of this object; return true iff the RR and this are exhausted.
next(K, U) - Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
Write key-value pair at the head of this stream to the objects provided; get next key-value pair from proxied RR.
next(Text, Text) - Method in class org.apache.hadoop.mapred.KeyValueLineRecordReader
Read key/value pair in a line.
next(K, V) - Method in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
 
next(LongWritable, T) - Method in class org.apache.hadoop.mapred.lib.db.DBInputFormat.DBRecordReader
Reads the next key/value pair from the input for processing.
next(K, V) - Method in interface org.apache.hadoop.mapred.RecordReader
Reads the next key/value pair from the input for processing.
next(BytesWritable, BytesWritable) - Method in class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
Read raw bytes from a SequenceFile.
next(Text, Text) - Method in class org.apache.hadoop.mapred.SequenceFileAsTextRecordReader
Read key/value pair in a line.
next(K, V) - Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
 
next(K) - Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
 
next(LongWritable, T) - Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
Deprecated. Use DBRecordReader.nextKeyValue()
next(X) - Method in class org.apache.hadoop.mapreduce.lib.join.ArrayListBackedIterator
 
next(TupleWritable) - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader.JoinCollector
Populate Tuple from iterators.
next(TupleWritable) - Method in class org.apache.hadoop.mapreduce.lib.join.JoinRecordReader.JoinDelegationIterator
 
next(V) - Method in class org.apache.hadoop.mapreduce.lib.join.MultiFilterRecordReader.MultiFilterDelegationIterator
 
next(U) - Method in class org.apache.hadoop.mapreduce.lib.join.ResetableIterator.EMPTY
 
next(T) - Method in interface org.apache.hadoop.mapreduce.lib.join.ResetableIterator
Assign next value to actual.
next(X) - Method in class org.apache.hadoop.mapreduce.lib.join.StreamBackedIterator
 
next() - Method in class org.apache.hadoop.mapreduce.MarkableIterator
 
next() - Method in class org.apache.hadoop.mapreduce.task.ReduceContextImpl.ValueIterator
 
next(Text, Text) - Method in class org.apache.hadoop.streaming.StreamBaseRecordReader
Read a record.
next(Text, Text) - Method in class org.apache.hadoop.streaming.StreamXmlRecordReader
 
nextKey() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
nextKey() - Method in interface org.apache.hadoop.mapreduce.ReduceContext
Start processing next unique key.
nextKeyValue() - Method in class org.apache.hadoop.examples.MultiFileWordCount.CombineFileLineRecordReader
 
nextKeyValue() - Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
Read the next key, value pair.
nextKeyValue() - Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader
 
nextKeyValue() - Method in class org.apache.hadoop.mapreduce.lib.input.KeyValueLineRecordReader
Read key/value pair in a line.
nextKeyValue() - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
Read raw bytes from a SequenceFile.
nextKeyValue() - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsTextRecordReader
Read key/value pair in a line.
nextKeyValue() - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileRecordReader
 
nextKeyValue() - Method in class org.apache.hadoop.mapreduce.lib.join.JoinRecordReader
Emit the next set of key, value pairs as defined by the child RecordReaders and operation associated with this composite RR.
nextKeyValue() - Method in class org.apache.hadoop.mapreduce.lib.join.MultiFilterRecordReader
Read the next key, value pair.
nextKeyValue() - Method in class org.apache.hadoop.mapreduce.lib.join.WrappedRecordReader
Read the next k,v pair into the head of this object; return true iff the RR and this are exhausted.
nextKeyValue() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
nextKeyValue() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
nextKeyValue() - Method in class org.apache.hadoop.mapreduce.RecordReader
Read the next key, value pair.
nextKeyValue() - Method in interface org.apache.hadoop.mapreduce.TaskInputOutputContext
Advance to the next key, value pair, returning null if at end.
NLineInputFormat - Class in org.apache.hadoop.mapred.lib
NLineInputFormat which splits N lines of input as one split.
NLineInputFormat() - Constructor for class org.apache.hadoop.mapred.lib.NLineInputFormat
 
NLineInputFormat - Class in org.apache.hadoop.mapreduce.lib.input
NLineInputFormat which splits N lines of input as one split.
NLineInputFormat() - Constructor for class org.apache.hadoop.mapreduce.lib.input.NLineInputFormat
 
normalizeMemoryConfigValue(long) - Static method in class org.apache.hadoop.mapred.JobConf
Normalize the negative values in configuration
normalizePath(String) - Static method in class org.apache.hadoop.contrib.index.mapred.Shard
 
NULL - Static variable in interface org.apache.hadoop.mapred.Reporter
A constant of Reporter type that does nothing.
NullOutputFormat<K,V> - Class in org.apache.hadoop.mapred.lib
Consume all outputs and put them in /dev/null.
NullOutputFormat() - Constructor for class org.apache.hadoop.mapred.lib.NullOutputFormat
 
NullOutputFormat<K,V> - Class in org.apache.hadoop.mapreduce.lib.output
Consume all outputs and put them in /dev/null.
NullOutputFormat() - Constructor for class org.apache.hadoop.mapreduce.lib.output.NullOutputFormat
 
NUM_INPUT_FILES - Static variable in class org.apache.hadoop.mapred.FileInputFormat
 
NUM_INPUT_FILES - Static variable in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
 
NUM_OF_VALUES_FIELD - Static variable in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
 
NUM_ROWS - Static variable in class org.apache.hadoop.examples.terasort.TeraGen
 
NUM_THREADS - Static variable in class org.apache.hadoop.mapreduce.lib.map.MultithreadedMapper
 
numOfValues - Variable in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
 
numReduceTasksSpec_ - Variable in class org.apache.hadoop.streaming.StreamJob
 

O

obtainTokensForNamenodes(Path[], Configuration) - Static method in class org.apache.hadoop.mapreduce.security.TokenCache
Convenience method to obtain delegation tokens from namenodes corresponding to the paths passed.
ONE - Static variable in interface org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorDescriptor
 
ONE - Static variable in interface org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorDescriptor
 
oneRotation - Static variable in class org.apache.hadoop.examples.dancing.Pentomino
Is the piece fixed under rotation?
OneSidedPentomino - Class in org.apache.hadoop.examples.dancing
Of the "normal" 12 pentominos, 6 of them have distinct shapes when flipped.
OneSidedPentomino() - Constructor for class org.apache.hadoop.examples.dancing.OneSidedPentomino
 
OneSidedPentomino(int, int) - Constructor for class org.apache.hadoop.examples.dancing.OneSidedPentomino
 
openInput(String) - Method in class org.apache.hadoop.contrib.index.lucene.FileSystemDirectory
 
openInput(String, int) - Method in class org.apache.hadoop.contrib.index.lucene.FileSystemDirectory
 
OracleDataDrivenDBInputFormat<T extends DBWritable> - Class in org.apache.hadoop.mapreduce.lib.db
A InputFormat that reads input data from an SQL table in an Oracle db.
OracleDataDrivenDBInputFormat() - Constructor for class org.apache.hadoop.mapreduce.lib.db.OracleDataDrivenDBInputFormat
 
OracleDataDrivenDBRecordReader<T extends DBWritable> - Class in org.apache.hadoop.mapreduce.lib.db
A RecordReader that reads records from a Oracle table via DataDrivenDBRecordReader
OracleDataDrivenDBRecordReader(DBInputFormat.DBInputSplit, Class<T>, Configuration, Connection, DBConfiguration, String, String[], String) - Constructor for class org.apache.hadoop.mapreduce.lib.db.OracleDataDrivenDBRecordReader
 
OracleDateSplitter - Class in org.apache.hadoop.mapreduce.lib.db
Implement DBSplitter over date/time values returned by an Oracle db.
OracleDateSplitter() - Constructor for class org.apache.hadoop.mapreduce.lib.db.OracleDateSplitter
 
OracleDBRecordReader<T extends DBWritable> - Class in org.apache.hadoop.mapreduce.lib.db
A RecordReader that reads records from an Oracle SQL table.
OracleDBRecordReader(DBInputFormat.DBInputSplit, Class<T>, Configuration, Connection, DBConfiguration, String, String[], String) - Constructor for class org.apache.hadoop.mapreduce.lib.db.OracleDBRecordReader
 
org.apache.hadoop.contrib.index.example - package org.apache.hadoop.contrib.index.example
 
org.apache.hadoop.contrib.index.lucene - package org.apache.hadoop.contrib.index.lucene
 
org.apache.hadoop.contrib.index.main - package org.apache.hadoop.contrib.index.main
 
org.apache.hadoop.contrib.index.mapred - package org.apache.hadoop.contrib.index.mapred
 
org.apache.hadoop.contrib.utils.join - package org.apache.hadoop.contrib.utils.join
 
org.apache.hadoop.examples - package org.apache.hadoop.examples
Hadoop example code.
org.apache.hadoop.examples.dancing - package org.apache.hadoop.examples.dancing
This package is a distributed implementation of Knuth's dancing links algorithm that can run under Hadoop.
org.apache.hadoop.examples.pi - package org.apache.hadoop.examples.pi
This package consists of a map/reduce application, distbbp, which computes exact binary digits of the mathematical constant π.
org.apache.hadoop.examples.pi.math - package org.apache.hadoop.examples.pi.math
This package provides useful mathematical library classes for the distbbp program.
org.apache.hadoop.examples.terasort - package org.apache.hadoop.examples.terasort
This package consists of 3 map/reduce applications for Hadoop to compete in the annual terabyte sort competition.
org.apache.hadoop.filecache - package org.apache.hadoop.filecache
Deprecated. Use Job instead.
org.apache.hadoop.mapred - package org.apache.hadoop.mapred
A software framework for easily writing applications which process vast amounts of data (multi-terabyte data-sets) parallelly on large clusters (thousands of nodes) built of commodity hardware in a reliable, fault-tolerant manner.
org.apache.hadoop.mapred.gridmix - package org.apache.hadoop.mapred.gridmix
 
org.apache.hadoop.mapred.jobcontrol - package org.apache.hadoop.mapred.jobcontrol
Utilities for managing dependent jobs.
org.apache.hadoop.mapred.join - package org.apache.hadoop.mapred.join
Given a set of sorted datasets keyed with the same class and yielding equal partitions, it is possible to effect a join of those datasets prior to the map.
org.apache.hadoop.mapred.lib - package org.apache.hadoop.mapred.lib
Library of generally useful mappers, reducers, and partitioners.
org.apache.hadoop.mapred.lib.aggregate - package org.apache.hadoop.mapred.lib.aggregate
Classes for performing various counting and aggregations.
org.apache.hadoop.mapred.lib.db - package org.apache.hadoop.mapred.lib.db
org.apache.hadoop.mapred.lib.db Package
org.apache.hadoop.mapred.pipes - package org.apache.hadoop.mapred.pipes
Hadoop Pipes allows C++ code to use Hadoop DFS and map/reduce.
org.apache.hadoop.mapred.tools - package org.apache.hadoop.mapred.tools
 
org.apache.hadoop.mapreduce - package org.apache.hadoop.mapreduce
 
org.apache.hadoop.mapreduce.lib.aggregate - package org.apache.hadoop.mapreduce.lib.aggregate
Classes for performing various counting and aggregations.
org.apache.hadoop.mapreduce.lib.chain - package org.apache.hadoop.mapreduce.lib.chain
 
org.apache.hadoop.mapreduce.lib.db - package org.apache.hadoop.mapreduce.lib.db
org.apache.hadoop.mapred.lib.db Package
org.apache.hadoop.mapreduce.lib.fieldsel - package org.apache.hadoop.mapreduce.lib.fieldsel
 
org.apache.hadoop.mapreduce.lib.input - package org.apache.hadoop.mapreduce.lib.input
 
org.apache.hadoop.mapreduce.lib.jobcontrol - package org.apache.hadoop.mapreduce.lib.jobcontrol
Utilities for managing dependent jobs.
org.apache.hadoop.mapreduce.lib.join - package org.apache.hadoop.mapreduce.lib.join
Given a set of sorted datasets keyed with the same class and yielding equal partitions, it is possible to effect a join of those datasets prior to the map.
org.apache.hadoop.mapreduce.lib.map - package org.apache.hadoop.mapreduce.lib.map
 
org.apache.hadoop.mapreduce.lib.output - package org.apache.hadoop.mapreduce.lib.output
 
org.apache.hadoop.mapreduce.lib.partition - package org.apache.hadoop.mapreduce.lib.partition
 
org.apache.hadoop.mapreduce.lib.reduce - package org.apache.hadoop.mapreduce.lib.reduce
 
org.apache.hadoop.mapreduce.security - package org.apache.hadoop.mapreduce.security
 
org.apache.hadoop.mapreduce.server.jobtracker - package org.apache.hadoop.mapreduce.server.jobtracker
 
org.apache.hadoop.mapreduce.tools - package org.apache.hadoop.mapreduce.tools
 
org.apache.hadoop.streaming - package org.apache.hadoop.streaming
Hadoop Streaming is a utility which allows users to create and run Map-Reduce jobs with any executables (e.g.
org.apache.hadoop.streaming.io - package org.apache.hadoop.streaming.io
 
org.apache.hadoop.typedbytes - package org.apache.hadoop.typedbytes
Typed bytes are sequences of bytes in which the first byte is a type code.
out - Static variable in class org.apache.hadoop.examples.pi.Util
Output stream
out - Variable in class org.apache.hadoop.mapred.TextOutputFormat.LineRecordWriter
 
out - Variable in class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat.LineRecordWriter
 
OUTDIR - Static variable in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
 
OuterJoinRecordReader<K extends org.apache.hadoop.io.WritableComparable> - Class in org.apache.hadoop.mapred.join
Full outer join.
OuterJoinRecordReader<K extends org.apache.hadoop.io.WritableComparable<?>> - Class in org.apache.hadoop.mapreduce.lib.join
Full outer join.
outerrThreadsThrowable - Variable in class org.apache.hadoop.streaming.PipeMapRed
 
output_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
OUTPUT_FIELD_COUNT_PROPERTY - Static variable in class org.apache.hadoop.mapred.lib.db.DBConfiguration
Number of fields in the Output table
OUTPUT_FIELD_COUNT_PROPERTY - Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
Number of fields in the Output table
OUTPUT_FIELD_NAMES_PROPERTY - Static variable in class org.apache.hadoop.mapred.lib.db.DBConfiguration
Field names in the Output table
OUTPUT_FIELD_NAMES_PROPERTY - Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
Field names in the Output table
OUTPUT_FILTER - Static variable in class org.apache.hadoop.mapreduce.Job
 
OUTPUT_FORMAT - Static variable in class org.apache.hadoop.mapreduce.lib.output.LazyOutputFormat
 
OUTPUT_TABLE_NAME_PROPERTY - Static variable in class org.apache.hadoop.mapred.lib.db.DBConfiguration
Output table name
OUTPUT_TABLE_NAME_PROPERTY - Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
Output table name
OutputCollector<K,V> - Interface in org.apache.hadoop.mapred
Collects the <key, value> pairs output by Mappers and Reducers.
OutputCommitter - Class in org.apache.hadoop.mapred
OutputCommitter describes the commit of task output for a Map-Reduce job.
OutputCommitter() - Constructor for class org.apache.hadoop.mapred.OutputCommitter
 
OutputCommitter - Class in org.apache.hadoop.mapreduce
OutputCommitter describes the commit of task output for a Map-Reduce job.
OutputCommitter() - Constructor for class org.apache.hadoop.mapreduce.OutputCommitter
 
OutputFormat<K,V> - Interface in org.apache.hadoop.mapred
OutputFormat describes the output-specification for a Map-Reduce job.
OutputFormat<K,V> - Class in org.apache.hadoop.mapreduce
OutputFormat describes the output-specification for a Map-Reduce job.
OutputFormat() - Constructor for class org.apache.hadoop.mapreduce.OutputFormat
 
outputFormatSpec_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
OutputLogFilter - Class in org.apache.hadoop.mapred
Deprecated. Use Utils.OutputFileUtils.OutputLogFilter instead.
OutputLogFilter() - Constructor for class org.apache.hadoop.mapred.OutputLogFilter
Deprecated.  
OutputReader<K,V> - Class in org.apache.hadoop.streaming.io
Abstract base for classes that read the client's output.
OutputReader() - Constructor for class org.apache.hadoop.streaming.io.OutputReader
 
outputRecords(OutputStream, boolean, Unsigned16, Unsigned16, Unsigned16) - Static method in class org.apache.hadoop.examples.terasort.GenSort
 
outputSingleNode_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
OverrideRecordReader<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> - Class in org.apache.hadoop.mapred.join
Prefer the "rightmost" data source for this key.
OverrideRecordReader<K extends org.apache.hadoop.io.WritableComparable<?>,V extends org.apache.hadoop.io.Writable> - Class in org.apache.hadoop.mapreduce.lib.join
Prefer the "rightmost" data source for this key.

P

packageFiles_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
packageJobJar() - Method in class org.apache.hadoop.streaming.StreamJob
 
parseLongVariable(String, String) - Static method in class org.apache.hadoop.examples.pi.Util
Parse a variable.
parseOutputKeyValueSpec(String, List<Integer>, List<Integer>) - Static method in class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionHelper
 
Parser - Class in org.apache.hadoop.examples.pi
A class for parsing outputs
Parser(boolean) - Constructor for class org.apache.hadoop.examples.pi.Parser
 
Parser - Class in org.apache.hadoop.mapred.join
Very simple shift-reduce parser for join expressions.
Parser() - Constructor for class org.apache.hadoop.mapred.join.Parser
 
Parser - Class in org.apache.hadoop.mapreduce.lib.join
Very simple shift-reduce parser for join expressions.
Parser() - Constructor for class org.apache.hadoop.mapreduce.lib.join.Parser
 
Parser.Node - Class in org.apache.hadoop.mapred.join
 
Parser.Node(String) - Constructor for class org.apache.hadoop.mapred.join.Parser.Node
 
Parser.Node - Class in org.apache.hadoop.mapreduce.lib.join
 
Parser.Node(String) - Constructor for class org.apache.hadoop.mapreduce.lib.join.Parser.Node
 
Parser.NodeToken - Class in org.apache.hadoop.mapred.join
 
Parser.NodeToken - Class in org.apache.hadoop.mapreduce.lib.join
 
Parser.NumToken - Class in org.apache.hadoop.mapred.join
 
Parser.NumToken(double) - Constructor for class org.apache.hadoop.mapred.join.Parser.NumToken
 
Parser.NumToken - Class in org.apache.hadoop.mapreduce.lib.join
 
Parser.NumToken(double) - Constructor for class org.apache.hadoop.mapreduce.lib.join.Parser.NumToken
 
Parser.StrToken - Class in org.apache.hadoop.mapred.join
 
Parser.StrToken(Parser.TType, String) - Constructor for class org.apache.hadoop.mapred.join.Parser.StrToken
 
Parser.StrToken - Class in org.apache.hadoop.mapreduce.lib.join
 
Parser.StrToken(Parser.TType, String) - Constructor for class org.apache.hadoop.mapreduce.lib.join.Parser.StrToken
 
Parser.Token - Class in org.apache.hadoop.mapred.join
Tagged-union type for tokens from the join expression.
Parser.Token - Class in org.apache.hadoop.mapreduce.lib.join
Tagged-union type for tokens from the join expression.
Parser.TType - Enum in org.apache.hadoop.mapred.join
 
Parser.TType - Enum in org.apache.hadoop.mapreduce.lib.join
 
parseStringVariable(String, String) - Static method in class org.apache.hadoop.examples.pi.Util
Parse a variable.
PART - Static variable in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
 
partition(int) - Method in class org.apache.hadoop.examples.pi.math.Summation
Partition the summation.
Partitioner<K2,V2> - Interface in org.apache.hadoop.mapred
Partitions the key space.
PARTITIONER - Static variable in class org.apache.hadoop.mapred.pipes.Submitter
 
Partitioner<KEY,VALUE> - Class in org.apache.hadoop.mapreduce
Partitions the key space.
Partitioner() - Constructor for class org.apache.hadoop.mapreduce.Partitioner
 
PARTITIONER_OPTIONS - Static variable in class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedPartitioner
 
PARTITIONER_PATH - Static variable in class org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner
 
partitionerSpec_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
PASSWORD_PROPERTY - Static variable in class org.apache.hadoop.mapred.lib.db.DBConfiguration
Password to access the database
PASSWORD_PROPERTY - Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
Password to access the database
PATHFILTER_CLASS - Static variable in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
 
PathFinder - Class in org.apache.hadoop.streaming
Maps a relative pathname to an absolute pathname using the PATH enviroment.
PathFinder() - Constructor for class org.apache.hadoop.streaming.PathFinder
Construct a PathFinder object using the path from java.class.path
PathFinder(String) - Constructor for class org.apache.hadoop.streaming.PathFinder
Construct a PathFinder object using the path from the specified system environment variable.
PATTERN - Static variable in class org.apache.hadoop.mapreduce.lib.map.RegexMapper
 
Pentomino - Class in org.apache.hadoop.examples.dancing
 
Pentomino(int, int) - Constructor for class org.apache.hadoop.examples.dancing.Pentomino
Create the model for a given pentomino set of pieces and board size.
Pentomino() - Constructor for class org.apache.hadoop.examples.dancing.Pentomino
Create the object without initialization.
Pentomino.ColumnName - Interface in org.apache.hadoop.examples.dancing
This interface just is a marker for what types I expect to get back as column names.
Pentomino.Piece - Class in org.apache.hadoop.examples.dancing
Maintain information about a puzzle piece.
Pentomino.Piece(String, String, boolean, int[]) - Constructor for class org.apache.hadoop.examples.dancing.Pentomino.Piece
 
Pentomino.SolutionCategory - Enum in org.apache.hadoop.examples.dancing
 
pi2string(double, long) - Static method in class org.apache.hadoop.examples.pi.Util
Convert a pi value to a String.
pieces - Variable in class org.apache.hadoop.examples.dancing.Pentomino
 
PipeCombiner - Class in org.apache.hadoop.streaming
 
PipeCombiner() - Constructor for class org.apache.hadoop.streaming.PipeCombiner
 
PipeMapper - Class in org.apache.hadoop.streaming
A generic Mapper bridge.
PipeMapper() - Constructor for class org.apache.hadoop.streaming.PipeMapper
 
PipeMapRed - Class in org.apache.hadoop.streaming
Shared functionality for PipeMapper, PipeReducer.
PipeMapRed() - Constructor for class org.apache.hadoop.streaming.PipeMapRed
 
PipeMapRunner<K1,V1,K2,V2> - Class in org.apache.hadoop.streaming
 
PipeMapRunner() - Constructor for class org.apache.hadoop.streaming.PipeMapRunner
 
PipeReducer - Class in org.apache.hadoop.streaming
A generic Reducer bridge.
PipeReducer() - Constructor for class org.apache.hadoop.streaming.PipeReducer
 
PORT - Static variable in class org.apache.hadoop.mapred.pipes.Submitter
 
PREP - Static variable in class org.apache.hadoop.mapred.JobStatus
 
prependPathComponent(String) - Method in class org.apache.hadoop.streaming.PathFinder
Appends the specified component to the path list
PRESERVE_COMMANDFILE - Static variable in class org.apache.hadoop.mapred.pipes.Submitter
 
printAll() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.JobInfo
Print all the job information
printAll() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskAttemptInfo
Print all the information about this attempt.
printAll() - Method in class org.apache.hadoop.mapreduce.jobhistory.JobHistoryParser.TaskInfo
 
printBitSkipped(long) - Static method in class org.apache.hadoop.examples.pi.Util
Print a "bits skipped" message.
printUsage(String[], String) - Static method in class org.apache.hadoop.examples.pi.Util
Print usage messages
printUsage(PrintStream) - Method in class org.apache.hadoop.mapred.gridmix.Gridmix
 
process(IntermediateForm) - Method in class org.apache.hadoop.contrib.index.lucene.ShardWriter
Process an intermediate form by carrying out, on the Lucene instance of the shard, the deletes and the inserts (a ram index) in the form.
process(DocumentAndOp, Analyzer) - Method in class org.apache.hadoop.contrib.index.mapred.IntermediateForm
This method is used by the index update mapper and process a document operation into the current intermediate form.
process(IntermediateForm) - Method in class org.apache.hadoop.contrib.index.mapred.IntermediateForm
This method is used by the index update combiner and process an intermediate form into the current intermediate form.
progress - Variable in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
 
progress - Variable in class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader
 
progress() - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
progress() - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
PROGRESS_MONITOR_POLL_INTERVAL_KEY - Static variable in class org.apache.hadoop.mapreduce.Job
Key in mapred-*.xml that sets progMonitorPollIntervalMillis

Q

QuasiMonteCarlo - Class in org.apache.hadoop.examples
A map/reduce program that estimates the value of Pi using a quasi-Monte Carlo (qMC) method.
QuasiMonteCarlo() - Constructor for class org.apache.hadoop.examples.QuasiMonteCarlo
 
QuasiMonteCarlo.QmcMapper - Class in org.apache.hadoop.examples
Mapper class for Pi estimation.
QuasiMonteCarlo.QmcMapper() - Constructor for class org.apache.hadoop.examples.QuasiMonteCarlo.QmcMapper
 
QuasiMonteCarlo.QmcReducer - Class in org.apache.hadoop.examples
Reducer class for Pi estimation.
QuasiMonteCarlo.QmcReducer() - Constructor for class org.apache.hadoop.examples.QuasiMonteCarlo.QmcReducer
 
QueueAclsInfo - Class in org.apache.hadoop.mapreduce
Class to encapsulate Queue ACLs for a particular user.
QueueAclsInfo() - Constructor for class org.apache.hadoop.mapreduce.QueueAclsInfo
Default constructor for QueueAclsInfo.
QueueAclsInfo(String, String[]) - Constructor for class org.apache.hadoop.mapreduce.QueueAclsInfo
Construct a new QueueAclsInfo object using the queue name and the queue operations array
QueueInfo - Class in org.apache.hadoop.mapreduce
Class that contains the information regarding the Job Queues which are maintained by the Hadoop Map/Reduce framework.
QueueInfo() - Constructor for class org.apache.hadoop.mapreduce.QueueInfo
Default constructor for QueueInfo.
QueueInfo(String, String) - Constructor for class org.apache.hadoop.mapreduce.QueueInfo
Construct a new QueueInfo object using the queue name and the scheduling information passed.
QueueInfo(String, String, QueueState, JobStatus[]) - Constructor for class org.apache.hadoop.mapreduce.QueueInfo
 
QueueState - Enum in org.apache.hadoop.mapreduce
Enum representing queue state

R

RAMDirectoryUtil - Class in org.apache.hadoop.contrib.index.lucene
A utility class which writes an index in a ram dir into a DataOutput and read from a DataInput an index into a ram dir.
RAMDirectoryUtil() - Constructor for class org.apache.hadoop.contrib.index.lucene.RAMDirectoryUtil
 
RandomTextWriter - Class in org.apache.hadoop.examples
This program uses map/reduce to just run a distributed job where there is no interaction between the tasks and each task writes a large unsorted random sequence of words.
RandomTextWriter() - Constructor for class org.apache.hadoop.examples.RandomTextWriter
 
RandomWriter - Class in org.apache.hadoop.examples
This program uses map/reduce to just run a distributed job where there is no interaction between the tasks and each task write a large unsorted random binary sequence file of BytesWritable.
RandomWriter() - Constructor for class org.apache.hadoop.examples.RandomWriter
 
RAW_BYTES_ID - Static variable in class org.apache.hadoop.streaming.io.IdentifierResolver
 
RawBytesInputWriter - Class in org.apache.hadoop.streaming.io
InputWriter that writes the client's input as raw bytes.
RawBytesInputWriter() - Constructor for class org.apache.hadoop.streaming.io.RawBytesInputWriter
 
RawBytesOutputReader - Class in org.apache.hadoop.streaming.io
OutputReader that reads the client's output as raw bytes.
RawBytesOutputReader() - Constructor for class org.apache.hadoop.streaming.io.RawBytesOutputReader
 
rawWriter - Variable in class org.apache.hadoop.mapred.lib.FilterOutputFormat.FilterRecordWriter
 
rawWriter - Variable in class org.apache.hadoop.mapreduce.lib.output.FilterOutputFormat.FilterRecordWriter
 
read(Class<?>, Configuration) - Static method in class org.apache.hadoop.examples.pi.SummationWritable
Read sigma from conf
read(DataInput) - Static method in class org.apache.hadoop.mapred.JobID
Deprecated. 
read(DataInput) - Static method in class org.apache.hadoop.mapred.TaskAttemptID
Deprecated. 
read(DataInput) - Static method in class org.apache.hadoop.mapred.TaskID
Deprecated. 
read() - Method in class org.apache.hadoop.mapred.TaskLog.Reader
 
read(byte[], int, int) - Method in class org.apache.hadoop.mapred.TaskLog.Reader
 
read() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads a typed bytes sequence and converts it to a Java object.
read() - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readArray(ArrayWritable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readArray() - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readBool() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads the boolean following a Type.BOOL code.
readBool(String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
 
readBoolean(BooleanWritable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readBoolean() - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readBuffer(String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
 
readByte() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads the byte following a Type.BYTE code.
readByte(String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
 
readByte(ByteWritable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readByte() - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readBytes() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads the bytes following a Type.BYTES code.
readBytes(BytesWritable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readBytes() - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readDouble() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads the double following a Type.DOUBLE code.
readDouble(String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
 
readDouble(DoubleWritable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readDouble() - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readFields(DataInput) - Method in class org.apache.hadoop.contrib.index.example.LineDocTextAndOp
 
readFields(DataInput) - Method in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
 
readFields(DataInput) - Method in class org.apache.hadoop.contrib.index.mapred.DocumentID
 
readFields(DataInput) - Method in class org.apache.hadoop.contrib.index.mapred.IntermediateForm
 
readFields(DataInput) - Method in class org.apache.hadoop.contrib.index.mapred.Shard
 
readFields(DataInput) - Method in class org.apache.hadoop.examples.BaileyBorweinPlouffe.BbpSplit
readFields(DataInput) - Method in class org.apache.hadoop.examples.MultiFileWordCount.WordOffset
 
readFields(DataInput) - Method in class org.apache.hadoop.examples.pi.DistSum.Machine.SummationSplit
readFields(DataInput) - Method in class org.apache.hadoop.examples.pi.SummationWritable
readFields(DataInput) - Method in class org.apache.hadoop.examples.pi.TaskResult
readFields(DataInput) - Method in class org.apache.hadoop.examples.SecondarySort.IntPair
Read the two integers.
readFields(DataInput) - Method in class org.apache.hadoop.mapred.ClusterStatus.BlackListInfo
 
readFields(DataInput) - Method in class org.apache.hadoop.mapred.ClusterStatus
 
readFields(DataInput) - Method in class org.apache.hadoop.mapred.Counters.Group
 
readFields(DataInput) - Method in class org.apache.hadoop.mapred.Counters
Read a set of groups.
readFields(DataInput) - Method in class org.apache.hadoop.mapred.FileSplit
 
readFields(DataInput) - Method in class org.apache.hadoop.mapred.join.CompositeInputSplit
readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.ClusterMetrics
 
readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.Counter
Read the binary representation of the counter
readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.CounterGroup
 
readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.Counters
Read a set of groups.
readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.ID
 
readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.JobID
 
readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.JobStatus
 
readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat.DataDrivenDBInputSplit
readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit
readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat.NullDBWritable
 
readFields(ResultSet) - Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat.NullDBWritable
 
readFields(ResultSet) - Method in interface org.apache.hadoop.mapreduce.lib.db.DBWritable
Reads the fields of the object from the ResultSet.
readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
 
readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.lib.input.FileSplit
 
readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeInputSplit
readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.lib.join.TupleWritable
readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.QueueAclsInfo
 
readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.QueueInfo
 
readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.split.JobSplit.SplitMetaInfo
 
readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitIndex
 
readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.TaskAttemptID
 
readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.TaskCompletionEvent
 
readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.TaskID
 
readFields(DataInput) - Method in class org.apache.hadoop.mapreduce.TaskTrackerInfo
 
readFloat() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads the float following a Type.FLOAT code.
readFloat(String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
 
readFloat(FloatWritable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readFloat() - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readInt() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads the integer following a Type.INT code.
readInt(String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
 
readInt(IntWritable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readInt() - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readKeyValue() - Method in class org.apache.hadoop.streaming.io.OutputReader
Read the next key/value pair outputted by the client.
readKeyValue() - Method in class org.apache.hadoop.streaming.io.RawBytesOutputReader
 
readKeyValue() - Method in class org.apache.hadoop.streaming.io.TextOutputReader
 
readKeyValue() - Method in class org.apache.hadoop.streaming.io.TypedBytesOutputReader
 
readLine(LineReader, Text) - Static method in class org.apache.hadoop.streaming.StreamKeyValUtil
Read a utf8 encoded line from a data input stream.
readLine(LineReader, Text) - Static method in class org.apache.hadoop.streaming.UTF8ByteArrayUtils
Deprecated. use StreamKeyValUtil.readLine(LineReader, Text)
readList() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads the list following a Type.LIST code.
readLong() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads the long following a Type.LONG code.
readLong(String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
 
readLong(LongWritable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readLong() - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readMap() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads the map following a Type.MAP code.
readMap(MapWritable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readMap() - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readMapHeader() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads the header following a Type.MAP code.
readRAMFiles(DataInput, RAMDirectory) - Static method in class org.apache.hadoop.contrib.index.lucene.RAMDirectoryUtil
Read a number of files from a data input to a ram directory.
readRaw() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads a typed bytes sequence.
readRawBool() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads the raw bytes following a Type.BOOL code.
readRawByte() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads the raw byte following a Type.BYTE code.
readRawBytes(int) - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads the raw bytes following a custom code.
readRawBytes() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads the raw bytes following a Type.BYTES code.
readRawDouble() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads the raw bytes following a Type.DOUBLE code.
readRawFloat() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads the raw bytes following a Type.FLOAT code.
readRawInt() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads the raw bytes following a Type.INT code.
readRawList() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads the raw bytes following a Type.LIST code.
readRawLong() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads the raw bytes following a Type.LONG code.
readRawMap() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads the raw bytes following a Type.MAP code.
readRawString() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads the raw bytes following a Type.STRING code.
readRawVector() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads the raw bytes following a Type.VECTOR code.
readSortedMap(SortedMapWritable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readSortedMap() - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readString() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads the string following a Type.STRING code.
readString(String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
 
readText(Text) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readText() - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readType() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads a type byte and returns the corresponding Type.
readType() - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readVector() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads the vector following a Type.VECTOR code.
readVectorHeader() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Reads the header following a Type.VECTOR code.
readVInt(VIntWritable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readVInt() - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readVLong(VLongWritable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readVLong() - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readWritable(Writable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
readWritable() - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
READY - Static variable in class org.apache.hadoop.mapred.jobcontrol.Job
 
RecordReader<K,V> - Interface in org.apache.hadoop.mapred
RecordReader reads <key, value> pairs from an InputSplit.
RecordReader<KEYIN,VALUEIN> - Class in org.apache.hadoop.mapreduce
The record reader breaks the data into key/value pairs for input to the Mapper.
RecordReader() - Constructor for class org.apache.hadoop.mapreduce.RecordReader
 
RecordWriter<K,V> - Interface in org.apache.hadoop.mapred
RecordWriter writes the output <key, value> pairs to an output file.
RecordWriter<K,V> - Class in org.apache.hadoop.mapreduce
RecordWriter writes the output <key, value> pairs to an output file.
RecordWriter() - Constructor for class org.apache.hadoop.mapreduce.RecordWriter
 
redCmd_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
reduce(Shard, Iterator<IntermediateForm>, OutputCollector<Shard, IntermediateForm>, Reporter) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateCombiner
 
reduce(Shard, Iterator<IntermediateForm>, OutputCollector<Shard, Text>, Reporter) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateReducer
 
reduce(Object, Iterator, OutputCollector, Reporter) - Method in class org.apache.hadoop.contrib.utils.join.DataJoinMapperBase
 
reduce(Object, Iterator, OutputCollector, Reporter) - Method in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
 
reduce(LongWritable, Iterable<BytesWritable>, Reducer<LongWritable, BytesWritable, LongWritable, BytesWritable>.Context) - Method in class org.apache.hadoop.examples.BaileyBorweinPlouffe.BbpReducer
Concatenate map outputs.
reduce(IntWritable, Iterable<SummationWritable>, Reducer<IntWritable, SummationWritable, NullWritable, TaskResult>.Context) - Method in class org.apache.hadoop.examples.pi.DistSum.ReduceSide.SummingReducer
 
reduce(BooleanWritable, Iterable<LongWritable>, Reducer<BooleanWritable, LongWritable, WritableComparable<?>, Writable>.Context) - Method in class org.apache.hadoop.examples.QuasiMonteCarlo.QmcReducer
Accumulate number of points inside/outside results from the mappers.
reduce(SecondarySort.IntPair, Iterable<IntWritable>, Reducer<SecondarySort.IntPair, IntWritable, Text, IntWritable>.Context) - Method in class org.apache.hadoop.examples.SecondarySort.Reduce
 
reduce(Text, Iterable<IntWritable>, Reducer<Text, IntWritable, Text, IntWritable>.Context) - Method in class org.apache.hadoop.examples.WordCount.IntSumReducer
 
reduce(Text, Iterator<Text>, OutputCollector<Text, Text>, Reporter) - Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorCombiner
Combines values for a given key.
reduce(Text, Iterator<Text>, OutputCollector<Text, Text>, Reporter) - Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorMapper
Do nothing.
reduce(Text, Iterator<Text>, OutputCollector<Text, Text>, Reporter) - Method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorReducer
 
reduce(Object, Iterator, OutputCollector, Reporter) - Method in class org.apache.hadoop.mapred.lib.ChainReducer
Chains the reduce(...) method of the Reducer with the map(...) methods of the Mappers in the chain.
reduce(Text, Iterator<Text>, OutputCollector<Text, Text>, Reporter) - Method in class org.apache.hadoop.mapred.lib.FieldSelectionMapReduce
 
reduce(K, Iterator<V>, OutputCollector<K, V>, Reporter) - Method in class org.apache.hadoop.mapred.lib.IdentityReducer
Writes all keys and values directly to output.
reduce(K, Iterator<LongWritable>, OutputCollector<K, LongWritable>, Reporter) - Method in class org.apache.hadoop.mapred.lib.LongSumReducer
 
reduce(K2, Iterator<V2>, OutputCollector<K3, V3>, Reporter) - Method in interface org.apache.hadoop.mapred.Reducer
Reduces values for a given key.
reduce(Text, Iterable<Text>, Reducer<Text, Text, Text, Text>.Context) - Method in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorCombiner
Combines values for a given key.
reduce(Text, Iterable<Text>, Reducer<Text, Text, Text, Text>.Context) - Method in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorReducer
 
reduce(Text, Iterable<Text>, Reducer<Text, Text, Text, Text>.Context) - Method in class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionReducer
 
reduce(Key, Iterable<IntWritable>, Reducer<Key, IntWritable, Key, IntWritable>.Context) - Method in class org.apache.hadoop.mapreduce.lib.reduce.IntSumReducer
 
reduce(KEY, Iterable<LongWritable>, Reducer<KEY, LongWritable, KEY, LongWritable>.Context) - Method in class org.apache.hadoop.mapreduce.lib.reduce.LongSumReducer
 
reduce(KEYIN, Iterable<VALUEIN>, Reducer<KEYIN, VALUEIN, KEYOUT, VALUEOUT>.Context) - Method in class org.apache.hadoop.mapreduce.Reducer
This method is called once for each key.
reduce(Object, Iterator, OutputCollector, Reporter) - Method in class org.apache.hadoop.streaming.PipeReducer
 
REDUCE_OUTPUT_KEY_VALUE_SPEC - Static variable in class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionHelper
 
reduceContext - Variable in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
ReduceContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT> - Interface in org.apache.hadoop.mapreduce
The context passed to the Reducer.
ReduceContext.ValueIterator<VALUEIN> - Interface in org.apache.hadoop.mapreduce
Iterator to iterate over values for a given group of records.
ReduceContextImpl.ValueIterable - Class in org.apache.hadoop.mapreduce.task
 
ReduceContextImpl.ValueIterable() - Constructor for class org.apache.hadoop.mapreduce.task.ReduceContextImpl.ValueIterable
 
ReduceContextImpl.ValueIterator - Class in org.apache.hadoop.mapreduce.task
 
ReduceContextImpl.ValueIterator() - Constructor for class org.apache.hadoop.mapreduce.task.ReduceContextImpl.ValueIterator
 
reduceDebugSpec_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
reduceProgress() - Method in class org.apache.hadoop.mapred.JobStatus
 
reduceProgress() - Method in interface org.apache.hadoop.mapred.RunningJob
Get the progress of the job's reduce-tasks, as a float between 0.0 and 1.0.
reduceProgress() - Method in class org.apache.hadoop.mapreduce.Job
Get the progress of the job's reduce-tasks, as a float between 0.0 and 1.0.
Reducer<K2,V2,K3,V3> - Interface in org.apache.hadoop.mapred
Reduces a set of intermediate values which share a key to a smaller set of values.
Reducer<KEYIN,VALUEIN,KEYOUT,VALUEOUT> - Class in org.apache.hadoop.mapreduce
Reduces a set of intermediate values which share a key to a smaller set of values.
Reducer() - Constructor for class org.apache.hadoop.mapreduce.Reducer
 
Reducer.Context - Class in org.apache.hadoop.mapreduce
The Context passed on to the Reducer implementations.
Reducer.Context() - Constructor for class org.apache.hadoop.mapreduce.Reducer.Context
 
REDUCES_PER_HOST - Static variable in class org.apache.hadoop.examples.Join
 
REDUCES_PER_HOST - Static variable in class org.apache.hadoop.examples.Sort
 
RegexMapper<K> - Class in org.apache.hadoop.mapred.lib
A Mapper that extracts text matching a regular expression.
RegexMapper() - Constructor for class org.apache.hadoop.mapred.lib.RegexMapper
 
RegexMapper<K> - Class in org.apache.hadoop.mapreduce.lib.map
A Mapper that extracts text matching a regular expression.
RegexMapper() - Constructor for class org.apache.hadoop.mapreduce.lib.map.RegexMapper
 
regexpEscape(String) - Static method in class org.apache.hadoop.streaming.StreamUtil
 
remainingTerms(List<T>) - Method in class org.apache.hadoop.examples.pi.math.Summation
Find the remaining terms.
remove() - Method in class org.apache.hadoop.contrib.utils.join.ArrayListBackedIterator
 
remove() - Method in class org.apache.hadoop.mapreduce.MarkableIterator
 
remove() - Method in class org.apache.hadoop.mapreduce.task.ReduceContextImpl.ValueIterator
 
renameFile(String, String) - Method in class org.apache.hadoop.contrib.index.lucene.FileSystemDirectory
 
renewDelegationToken(Token<DelegationTokenIdentifier>) - Method in class org.apache.hadoop.mapred.JobClient
Renew a delegation token
renewDelegationToken(Token<DelegationTokenIdentifier>) - Method in class org.apache.hadoop.mapreduce.Cluster
Renew a delegation token
replay(TupleWritable) - Method in class org.apache.hadoop.mapred.join.JoinRecordReader.JoinDelegationIterator
 
replay(V) - Method in class org.apache.hadoop.mapred.join.MultiFilterRecordReader.MultiFilterDelegationIterator
 
replay(X) - Method in class org.apache.hadoop.mapreduce.lib.join.ArrayListBackedIterator
 
replay(TupleWritable) - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader.JoinCollector
Replay the last Tuple emitted.
replay(TupleWritable) - Method in class org.apache.hadoop.mapreduce.lib.join.JoinRecordReader.JoinDelegationIterator
 
replay(V) - Method in class org.apache.hadoop.mapreduce.lib.join.MultiFilterRecordReader.MultiFilterDelegationIterator
 
replay(U) - Method in class org.apache.hadoop.mapreduce.lib.join.ResetableIterator.EMPTY
 
replay(T) - Method in interface org.apache.hadoop.mapreduce.lib.join.ResetableIterator
Assign last value returned to actual.
replay(X) - Method in class org.apache.hadoop.mapreduce.lib.join.StreamBackedIterator
 
report() - Method in class org.apache.hadoop.contrib.utils.join.JobBase
log the counters
reporter - Variable in class org.apache.hadoop.contrib.utils.join.DataJoinMapperBase
 
reporter - Variable in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
 
reporter - Variable in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
 
Reporter - Interface in org.apache.hadoop.mapred
A facility for Map-Reduce applications to report progress and update counters, status information etc.
reset() - Method in class org.apache.hadoop.contrib.utils.join.ArrayListBackedIterator
 
reset() - Method in interface org.apache.hadoop.contrib.utils.join.ResetableIterator
 
reset() - Method in class org.apache.hadoop.mapred.join.JoinRecordReader.JoinDelegationIterator
 
reset() - Method in class org.apache.hadoop.mapred.join.MultiFilterRecordReader.MultiFilterDelegationIterator
 
reset() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.DoubleValueSum
reset the aggregator
reset() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.LongValueMax
reset the aggregator
reset() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.LongValueMin
reset the aggregator
reset() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.LongValueSum
reset the aggregator
reset() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.StringValueMax
reset the aggregator
reset() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.StringValueMin
reset the aggregator
reset() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.UniqValueCount
reset the aggregator
reset() - Method in interface org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregator
reset the aggregator
reset() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.ValueHistogram
reset the aggregator
reset() - Method in class org.apache.hadoop.mapreduce.lib.join.ArrayListBackedIterator
 
reset(K) - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader.JoinCollector
Codify the contents of the collector to be iterated over.
reset() - Method in class org.apache.hadoop.mapreduce.lib.join.JoinRecordReader.JoinDelegationIterator
 
reset() - Method in class org.apache.hadoop.mapreduce.lib.join.MultiFilterRecordReader.MultiFilterDelegationIterator
 
reset() - Method in class org.apache.hadoop.mapreduce.lib.join.ResetableIterator.EMPTY
 
reset() - Method in interface org.apache.hadoop.mapreduce.lib.join.ResetableIterator
Set iterator to return to the start of its range.
reset() - Method in class org.apache.hadoop.mapreduce.lib.join.StreamBackedIterator
 
reset(BytesWritable) - Method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat.WritableValueBytes
 
reset() - Method in class org.apache.hadoop.mapreduce.MarkableIterator
 
reset() - Method in class org.apache.hadoop.mapreduce.task.ReduceContextImpl.ValueIterator
 
ResetableIterator - Interface in org.apache.hadoop.contrib.utils.join
This defines an iterator interface that will help the reducer class re-group its input by source tags.
ResetableIterator<T extends org.apache.hadoop.io.Writable> - Interface in org.apache.hadoop.mapred.join
This defines an interface to a stateful Iterator that can replay elements added to it directly.
ResetableIterator<T extends org.apache.hadoop.io.Writable> - Interface in org.apache.hadoop.mapreduce.lib.join
This defines an interface to a stateful Iterator that can replay elements added to it directly.
ResetableIterator.EMPTY<U extends org.apache.hadoop.io.Writable> - Class in org.apache.hadoop.mapred.join
 
ResetableIterator.EMPTY() - Constructor for class org.apache.hadoop.mapred.join.ResetableIterator.EMPTY
 
ResetableIterator.EMPTY<U extends org.apache.hadoop.io.Writable> - Class in org.apache.hadoop.mapreduce.lib.join
 
ResetableIterator.EMPTY() - Constructor for class org.apache.hadoop.mapreduce.lib.join.ResetableIterator.EMPTY
 
resetBackupStore() - Method in interface org.apache.hadoop.mapreduce.ReduceContext.ValueIterator
This method is called when the reducer moves from one key to another.
resetBackupStore() - Method in class org.apache.hadoop.mapreduce.task.ReduceContextImpl.ValueIterator
This method is called when the reducer moves from one key to another.
resolve(String) - Method in class org.apache.hadoop.streaming.io.IdentifierResolver
Resolves a given identifier.
resume() - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl
resume the suspended thread
RIGHT_OFFSET_PROPERTY_NAME - Static variable in class org.apache.hadoop.mapreduce.lib.partition.BinaryPartitioner
 
rjustify(String, int) - Static method in class org.apache.hadoop.streaming.StreamUtil
 
RoundRobinDistributionPolicy - Class in org.apache.hadoop.contrib.index.example
Choose a shard for each insert in a round-robin fashion.
RoundRobinDistributionPolicy() - Constructor for class org.apache.hadoop.contrib.index.example.RoundRobinDistributionPolicy
 
rrClass - Variable in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
 
rrClass - Variable in class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader
 
rrConstructor - Variable in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
 
rrConstructor - Variable in class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader
 
rrCstrMap - Static variable in class org.apache.hadoop.mapred.join.Parser.Node
 
rrCstrMap - Static variable in class org.apache.hadoop.mapreduce.lib.join.Parser.Node
 
run(Configuration, Path[], Path, int, Shard[]) - Method in interface org.apache.hadoop.contrib.index.mapred.IIndexUpdater
Create a Map/Reduce job configuration and run the Map/Reduce job to analyze documents and update Lucene instances in parallel.
run(Configuration, Path[], Path, int, Shard[]) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdater
 
run(String[]) - Method in class org.apache.hadoop.examples.BaileyBorweinPlouffe
Parse arguments and then runs a map/reduce job.
run(String[]) - Method in class org.apache.hadoop.examples.dancing.DistributedPentomino
 
run(String[]) - Method in class org.apache.hadoop.examples.DBCountPageView
 
run(String[]) - Method in class org.apache.hadoop.examples.Grep
 
run(String[]) - Method in class org.apache.hadoop.examples.Join
The main driver for sort program.
run(String[]) - Method in class org.apache.hadoop.examples.MultiFileWordCount
 
run(String[]) - Method in class org.apache.hadoop.examples.pi.DistBbp
run(String[]) - Method in class org.apache.hadoop.examples.pi.DistSum
run(String[]) - Method in class org.apache.hadoop.examples.QuasiMonteCarlo
Parse arguments and then runs a map/reduce job.
run(String[]) - Method in class org.apache.hadoop.examples.RandomTextWriter
This is the main routine for launching a distributed random write job.
run(String[]) - Method in class org.apache.hadoop.examples.RandomWriter
This is the main routine for launching a distributed random write job.
run(String[]) - Method in class org.apache.hadoop.examples.Sort
The main driver for sort program.
run(String[]) - Method in class org.apache.hadoop.examples.terasort.TeraChecksum
 
run(String[]) - Method in class org.apache.hadoop.examples.terasort.TeraGen
 
run(String[]) - Method in class org.apache.hadoop.examples.terasort.TeraSort
 
run(String[]) - Method in class org.apache.hadoop.examples.terasort.TeraValidate
 
run(String[]) - Method in class org.apache.hadoop.mapred.gridmix.Gridmix
 
run(RecordReader<K1, V1>, OutputCollector<K2, V2>, Reporter) - Method in class org.apache.hadoop.mapred.lib.MultithreadedMapRunner
 
run(RecordReader<K1, V1>, OutputCollector<K2, V2>, Reporter) - Method in interface org.apache.hadoop.mapred.MapRunnable
Start mapping input <key, value> pairs.
run(RecordReader<K1, V1>, OutputCollector<K2, V2>, Reporter) - Method in class org.apache.hadoop.mapred.MapRunner
 
run(String[]) - Method in class org.apache.hadoop.mapred.pipes.Submitter
 
run(String[]) - Method in class org.apache.hadoop.mapred.tools.MRAdmin
 
run(Mapper<KEYIN, VALUEIN, KEYOUT, VALUEOUT>.Context) - Method in class org.apache.hadoop.mapreduce.lib.chain.ChainMapper
 
run(Reducer<KEYIN, VALUEIN, KEYOUT, VALUEOUT>.Context) - Method in class org.apache.hadoop.mapreduce.lib.chain.ChainReducer
 
run() - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl
The main loop for the thread.
run(Mapper<K1, V1, K2, V2>.Context) - Method in class org.apache.hadoop.mapreduce.lib.map.MultithreadedMapper
Run the application's maps using a thread pool.
run(String[]) - Method in class org.apache.hadoop.mapreduce.lib.partition.InputSampler
Driver for InputSampler from the command line.
run(Mapper<KEYIN, VALUEIN, KEYOUT, VALUEOUT>.Context) - Method in class org.apache.hadoop.mapreduce.Mapper
Expert users can override this method for more complete control over the execution of the Mapper.
run(Reducer<KEYIN, VALUEIN, KEYOUT, VALUEOUT>.Context) - Method in class org.apache.hadoop.mapreduce.Reducer
Advanced application writers can use the Reducer.run(org.apache.hadoop.mapreduce.Reducer.Context) method to control how the reduce task works.
run(String[]) - Method in class org.apache.hadoop.mapreduce.tools.CLI
 
run(String[]) - Method in class org.apache.hadoop.streaming.DumpTypedBytes
The main driver for DumpTypedBytes.
run(String[]) - Method in class org.apache.hadoop.streaming.LoadTypedBytes
The main driver for LoadTypedBytes.
run(RecordReader<K1, V1>, OutputCollector<K2, V2>, Reporter) - Method in class org.apache.hadoop.streaming.PipeMapRunner
 
run(String[]) - Method in class org.apache.hadoop.streaming.StreamJob
 
runJob(JobConf) - Static method in class org.apache.hadoop.contrib.utils.join.DataJoinJob
Submit/run a map/reduce job.
runJob(JobConf) - Static method in class org.apache.hadoop.mapred.JobClient
Utility that submits a job, then polls for progress until the job is complete.
runJob(JobConf) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
Submit a job to the map/reduce cluster.
RUNNING - Static variable in class org.apache.hadoop.mapred.jobcontrol.Job
 
RUNNING - Static variable in class org.apache.hadoop.mapred.JobStatus
 
running_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
RunningJob - Interface in org.apache.hadoop.mapred
RunningJob is the user-interface to query for details on a running Map-Reduce job.

S

safeGetCanonicalPath(File) - Static method in class org.apache.hadoop.streaming.StreamUtil
 
SecondarySort - Class in org.apache.hadoop.examples
This is an example Hadoop Map/Reduce application.
SecondarySort() - Constructor for class org.apache.hadoop.examples.SecondarySort
 
SecondarySort.FirstGroupingComparator - Class in org.apache.hadoop.examples
Compare only the first part of the pair, so that reduce is called once for each value of the first part.
SecondarySort.FirstGroupingComparator() - Constructor for class org.apache.hadoop.examples.SecondarySort.FirstGroupingComparator
 
SecondarySort.FirstPartitioner - Class in org.apache.hadoop.examples
Partition based on the first part of the pair.
SecondarySort.FirstPartitioner() - Constructor for class org.apache.hadoop.examples.SecondarySort.FirstPartitioner
 
SecondarySort.IntPair - Class in org.apache.hadoop.examples
Define a pair of integers that are writable.
SecondarySort.IntPair() - Constructor for class org.apache.hadoop.examples.SecondarySort.IntPair
 
SecondarySort.IntPair.Comparator - Class in org.apache.hadoop.examples
A Comparator that compares serialized IntPair.
SecondarySort.IntPair.Comparator() - Constructor for class org.apache.hadoop.examples.SecondarySort.IntPair.Comparator
 
SecondarySort.MapClass - Class in org.apache.hadoop.examples
Read two integers from each line and generate a key, value pair as ((left, right), right).
SecondarySort.MapClass() - Constructor for class org.apache.hadoop.examples.SecondarySort.MapClass
 
SecondarySort.Reduce - Class in org.apache.hadoop.examples
A reducer class that just emits the sum of the input values.
SecondarySort.Reduce() - Constructor for class org.apache.hadoop.examples.SecondarySort.Reduce
 
seek(long) - Method in class org.apache.hadoop.mapred.SequenceFileRecordReader
 
seekNextRecordBoundary() - Method in class org.apache.hadoop.streaming.StreamBaseRecordReader
Implementation should seek forward in_ to the first byte of the next record.
seekNextRecordBoundary() - Method in class org.apache.hadoop.streaming.StreamXmlRecordReader
 
seenPrimary_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
SEPARATOR - Static variable in class org.apache.hadoop.mapreduce.ID
 
SEPERATOR - Static variable in class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat
 
SequenceFileAsBinaryInputFormat - Class in org.apache.hadoop.mapred
InputFormat reading keys, values from SequenceFiles in binary (raw) format.
SequenceFileAsBinaryInputFormat() - Constructor for class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat
 
SequenceFileAsBinaryInputFormat - Class in org.apache.hadoop.mapreduce.lib.input
InputFormat reading keys, values from SequenceFiles in binary (raw) format.
SequenceFileAsBinaryInputFormat() - Constructor for class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsBinaryInputFormat
 
SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader - Class in org.apache.hadoop.mapred
Read records from a SequenceFile as binary (raw) bytes.
SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader(Configuration, FileSplit) - Constructor for class org.apache.hadoop.mapred.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
 
SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader - Class in org.apache.hadoop.mapreduce.lib.input
Read records from a SequenceFile as binary (raw) bytes.
SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader() - Constructor for class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsBinaryInputFormat.SequenceFileAsBinaryRecordReader
 
SequenceFileAsBinaryOutputFormat - Class in org.apache.hadoop.mapred
An OutputFormat that writes keys, values to SequenceFiles in binary(raw) format
SequenceFileAsBinaryOutputFormat() - Constructor for class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat
 
SequenceFileAsBinaryOutputFormat - Class in org.apache.hadoop.mapreduce.lib.output
An OutputFormat that writes keys, values to SequenceFiles in binary(raw) format
SequenceFileAsBinaryOutputFormat() - Constructor for class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat
 
SequenceFileAsBinaryOutputFormat.WritableValueBytes - Class in org.apache.hadoop.mapred
Inner class used for appendRaw
SequenceFileAsBinaryOutputFormat.WritableValueBytes() - Constructor for class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat.WritableValueBytes
 
SequenceFileAsBinaryOutputFormat.WritableValueBytes - Class in org.apache.hadoop.mapreduce.lib.output
Inner class used for appendRaw
SequenceFileAsBinaryOutputFormat.WritableValueBytes() - Constructor for class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat.WritableValueBytes
 
SequenceFileAsBinaryOutputFormat.WritableValueBytes(BytesWritable) - Constructor for class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat.WritableValueBytes
 
SequenceFileAsTextInputFormat - Class in org.apache.hadoop.mapred
This class is similar to SequenceFileInputFormat, except it generates SequenceFileAsTextRecordReader which converts the input keys and values to their String forms by calling toString() method.
SequenceFileAsTextInputFormat() - Constructor for class org.apache.hadoop.mapred.SequenceFileAsTextInputFormat
 
SequenceFileAsTextInputFormat - Class in org.apache.hadoop.mapreduce.lib.input
This class is similar to SequenceFileInputFormat, except it generates SequenceFileAsTextRecordReader which converts the input keys and values to their String forms by calling toString() method.
SequenceFileAsTextInputFormat() - Constructor for class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsTextInputFormat
 
SequenceFileAsTextRecordReader - Class in org.apache.hadoop.mapred
This class converts the input keys and values to their String forms by calling toString() method.
SequenceFileAsTextRecordReader(Configuration, FileSplit) - Constructor for class org.apache.hadoop.mapred.SequenceFileAsTextRecordReader
 
SequenceFileAsTextRecordReader - Class in org.apache.hadoop.mapreduce.lib.input
This class converts the input keys and values to their String forms by calling toString() method.
SequenceFileAsTextRecordReader() - Constructor for class org.apache.hadoop.mapreduce.lib.input.SequenceFileAsTextRecordReader
 
SequenceFileInputFilter<K,V> - Class in org.apache.hadoop.mapred
A class that allows a map/red job to work on a sample of sequence files.
SequenceFileInputFilter() - Constructor for class org.apache.hadoop.mapred.SequenceFileInputFilter
 
SequenceFileInputFilter<K,V> - Class in org.apache.hadoop.mapreduce.lib.input
A class that allows a map/red job to work on a sample of sequence files.
SequenceFileInputFilter() - Constructor for class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter
 
SequenceFileInputFilter.Filter - Interface in org.apache.hadoop.mapred
filter interface
SequenceFileInputFilter.Filter - Interface in org.apache.hadoop.mapreduce.lib.input
filter interface
SequenceFileInputFilter.FilterBase - Class in org.apache.hadoop.mapred
base class for Filters
SequenceFileInputFilter.FilterBase() - Constructor for class org.apache.hadoop.mapred.SequenceFileInputFilter.FilterBase
 
SequenceFileInputFilter.FilterBase - Class in org.apache.hadoop.mapreduce.lib.input
base class for Filters
SequenceFileInputFilter.FilterBase() - Constructor for class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.FilterBase
 
SequenceFileInputFilter.MD5Filter - Class in org.apache.hadoop.mapred
This class returns a set of records by examing the MD5 digest of its key against a filtering frequency f.
SequenceFileInputFilter.MD5Filter() - Constructor for class org.apache.hadoop.mapred.SequenceFileInputFilter.MD5Filter
 
SequenceFileInputFilter.MD5Filter - Class in org.apache.hadoop.mapreduce.lib.input
This class returns a set of records by examing the MD5 digest of its key against a filtering frequency f.
SequenceFileInputFilter.MD5Filter() - Constructor for class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.MD5Filter
 
SequenceFileInputFilter.PercentFilter - Class in org.apache.hadoop.mapred
This class returns a percentage of records The percentage is determined by a filtering frequency f using the criteria record# % f == 0.
SequenceFileInputFilter.PercentFilter() - Constructor for class org.apache.hadoop.mapred.SequenceFileInputFilter.PercentFilter
 
SequenceFileInputFilter.PercentFilter - Class in org.apache.hadoop.mapreduce.lib.input
This class returns a percentage of records The percentage is determined by a filtering frequency f using the criteria record# % f == 0.
SequenceFileInputFilter.PercentFilter() - Constructor for class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.PercentFilter
 
SequenceFileInputFilter.RegexFilter - Class in org.apache.hadoop.mapred
Records filter by matching key to regex
SequenceFileInputFilter.RegexFilter() - Constructor for class org.apache.hadoop.mapred.SequenceFileInputFilter.RegexFilter
 
SequenceFileInputFilter.RegexFilter - Class in org.apache.hadoop.mapreduce.lib.input
Records filter by matching key to regex
SequenceFileInputFilter.RegexFilter() - Constructor for class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.RegexFilter
 
SequenceFileInputFormat<K,V> - Class in org.apache.hadoop.mapred
An InputFormat for SequenceFiles.
SequenceFileInputFormat() - Constructor for class org.apache.hadoop.mapred.SequenceFileInputFormat
 
SequenceFileInputFormat<K,V> - Class in org.apache.hadoop.mapreduce.lib.input
An InputFormat for SequenceFiles.
SequenceFileInputFormat() - Constructor for class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFormat
 
SequenceFileOutputFormat<K,V> - Class in org.apache.hadoop.mapred
An OutputFormat that writes SequenceFiles.
SequenceFileOutputFormat() - Constructor for class org.apache.hadoop.mapred.SequenceFileOutputFormat
 
SequenceFileOutputFormat<K,V> - Class in org.apache.hadoop.mapreduce.lib.output
An OutputFormat that writes SequenceFiles.
SequenceFileOutputFormat() - Constructor for class org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat
 
SequenceFileRecordReader<K,V> - Class in org.apache.hadoop.mapred
An RecordReader for SequenceFiles.
SequenceFileRecordReader(Configuration, FileSplit) - Constructor for class org.apache.hadoop.mapred.SequenceFileRecordReader
 
SequenceFileRecordReader<K,V> - Class in org.apache.hadoop.mapreduce.lib.input
An RecordReader for SequenceFiles.
SequenceFileRecordReader() - Constructor for class org.apache.hadoop.mapreduce.lib.input.SequenceFileRecordReader
 
SerialJobFactory - Class in org.apache.hadoop.mapred.gridmix
 
SerialJobFactory(JobSubmitter, JobStoryProducer, Path, Configuration, CountDownLatch) - Constructor for class org.apache.hadoop.mapred.gridmix.SerialJobFactory
Creating a new instance does not start the thread.
SESSION_TIMEZONE_KEY - Static variable in class org.apache.hadoop.mapreduce.lib.db.OracleDBRecordReader
Configuration key to set to a timezone string.
set(int, int) - Method in class org.apache.hadoop.examples.SecondarySort.IntPair
Set the left and right values.
setAggregatorDescriptors(JobConf, Class<? extends ValueAggregatorDescriptor>[]) - Static method in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJob
 
setAggregatorDescriptors(Class<? extends ValueAggregatorDescriptor>[]) - Static method in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorJob
 
setAssignedJobID(JobID) - Method in class org.apache.hadoop.mapred.jobcontrol.Job
Deprecated. setAssignedJobID should not be called. JOBID is set by the framework.
setAttemptsToStartSkipping(Configuration, int) - Static method in class org.apache.hadoop.mapred.SkipBadRecords
Set the number of Task attempts AFTER which skip mode will be kicked off.
setAutoIncrMapperProcCount(Configuration, boolean) - Static method in class org.apache.hadoop.mapred.SkipBadRecords
Set the flag which if set to true, SkipBadRecords.COUNTER_MAP_PROCESSED_RECORDS is incremented by MapRunner after invoking the map function.
setAutoIncrReducerProcCount(Configuration, boolean) - Static method in class org.apache.hadoop.mapred.SkipBadRecords
Set the flag which if set to true, SkipBadRecords.COUNTER_REDUCE_PROCESSED_GROUPS is incremented by framework after invoking the reduce function.
setBoundingQuery(Configuration, String) - Static method in class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat
Set the user-defined bounding query to use with a user-defined query.
setCacheArchives(URI[]) - Method in class org.apache.hadoop.mapreduce.Job
Set the given set of archives
setCacheFiles(URI[]) - Method in class org.apache.hadoop.mapreduce.Job
Set the given set of files
setCancelDelegationTokenUponJobCompletion(boolean) - Method in class org.apache.hadoop.mapreduce.Job
Sets the flag that will allow the JobTracker to cancel the HDFS delegation tokens upon job completion.
setChildren(List<JobQueueInfo>) - Method in class org.apache.hadoop.mapred.JobQueueInfo
 
setCleanupProgress(float) - Method in class org.apache.hadoop.mapred.JobStatus
Sets the cleanup progress of this job
setCleanupProgress(float) - Method in class org.apache.hadoop.mapreduce.JobStatus
Sets the cleanup progress of this job
setCombinerClass(Class<? extends Reducer>) - Method in class org.apache.hadoop.mapred.JobConf
Set the user-defined combiner class used to combine map-outputs before being sent to the reducers.
setCombinerClass(Class<? extends Reducer>) - Method in class org.apache.hadoop.mapreduce.Job
Set the combiner class for the job.
setCompressMapOutput(boolean) - Method in class org.apache.hadoop.mapred.JobConf
Should the map outputs be compressed before transfer? Uses the SequenceFile compression.
setCompressOutput(JobConf, boolean) - Static method in class org.apache.hadoop.mapred.FileOutputFormat
Set whether the output of the job is compressed.
setCompressOutput(Job, boolean) - Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
Set whether the output of the job is compressed.
setConf(Configuration) - Method in class org.apache.hadoop.examples.terasort.TeraSort.SimplePartitioner
 
setConf(Configuration) - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
setConf(Configuration) - Method in class org.apache.hadoop.mapred.SequenceFileInputFilter.MD5Filter
configure the filter according to configuration
setConf(Configuration) - Method in class org.apache.hadoop.mapred.SequenceFileInputFilter.PercentFilter
configure the filter by checking the configuration
setConf(Configuration) - Method in class org.apache.hadoop.mapred.SequenceFileInputFilter.RegexFilter
configure the Filter by checking the configuration
setConf(Configuration) - Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat
setConf(Configuration) - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.MD5Filter
configure the filter according to configuration
setConf(Configuration) - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.PercentFilter
configure the filter by checking the configuration
setConf(Configuration) - Method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.RegexFilter
configure the Filter by checking the configuration
setConf(Configuration) - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
setConf(Configuration) - Method in class org.apache.hadoop.mapreduce.lib.partition.BinaryPartitioner
 
setConf(Configuration) - Method in class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedComparator
 
setConf(Configuration) - Method in class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedPartitioner
 
setConf(Configuration) - Method in class org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner
Read in the partition file and build indexing data structures.
setConf(Configuration) - Method in class org.apache.hadoop.streaming.DumpTypedBytes
 
setConf(Configuration) - Method in class org.apache.hadoop.streaming.LoadTypedBytes
 
setConf(Configuration) - Method in class org.apache.hadoop.streaming.StreamJob
 
setConf(Configuration) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableInput
 
setCountersEnabled(JobConf, boolean) - Static method in class org.apache.hadoop.mapred.lib.MultipleOutputs
Enables or disables counters for the named outputs.
setCountersEnabled(Job, boolean) - Static method in class org.apache.hadoop.mapreduce.lib.output.MultipleOutputs
Enables or disables counters for the named outputs.
setDelete(Term) - Method in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
Set the instance to be a delete operation.
setDisplayName(String) - Method in class org.apache.hadoop.mapred.Counters.Counter
 
setDisplayName(String) - Method in class org.apache.hadoop.mapred.Counters.Group
Set the display name
setDisplayName(String) - Method in class org.apache.hadoop.mapreduce.Counter
Deprecated. 
setDistributionPolicyClass(Class<? extends IDistributionPolicy>) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Set the distribution policy class.
setDocumentAnalyzerClass(Class<? extends Analyzer>) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Set the analyzer class.
setDoubleValue(Object, double) - Method in class org.apache.hadoop.contrib.utils.join.JobBase
Set the given counter to the given value
setEventId(int) - Method in class org.apache.hadoop.mapred.TaskCompletionEvent
set event Id.
setEventId(int) - Method in class org.apache.hadoop.mapreduce.TaskCompletionEvent
set event Id.
setExecutable(JobConf, String) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
Set the URI for the application's executable.
setFilterClass(Configuration, Class) - Static method in class org.apache.hadoop.mapred.SequenceFileInputFilter
set the filter class
setFilterClass(Job, Class<?>) - Static method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter
set the filter class
setFinishTime(long) - Method in class org.apache.hadoop.mapred.JobStatus
Set the finish time of the job
setFinishTime(long) - Method in class org.apache.hadoop.mapred.TaskReport
set finish time of task.
setFinishTime(long) - Method in class org.apache.hadoop.mapreduce.JobStatus
Set the finish time of the job
setFormat(JobConf) - Method in class org.apache.hadoop.mapred.join.CompositeInputFormat
Interpret a given string as a composite expression.
setFormat(Configuration) - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeInputFormat
Interpret a given string as a composite expression.
setFrequency(Configuration, int) - Static method in class org.apache.hadoop.mapred.SequenceFileInputFilter.MD5Filter
set the filtering frequency in configuration
setFrequency(Configuration, int) - Static method in class org.apache.hadoop.mapred.SequenceFileInputFilter.PercentFilter
set the frequency and stores it in conf
setFrequency(Configuration, int) - Static method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.MD5Filter
set the filtering frequency in configuration
setFrequency(Configuration, int) - Static method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.PercentFilter
set the frequency and stores it in conf
setGroupingComparatorClass(Class<? extends RawComparator>) - Method in class org.apache.hadoop.mapreduce.Job
Define the comparator that controls which keys are grouped together for a single call to Reducer.reduce(Object, Iterable, org.apache.hadoop.mapreduce.Reducer.Context)
setHistoryFile(String) - Method in class org.apache.hadoop.mapred.JobStatus
Set the job history file url for a completed job
setHistoryFile(String) - Method in class org.apache.hadoop.mapreduce.JobStatus
Set the job history file url for a completed job
setID(int) - Method in class org.apache.hadoop.mapred.join.Parser.Node
 
setID(int) - Method in class org.apache.hadoop.mapreduce.lib.join.Parser.Node
 
setIndexInputFormatClass(Class<? extends InputFormat>) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Set the index input format class.
setIndexMaxFieldLength(int) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Set the max field length for a Lucene instance.
setIndexMaxNumSegments(int) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Set the max number of segments for a Lucene instance.
setIndexShards(String) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Set the string representation of a number of shards.
setIndexShards(IndexUpdateConfiguration, Shard[]) - Static method in class org.apache.hadoop.contrib.index.mapred.Shard
 
setIndexUpdaterClass(Class<? extends IIndexUpdater>) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Set the index updater class.
setIndexUseCompoundFile(boolean) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Set whether use the compound file format for a Lucene instance.
setInput(JobConf, Class<? extends DBWritable>, String, String, String, String...) - Static method in class org.apache.hadoop.mapred.lib.db.DBInputFormat
Initializes the map-part of the job with the appropriate input settings.
setInput(JobConf, Class<? extends DBWritable>, String, String) - Static method in class org.apache.hadoop.mapred.lib.db.DBInputFormat
Initializes the map-part of the job with the appropriate input settings.
setInput(Job, Class<? extends DBWritable>, String, String, String, String...) - Static method in class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat
Note that the "orderBy" column is called the "splitBy" in this version.
setInput(Job, Class<? extends DBWritable>, String, String) - Static method in class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat
setInput() takes a custom query and a separate "bounding query" to use instead of the custom "count query" used by DBInputFormat.
setInput(Job, Class<? extends DBWritable>, String, String, String, String...) - Static method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat
Initializes the map-part of the job with the appropriate input settings.
setInput(Job, Class<? extends DBWritable>, String, String) - Static method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat
Initializes the map-part of the job with the appropriate input settings.
setInputBoundingQuery(String) - Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
 
setInputClass(Class<? extends DBWritable>) - Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
 
setInputConditions(String) - Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
 
setInputCountQuery(String) - Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
 
setInputDataLength(long) - Method in class org.apache.hadoop.mapreduce.split.JobSplit.SplitMetaInfo
 
setInputDataLocations(String[]) - Method in class org.apache.hadoop.mapreduce.split.JobSplit.SplitMetaInfo
 
setInputFieldNames(String...) - Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
 
setInputFormat(Class<? extends InputFormat>) - Method in class org.apache.hadoop.mapred.JobConf
Set the InputFormat implementation for the map-reduce job.
setInputFormatClass(Class<? extends InputFormat>) - Method in class org.apache.hadoop.mapreduce.Job
Set the InputFormat for the job.
setInputOrderBy(String) - Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
 
setInputPathFilter(JobConf, Class<? extends PathFilter>) - Static method in class org.apache.hadoop.mapred.FileInputFormat
Set a PathFilter to be applied to the input paths for the map-reduce job.
setInputPathFilter(Job, Class<? extends PathFilter>) - Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
Set a PathFilter to be applied to the input paths for the map-reduce job.
setInputPaths(JobConf, String) - Static method in class org.apache.hadoop.mapred.FileInputFormat
Sets the given comma separated paths as the list of inputs for the map-reduce job.
setInputPaths(JobConf, Path...) - Static method in class org.apache.hadoop.mapred.FileInputFormat
Set the array of Paths as the list of inputs for the map-reduce job.
setInputPaths(Job, String) - Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
Sets the given comma separated paths as the list of inputs for the map-reduce job.
setInputPaths(Job, Path...) - Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
Set the array of Paths as the list of inputs for the map-reduce job.
setInputQuery(String) - Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
 
setInputTableName(String) - Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
 
setInputWriterClass(Class<? extends InputWriter>) - Method in class org.apache.hadoop.streaming.io.IdentifierResolver
Sets the InputWriter class.
setInsert(Document) - Method in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
Set the instance to be an insert operation.
setIOSortMB(int) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Set the IO sort space in MB.
setIsCleanup(boolean) - Method in class org.apache.hadoop.mapred.TaskLogAppender
Set whether the task is a cleanup attempt or not.
setIsJavaMapper(JobConf, boolean) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
Set whether the Mapper is written in Java.
setIsJavaRecordReader(JobConf, boolean) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
Set whether the job is using a Java RecordReader.
setIsJavaRecordWriter(JobConf, boolean) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
Set whether the job will use a Java RecordWriter.
setIsJavaReducer(JobConf, boolean) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
Set whether the Reducer is written in Java.
setJar(String) - Method in class org.apache.hadoop.mapred.JobConf
Set the user jar for the map-reduce job.
setJar(String) - Method in class org.apache.hadoop.mapreduce.Job
Set the job jar
setJarByClass(Class) - Method in class org.apache.hadoop.mapred.JobConf
Set the job's jar file by finding an example class location.
setJarByClass(Class<?>) - Method in class org.apache.hadoop.mapreduce.Job
Set the Jar by finding where a given class came from.
setJob(Job) - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
Set the mapreduce job
setJobACLs(Map<JobACL, AccessControlList>) - Method in class org.apache.hadoop.mapred.JobStatus
 
setJobACLs(Map<JobACL, AccessControlList>) - Method in class org.apache.hadoop.mapreduce.JobStatus
 
setJobConf(JobConf) - Method in class org.apache.hadoop.mapred.jobcontrol.Job
Set the mapred job conf for this job.
setJobConf() - Method in class org.apache.hadoop.streaming.StreamJob
 
setJobEndNotificationURI(String) - Method in class org.apache.hadoop.mapred.JobConf
Set the uri to be invoked in-order to send a notification after the job has completed (success/failure).
setJobID(String) - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
Set the job ID for this job.
setJobName(String) - Method in class org.apache.hadoop.mapred.JobConf
Set the user-specified job name.
setJobName(String) - Method in class org.apache.hadoop.mapreduce.Job
Set the user-specified job name.
setJobName(String) - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
Set the job name for this job.
setJobPriority(JobPriority) - Method in class org.apache.hadoop.mapred.JobConf
Set JobPriority for this job.
setJobPriority(JobPriority) - Method in class org.apache.hadoop.mapred.JobStatus
Set the priority of the job, defaulting to NORMAL.
setJobPriority(String) - Method in interface org.apache.hadoop.mapred.RunningJob
Set the priority of a running job.
setJobSetupCleanupNeeded(boolean) - Method in class org.apache.hadoop.mapreduce.Job
Specify whether job-setup and job-cleanup is needed for the job
setJobState(ControlledJob.State) - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
Set the state for this job.
setJobStatuses(JobStatus[]) - Method in class org.apache.hadoop.mapred.JobQueueInfo
 
setJobStatuses(JobStatus[]) - Method in class org.apache.hadoop.mapreduce.QueueInfo
 
setJobToken(Token<? extends TokenIdentifier>, TokenStorage) - Static method in class org.apache.hadoop.mapreduce.security.TokenCache
store job token
setKeepCommandFile(JobConf, boolean) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
Set whether to keep the command file for debugging
setKeepFailedTaskFiles(boolean) - Method in class org.apache.hadoop.mapred.JobConf
Set whether the framework should keep the intermediate files for failed tasks.
setKeepTaskFilesPattern(String) - Method in class org.apache.hadoop.mapred.JobConf
Set a regular expression for task names that should be kept.
setKeyComparator(Class<? extends WritableComparator>) - Method in class org.apache.hadoop.mapred.join.Parser.Node
 
setKeyComparator(Class<? extends WritableComparator>) - Method in class org.apache.hadoop.mapreduce.lib.join.Parser.Node
 
setKeyFieldComparatorOptions(String) - Method in class org.apache.hadoop.mapred.JobConf
Set the KeyFieldBasedComparator options used to compare keys.
setKeyFieldComparatorOptions(Job, String) - Static method in class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedComparator
Set the KeyFieldBasedComparator options used to compare keys.
setKeyFieldPartitionerOptions(String) - Method in class org.apache.hadoop.mapred.JobConf
Set the KeyFieldBasedPartitioner options used for Partitioner
setKeyFieldPartitionerOptions(Job, String) - Method in class org.apache.hadoop.mapreduce.lib.partition.KeyFieldBasedPartitioner
Set the KeyFieldBasedPartitioner options used for Partitioner
setKeyValue(Text, Text, byte[], int, int) - Static method in class org.apache.hadoop.mapreduce.lib.input.KeyValueLineRecordReader
 
setLeftOffset(Configuration, int) - Static method in class org.apache.hadoop.mapreduce.lib.partition.BinaryPartitioner
Set the subarray to be used for partitioning to bytes[offset:] in Python syntax.
setLocalAnalysisClass(Class<? extends ILocalAnalysis>) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Set the local analysis class.
setLongValue(Object, long) - Method in class org.apache.hadoop.contrib.utils.join.JobBase
Set the given counter to the given value
setMapDebugScript(String) - Method in class org.apache.hadoop.mapred.JobConf
Set the debug script to run when the map tasks fail.
setMapOutputCompressorClass(Class<? extends CompressionCodec>) - Method in class org.apache.hadoop.mapred.JobConf
Set the given class as the CompressionCodec for the map outputs.
setMapOutputKeyClass(Class<?>) - Method in class org.apache.hadoop.mapred.JobConf
Set the key class for the map output data.
setMapOutputKeyClass(Class<?>) - Method in class org.apache.hadoop.mapreduce.Job
Set the key class for the map output data.
setMapOutputValueClass(Class<?>) - Method in class org.apache.hadoop.mapred.JobConf
Set the value class for the map output data.
setMapOutputValueClass(Class<?>) - Method in class org.apache.hadoop.mapreduce.Job
Set the value class for the map output data.
setMapperClass(Class<? extends Mapper>) - Method in class org.apache.hadoop.mapred.JobConf
Set the Mapper class for the job.
setMapperClass(Class<? extends Mapper>) - Method in class org.apache.hadoop.mapreduce.Job
Set the Mapper for the job.
setMapperClass(Job, Class<? extends Mapper<K1, V1, K2, V2>>) - Static method in class org.apache.hadoop.mapreduce.lib.map.MultithreadedMapper
Set the application's mapper class.
setMapperMaxSkipRecords(Configuration, long) - Static method in class org.apache.hadoop.mapred.SkipBadRecords
Set the number of acceptable skip records surrounding the bad record PER bad record in mapper.
setMapProgress(float) - Method in class org.apache.hadoop.mapred.JobStatus
Sets the map progress of this job
setMapProgress(float) - Method in class org.apache.hadoop.mapreduce.JobStatus
Sets the map progress of this job
setMapRunnerClass(Class<? extends MapRunnable>) - Method in class org.apache.hadoop.mapred.JobConf
Expert: Set the MapRunnable class for the job.
setMapSpeculativeExecution(boolean) - Method in class org.apache.hadoop.mapred.JobConf
Turn speculative execution on or off for this job for map tasks.
setMapSpeculativeExecution(boolean) - Method in class org.apache.hadoop.mapreduce.Job
Turn speculative execution on or off for this job for map tasks.
setMaxInputSplitSize(Job, long) - Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
Set the maximum split size
setMaxItems(long) - Method in class org.apache.hadoop.mapreduce.lib.aggregate.UniqValueCount
Set the limit on the number of unique values
setMaxMapAttempts(int) - Method in class org.apache.hadoop.mapred.JobConf
Expert: Set the number of maximum attempts that will be made to run a map task.
setMaxMapAttempts(int) - Method in class org.apache.hadoop.mapreduce.Job
Expert: Set the number of maximum attempts that will be made to run a map task.
setMaxMapTaskFailuresPercent(int) - Method in class org.apache.hadoop.mapred.JobConf
Expert: Set the maximum percentage of map tasks that can fail without the job being aborted.
setMaxPhysicalMemoryForTask(long) - Method in class org.apache.hadoop.mapred.JobConf
Deprecated. 
setMaxRAMSizeInBytes(long) - Method in class org.apache.hadoop.contrib.index.mapred.IndexUpdateConfiguration
Set the max ram index size in bytes.
setMaxReduceAttempts(int) - Method in class org.apache.hadoop.mapred.JobConf
Expert: Set the number of maximum attempts that will be made to run a reduce task.
setMaxReduceAttempts(int) - Method in class org.apache.hadoop.mapreduce.Job
Expert: Set the number of maximum attempts that will be made to run a reduce task.
setMaxReduceTaskFailuresPercent(int) - Method in class org.apache.hadoop.mapred.JobConf
Set the maximum percentage of reduce tasks that can fail without the job being aborted.
setMaxSplitSize(long) - Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat
Specify the maximum size (in bytes) of each split.
setMaxTaskFailuresPerTracker(int) - Method in class org.apache.hadoop.mapred.JobConf
Set the maximum no.
setMaxVirtualMemoryForTask(long) - Method in class org.apache.hadoop.mapred.JobConf
Deprecated. Use JobConf.setMemoryForMapTask(long mem) and Use JobConf.setMemoryForReduceTask(long mem)
setMemoryForMapTask(long) - Method in class org.apache.hadoop.mapred.JobConf
 
setMemoryForReduceTask(long) - Method in class org.apache.hadoop.mapred.JobConf
 
setMessage(String) - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
Set the message for this job.
setMinInputSplitSize(Job, long) - Static method in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
Set the minimum input split size
setMinSplitSize(long) - Method in class org.apache.hadoop.mapred.FileInputFormat
 
setMinSplitSizeNode(long) - Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat
Specify the minimum size (in bytes) of each split per node.
setMinSplitSizeRack(long) - Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat
Specify the minimum size (in bytes) of each split per rack.
setNumberOfThreads(Job, int) - Static method in class org.apache.hadoop.mapreduce.lib.map.MultithreadedMapper
Set the number of threads in the pool for running maps.
setNumLinesPerSplit(Job, int) - Static method in class org.apache.hadoop.mapreduce.lib.input.NLineInputFormat
Set the number of lines per split
setNumMapTasks(int) - Method in class org.apache.hadoop.mapred.JobConf
Set the number of map tasks for this job.
setNumReduceTasks(int) - Method in class org.apache.hadoop.mapred.JobConf
Set the requisite number of reduce tasks for this job.
setNumReduceTasks(int) - Method in class org.apache.hadoop.mapreduce.Job
Set the number of reduce tasks for the job.
setNumTasksToExecutePerJvm(int) - Method in class org.apache.hadoop.mapred.JobConf
Sets the number of tasks that a spawned task JVM should run before it exits
setOffsets(Configuration, int, int) - Static method in class org.apache.hadoop.mapreduce.lib.partition.BinaryPartitioner
Set the subarray to be used for partitioning to bytes[left:(right+1)] in Python syntax.
setOp(DocumentAndOp.Op) - Method in class org.apache.hadoop.contrib.index.example.LineDocTextAndOp
Set the type of the operation.
setOutput(JobConf, String, String...) - Static method in class org.apache.hadoop.mapred.lib.db.DBOutputFormat
Initializes the reduce-part of the job with the appropriate output settings
setOutput(JobConf, String, int) - Static method in class org.apache.hadoop.mapred.lib.db.DBOutputFormat
Initializes the reduce-part of the job with the appropriate output settings
setOutput(Job, String, String...) - Static method in class org.apache.hadoop.mapreduce.lib.db.DBOutputFormat
Initializes the reduce-part of the job with the appropriate output settings
setOutput(Job, String, int) - Static method in class org.apache.hadoop.mapreduce.lib.db.DBOutputFormat
Initializes the reduce-part of the job with the appropriate output settings
setOutputCommitter(Class<? extends OutputCommitter>) - Method in class org.apache.hadoop.mapred.JobConf
Set the OutputCommitter implementation for the map-reduce job.
setOutputCompressionType(JobConf, SequenceFile.CompressionType) - Static method in class org.apache.hadoop.mapred.SequenceFileOutputFormat
Set the SequenceFile.CompressionType for the output SequenceFile.
setOutputCompressionType(Job, SequenceFile.CompressionType) - Static method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileOutputFormat
Set the SequenceFile.CompressionType for the output SequenceFile.
setOutputCompressorClass(JobConf, Class<? extends CompressionCodec>) - Static method in class org.apache.hadoop.mapred.FileOutputFormat
Set the CompressionCodec to be used to compress job outputs.
setOutputCompressorClass(Job, Class<? extends CompressionCodec>) - Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
Set the CompressionCodec to be used to compress job outputs.
setOutputFieldCount(int) - Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
 
setOutputFieldNames(String...) - Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
 
setOutputFormat(Class<? extends OutputFormat>) - Method in class org.apache.hadoop.mapred.JobConf
Set the OutputFormat implementation for the map-reduce job.
setOutputFormatClass(JobConf, Class<? extends OutputFormat>) - Static method in class org.apache.hadoop.mapred.lib.LazyOutputFormat
Set the underlying output format for LazyOutputFormat.
setOutputFormatClass(Class<? extends OutputFormat>) - Method in class org.apache.hadoop.mapreduce.Job
Set the OutputFormat for the job.
setOutputFormatClass(Job, Class<? extends OutputFormat>) - Static method in class org.apache.hadoop.mapreduce.lib.output.LazyOutputFormat
Set the underlying output format for LazyOutputFormat.
setOutputKeyClass(Class<?>) - Method in class org.apache.hadoop.mapred.JobConf
Set the key class for the job output data.
setOutputKeyClass(Class<?>) - Method in class org.apache.hadoop.mapreduce.Job
Set the key class for the job output data.
setOutputKeyClass(Class) - Method in class org.apache.hadoop.streaming.io.IdentifierResolver
Sets the output key class class.
setOutputKeyComparatorClass(Class<? extends RawComparator>) - Method in class org.apache.hadoop.mapred.JobConf
Set the RawComparator comparator used to compare keys.
setOutputName(JobContext, String) - Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
Set the base output name for output file to be created.
setOutputPath(JobConf, Path) - Static method in class org.apache.hadoop.mapred.FileOutputFormat
Set the Path of the output directory for the map-reduce job.
setOutputPath(Job, Path) - Static method in class org.apache.hadoop.mapreduce.lib.output.FileOutputFormat
Set the Path of the output directory for the map-reduce job.
setOutputReaderClass(Class<? extends OutputReader>) - Method in class org.apache.hadoop.streaming.io.IdentifierResolver
Sets the OutputReader class.
setOutputReplication(Job, int) - Static method in class org.apache.hadoop.examples.terasort.TeraSort
 
setOutputTableName(String) - Method in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
 
setOutputValueClass(Class<?>) - Method in class org.apache.hadoop.mapred.JobConf
Set the value class for job outputs.
setOutputValueClass(Class<?>) - Method in class org.apache.hadoop.mapreduce.Job
Set the value class for job outputs.
setOutputValueClass(Class) - Method in class org.apache.hadoop.streaming.io.IdentifierResolver
Sets the output value class.
setOutputValueGroupingComparator(Class<? extends RawComparator>) - Method in class org.apache.hadoop.mapred.JobConf
Set the user defined RawComparator comparator for grouping keys in the input to the reduce.
setPartitionerClass(Class<? extends Partitioner>) - Method in class org.apache.hadoop.mapred.JobConf
Set the Partitioner class used to partition Mapper-outputs to be sent to the Reducers.
setPartitionerClass(Class<? extends Partitioner>) - Method in class org.apache.hadoop.mapreduce.Job
Set the Partitioner for the job.
setPartitionFile(Configuration, Path) - Static method in class org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner
Set the path to the SequenceFile storing the sorted partition keyset.
setPattern(Configuration, String) - Static method in class org.apache.hadoop.mapred.SequenceFileInputFilter.RegexFilter
 
setPattern(Configuration, String) - Static method in class org.apache.hadoop.mapreduce.lib.input.SequenceFileInputFilter.RegexFilter
Define the filtering regex and stores it in conf
setPrinter(DancingLinks.SolutionAcceptor<Pentomino.ColumnName>) - Method in class org.apache.hadoop.examples.dancing.Pentomino
Set the printer for the puzzle.
setPriority(JobPriority) - Method in class org.apache.hadoop.mapreduce.Job
Set the priority of a running job.
setPriority(JobPriority) - Method in class org.apache.hadoop.mapreduce.JobStatus
Set the priority of the job, defaulting to NORMAL.
setProfileEnabled(boolean) - Method in class org.apache.hadoop.mapred.JobConf
Set whether the system should collect profiler information for some of the tasks in this job? The information is stored in the user log directory.
setProfileEnabled(boolean) - Method in class org.apache.hadoop.mapreduce.Job
Set whether the system should collect profiler information for some of the tasks in this job? The information is stored in the user log directory.
setProfileParams(String) - Method in class org.apache.hadoop.mapred.JobConf
Set the profiler configuration arguments.
setProfileParams(String) - Method in class org.apache.hadoop.mapreduce.Job
Set the profiler configuration arguments.
setProfileTaskRange(boolean, String) - Method in class org.apache.hadoop.mapred.JobConf
Set the ranges of maps or reduces to profile.
setProfileTaskRange(boolean, String) - Method in class org.apache.hadoop.mapreduce.Job
Set the ranges of maps or reduces to profile.
setProperties(Properties) - Method in class org.apache.hadoop.mapred.JobQueueInfo
 
setProperties(Properties) - Method in class org.apache.hadoop.mapreduce.QueueInfo
 
setQueueChildren(List<QueueInfo>) - Method in class org.apache.hadoop.mapreduce.QueueInfo
 
setQueueName(String) - Method in class org.apache.hadoop.mapred.JobConf
Set the name of the queue to which this job should be submitted.
setQueueName(String) - Method in class org.apache.hadoop.mapred.JobQueueInfo
Set the queue name of the JobQueueInfo
setQueueName(String) - Method in class org.apache.hadoop.mapreduce.QueueAclsInfo
 
setQueueName(String) - Method in class org.apache.hadoop.mapreduce.QueueInfo
Set the queue name of the JobQueueInfo
setQueueState(String) - Method in class org.apache.hadoop.mapred.JobQueueInfo
Set the state of the queue
setReduceDebugScript(String) - Method in class org.apache.hadoop.mapred.JobConf
Set the debug script to run when the reduce tasks fail.
setReduceProgress(float) - Method in class org.apache.hadoop.mapred.JobStatus
Sets the reduce progress of this Job
setReduceProgress(float) - Method in class org.apache.hadoop.mapreduce.JobStatus
Sets the reduce progress of this Job
setReducer(JobConf, Class<? extends Reducer<K1, V1, K2, V2>>, Class<? extends K1>, Class<? extends V1>, Class<? extends K2>, Class<? extends V2>, boolean, JobConf) - Static method in class org.apache.hadoop.mapred.lib.ChainReducer
Sets the Reducer class to the chain job's JobConf.
setReducer(Job, Class<? extends Reducer>, Class<?>, Class<?>, Class<?>, Class<?>, Configuration) - Static method in class org.apache.hadoop.mapreduce.lib.chain.ChainReducer
Sets the Reducer class to the chain job.
setReducerClass(Class<? extends Reducer>) - Method in class org.apache.hadoop.mapred.JobConf
Set the Reducer class for the job.
setReducerClass(Class<? extends Reducer>) - Method in class org.apache.hadoop.mapreduce.Job
Set the Reducer for the job.
setReducerMaxSkipGroups(Configuration, long) - Static method in class org.apache.hadoop.mapred.SkipBadRecords
Set the number of acceptable skip groups surrounding the bad group PER bad group in reducer.
setReduceSpeculativeExecution(boolean) - Method in class org.apache.hadoop.mapred.JobConf
Turn speculative execution on or off for this job for reduce tasks.
setReduceSpeculativeExecution(boolean) - Method in class org.apache.hadoop.mapreduce.Job
Turn speculative execution on or off for this job for reduce tasks.
setRetired() - Method in class org.apache.hadoop.mapred.JobStatus
Set the job retire flag to true.
setRetired() - Method in class org.apache.hadoop.mapreduce.JobStatus
Set the job retire flag to true.
setRightOffset(Configuration, int) - Static method in class org.apache.hadoop.mapreduce.lib.partition.BinaryPartitioner
Set the subarray to be used for partitioning to bytes[:(offset+1)] in Python syntax.
setRunningTaskAttempts(Collection<TaskAttemptID>) - Method in class org.apache.hadoop.mapred.TaskReport
set running attempt(s) of the task.
setRunState(int) - Method in class org.apache.hadoop.mapred.JobStatus
Change the current run state of the job.
setSchedulingInfo(String) - Method in class org.apache.hadoop.mapred.JobQueueInfo
Set the scheduling information associated to particular job queue
setSchedulingInfo(String) - Method in class org.apache.hadoop.mapred.JobStatus
Used to set the scheduling information associated to a particular Job.
setSchedulingInfo(String) - Method in class org.apache.hadoop.mapreduce.JobStatus
Used to set the scheduling information associated to a particular Job.
setSchedulingInfo(String) - Method in class org.apache.hadoop.mapreduce.QueueInfo
Set the scheduling information associated to particular job queue
setSequenceFileOutputKeyClass(JobConf, Class<?>) - Static method in class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat
Set the key class for the SequenceFile
setSequenceFileOutputKeyClass(Job, Class<?>) - Static method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat
Set the key class for the SequenceFile
setSequenceFileOutputValueClass(JobConf, Class<?>) - Static method in class org.apache.hadoop.mapred.SequenceFileAsBinaryOutputFormat
Set the value class for the SequenceFile
setSequenceFileOutputValueClass(Job, Class<?>) - Static method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat
Set the value class for the SequenceFile
setSessionId(String) - Method in class org.apache.hadoop.mapred.JobConf
Deprecated. 
setSessionTimeZone(Configuration, Connection) - Static method in class org.apache.hadoop.mapreduce.lib.db.OracleDBRecordReader
Set session time zone
setSetupProgress(float) - Method in class org.apache.hadoop.mapred.JobStatus
Sets the setup progress of this job
setSetupProgress(float) - Method in class org.apache.hadoop.mapreduce.JobStatus
Sets the setup progress of this job
setSkipOutputPath(JobConf, Path) - Static method in class org.apache.hadoop.mapred.SkipBadRecords
Set the directory to which skipped records are written.
setSortComparatorClass(Class<? extends RawComparator>) - Method in class org.apache.hadoop.mapreduce.Job
Define the comparator that controls how the keys are sorted before they are passed to the Reducer.
setSpeculativeExecution(boolean) - Method in class org.apache.hadoop.mapred.JobConf
Turn speculative execution on or off for this job.
setSpeculativeExecution(boolean) - Method in class org.apache.hadoop.mapreduce.Job
Turn speculative execution on or off for this job.
setStartTime(long) - Method in class org.apache.hadoop.mapred.JobStatus
Set the start time of the job
setStartTime(long) - Method in class org.apache.hadoop.mapred.TaskReport
set start time of the task.
setStartTime(long) - Method in class org.apache.hadoop.mapreduce.JobStatus
Set the start time of the job
setState(JobStatus.State) - Method in class org.apache.hadoop.mapreduce.JobStatus
Change the current run state of the job.
setState(QueueState) - Method in class org.apache.hadoop.mapreduce.QueueInfo
Set the state of the queue
setStatement(PreparedStatement) - Method in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
 
setStatus(String) - Method in interface org.apache.hadoop.mapred.Reporter
Set the status description for the task.
setStatus(String) - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
setStatus(String) - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
setStatus(String) - Method in interface org.apache.hadoop.mapreduce.TaskAttemptContext
Set the current status of the task to the given string.
setSuccessfulAttempt(TaskAttemptID) - Method in class org.apache.hadoop.mapred.TaskReport
set successful attempt ID of the task.
setTag(Text) - Method in class org.apache.hadoop.contrib.utils.join.TaggedMapOutput
 
setTaskAttemptId(TaskAttemptID) - Method in class org.apache.hadoop.mapred.TaskCompletionEvent
Sets task id.
setTaskAttemptId(TaskAttemptID) - Method in class org.apache.hadoop.mapreduce.TaskCompletionEvent
Sets task id.
setTaskId(String) - Method in class org.apache.hadoop.mapred.TaskCompletionEvent
Deprecated. use TaskCompletionEvent.setTaskAttemptId(TaskAttemptID) instead.
setTaskId(String) - Method in class org.apache.hadoop.mapred.TaskLogAppender
 
setTaskOutputFilter(JobClient.TaskStatusFilter) - Method in class org.apache.hadoop.mapred.JobClient
Deprecated. 
setTaskOutputFilter(JobConf, JobClient.TaskStatusFilter) - Static method in class org.apache.hadoop.mapred.JobClient
Modify the JobConf to set the task output filter.
setTaskOutputFilter(Configuration, Job.TaskStatusFilter) - Static method in class org.apache.hadoop.mapreduce.Job
Modify the Configuration to set the task output filter.
setTaskRunTime(int) - Method in class org.apache.hadoop.mapred.TaskCompletionEvent
Set the task completion time
setTaskRunTime(int) - Method in class org.apache.hadoop.mapreduce.TaskCompletionEvent
Set the task completion time
setTaskStatus(TaskCompletionEvent.Status) - Method in class org.apache.hadoop.mapred.TaskCompletionEvent
Set task status.
setTaskStatus(TaskCompletionEvent.Status) - Method in class org.apache.hadoop.mapreduce.TaskCompletionEvent
Set task status.
setTaskTrackerHttp(String) - Method in class org.apache.hadoop.mapred.TaskCompletionEvent
Set task tracker http location.
setTaskTrackerHttp(String) - Method in class org.apache.hadoop.mapreduce.TaskCompletionEvent
Set task tracker http location.
setTokenStorage(TokenStorage) - Static method in class org.apache.hadoop.mapreduce.security.TokenCache
sets TokenStorage
setTotalLogFileSize(long) - Method in class org.apache.hadoop.mapred.TaskLogAppender
 
setTrackingUrl(String) - Method in class org.apache.hadoop.mapred.JobStatus
Set the link to the web-ui for details of the job.
setTrackingUrl(String) - Method in class org.apache.hadoop.mapreduce.JobStatus
Set the link to the web-ui for details of the job.
setup(Mapper<WritableComparable<?>, Text, Text, Text>.Context) - Method in class org.apache.hadoop.examples.dancing.DistributedPentomino.PentMap
 
setup(Configuration) - Static method in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorJobBase
 
setup(Mapper<K1, V1, Text, Text>.Context) - Method in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorMapper
 
setup(Reducer<Text, Text, Text, Text>.Context) - Method in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorReducer
 
setup(Mapper<KEYIN, VALUEIN, KEYOUT, VALUEOUT>.Context) - Method in class org.apache.hadoop.mapreduce.lib.chain.ChainMapper
 
setup(Reducer<KEYIN, VALUEIN, KEYOUT, VALUEOUT>.Context) - Method in class org.apache.hadoop.mapreduce.lib.chain.ChainReducer
 
setup(Mapper<K, V, Text, Text>.Context) - Method in class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionMapper
 
setup(Reducer<Text, Text, Text, Text>.Context) - Method in class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionReducer
 
setup(Mapper<K, Text, Text, LongWritable>.Context) - Method in class org.apache.hadoop.mapreduce.lib.map.RegexMapper
 
setup(Mapper<KEYIN, VALUEIN, KEYOUT, VALUEOUT>.Context) - Method in class org.apache.hadoop.mapreduce.Mapper
Called once at the beginning of the task.
setup(Reducer<KEYIN, VALUEIN, KEYOUT, VALUEOUT>.Context) - Method in class org.apache.hadoop.mapreduce.Reducer
Called once at the start of the task.
setUpdate(Document, Term) - Method in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
Set the instance to be an update operation.
setupJob(JobContext) - Method in class org.apache.hadoop.examples.terasort.TeraOutputFormat.TeraOutputCommitter
 
setupJob(JobContext) - Method in class org.apache.hadoop.mapred.FileOutputCommitter
 
setupJob(JobContext) - Method in class org.apache.hadoop.mapred.OutputCommitter
For the framework to setup the job output during initialization
setupJob(JobContext) - Method in class org.apache.hadoop.mapred.OutputCommitter
This method implements the new interface by calling the old method.
setupJob(JobContext) - Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
Create the temporary directory that is the root of all of the task work directories.
setupJob(JobContext) - Method in class org.apache.hadoop.mapreduce.OutputCommitter
For the framework to setup the job output during initialization
setupProgress() - Method in class org.apache.hadoop.mapred.JobStatus
 
setupProgress() - Method in interface org.apache.hadoop.mapred.RunningJob
Get the progress of the job's setup-tasks, as a float between 0.0 and 1.0.
setupProgress() - Method in class org.apache.hadoop.mapreduce.Job
Get the progress of the job's setup-tasks, as a float between 0.0 and 1.0.
setupTask(TaskAttemptContext) - Method in class org.apache.hadoop.examples.terasort.TeraOutputFormat.TeraOutputCommitter
 
setupTask(TaskAttemptContext) - Method in class org.apache.hadoop.mapred.FileOutputCommitter
 
setupTask(TaskAttemptContext) - Method in class org.apache.hadoop.mapred.OutputCommitter
Sets up output for the task.
setupTask(TaskAttemptContext) - Method in class org.apache.hadoop.mapred.OutputCommitter
This method implements the new interface by calling the old method.
setupTask(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
No task setup required.
setupTask(TaskAttemptContext) - Method in class org.apache.hadoop.mapreduce.OutputCommitter
Sets up output for the task.
setUseNewMapper(boolean) - Method in class org.apache.hadoop.mapred.JobConf
Set whether the framework should use the new api for the mapper.
setUseNewReducer(boolean) - Method in class org.apache.hadoop.mapred.JobConf
Set whether the framework should use the new api for the reducer.
setUser(String) - Method in class org.apache.hadoop.mapred.JobConf
Set the reported username for this job.
setUser(String) - Method in class org.apache.hadoop.mapreduce.Job
Set the reported username for this job.
setUsername(String) - Method in class org.apache.hadoop.mapred.JobStatus
 
setUsername(String) - Method in class org.apache.hadoop.mapreduce.JobStatus
 
setUseSimplePartitioner(Job, boolean) - Static method in class org.apache.hadoop.examples.terasort.TeraSort
 
setValue(Summation) - Method in class org.apache.hadoop.examples.pi.math.Bellard.Sum
Set the value of sigma
setValue(double) - Method in class org.apache.hadoop.examples.pi.math.Summation
Set the value of this summation
setValue(Object) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritable
Set the typed bytes from a given Java object.
setVerbose(boolean) - Method in class org.apache.hadoop.streaming.JarBuilder
 
setWorkingDirectory(Path) - Method in class org.apache.hadoop.mapred.JobConf
Set the current working directory for the default file system.
setWorkingDirectory(Path) - Method in class org.apache.hadoop.mapreduce.Job
Set the current working directory for the default file system.
Shard - Class in org.apache.hadoop.contrib.index.mapred
This class represents the metadata of a shard.
Shard() - Constructor for class org.apache.hadoop.contrib.index.mapred.Shard
Constructor.
Shard(long, String, long) - Constructor for class org.apache.hadoop.contrib.index.mapred.Shard
Construct a shard from a versio number, a directory and a generation number.
Shard(Shard) - Constructor for class org.apache.hadoop.contrib.index.mapred.Shard
Construct using a shard object.
ShardWriter - Class in org.apache.hadoop.contrib.index.lucene
The initial version of an index is stored in the perm dir.
ShardWriter(FileSystem, Shard, String, IndexUpdateConfiguration) - Constructor for class org.apache.hadoop.contrib.index.lucene.ShardWriter
Constructor
shippedCanonFiles_ - Variable in class org.apache.hadoop.streaming.StreamJob
 
Shuffle.ShuffleError - Exception in org.apache.hadoop.mapreduce.task.reduce
 
shutdown() - Method in class org.apache.hadoop.mapred.gridmix.Statistics
 
size() - Method in class org.apache.hadoop.mapred.Counters.Group
Returns the number of counters in this group.
size() - Method in class org.apache.hadoop.mapred.Counters
Returns the total number of counters, by summing the number of counters in each group.
size() - Method in class org.apache.hadoop.mapreduce.CounterGroup
Returns the number of counters in this group.
size() - Method in class org.apache.hadoop.mapreduce.lib.join.TupleWritable
The number of children in this Tuple.
skip(K) - Method in interface org.apache.hadoop.mapred.join.ComposableRecordReader
Skip key-value pairs with keys less than or equal to the key provided.
skip(K) - Method in class org.apache.hadoop.mapred.join.CompositeRecordReader
Pass skip key to child RRs.
skip(K) - Method in class org.apache.hadoop.mapred.join.WrappedRecordReader
Skip key-value pairs with keys less than or equal to the key provided.
skip(K) - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
Pass skip key to child RRs.
skip(K) - Method in class org.apache.hadoop.mapreduce.lib.join.WrappedRecordReader
Skip key-value pairs with keys less than or equal to the key provided.
SkipBadRecords - Class in org.apache.hadoop.mapred
Utility class for skip bad records functionality.
SkipBadRecords() - Constructor for class org.apache.hadoop.mapred.SkipBadRecords
 
skipType() - Method in class org.apache.hadoop.typedbytes.TypedBytesInput
Skips a type byte.
solution(List<List<ColumnName>>) - Method in interface org.apache.hadoop.examples.dancing.DancingLinks.SolutionAcceptor
A callback to return a solution to the application.
solve(int[], DancingLinks.SolutionAcceptor<ColumnName>) - Method in class org.apache.hadoop.examples.dancing.DancingLinks
Given a prefix, find solutions under it.
solve(DancingLinks.SolutionAcceptor<ColumnName>) - Method in class org.apache.hadoop.examples.dancing.DancingLinks
Solve a complete problem
solve(int[]) - Method in class org.apache.hadoop.examples.dancing.Pentomino
Find all of the solutions that start with the given prefix.
solve() - Method in class org.apache.hadoop.examples.dancing.Pentomino
Find all of the solutions to the puzzle.
solve() - Method in class org.apache.hadoop.examples.dancing.Sudoku
 
Sort<K,V> - Class in org.apache.hadoop.examples
This is the trivial map/reduce program that does absolutely nothing other than use the framework to fragment and sort the input values.
Sort() - Constructor for class org.apache.hadoop.examples.Sort
 
SOURCE_TAGS_FIELD - Static variable in class org.apache.hadoop.contrib.utils.join.DataJoinReducerBase
 
specToString(String, String, int, List<Integer>, List<Integer>) - Static method in class org.apache.hadoop.mapreduce.lib.fieldsel.FieldSelectionHelper
 
split(int) - Method in class org.apache.hadoop.examples.dancing.DancingLinks
Generate a list of row choices to cover the first moves.
split - Variable in class org.apache.hadoop.mapred.lib.CombineFileRecordReader
 
split(Configuration, ResultSet, String) - Method in class org.apache.hadoop.mapreduce.lib.db.BigDecimalSplitter
 
split(Configuration, ResultSet, String) - Method in class org.apache.hadoop.mapreduce.lib.db.BooleanSplitter
 
split(Configuration, ResultSet, String) - Method in class org.apache.hadoop.mapreduce.lib.db.DateSplitter
 
split(Configuration, ResultSet, String) - Method in interface org.apache.hadoop.mapreduce.lib.db.DBSplitter
Given a ResultSet containing one record (and already advanced to that record) with two columns (a low value, and a high value, both of the same type), determine a set of splits that span the given values.
split(Configuration, ResultSet, String) - Method in class org.apache.hadoop.mapreduce.lib.db.FloatSplitter
 
split(Configuration, ResultSet, String) - Method in class org.apache.hadoop.mapreduce.lib.db.IntegerSplitter
 
split(Configuration, ResultSet, String) - Method in class org.apache.hadoop.mapreduce.lib.db.TextSplitter
This method needs to determine the splits between two user-provided strings.
split - Variable in class org.apache.hadoop.mapreduce.lib.input.CombineFileRecordReader
 
SPLIT_MAXSIZE - Static variable in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
 
SPLIT_MINSIZE - Static variable in class org.apache.hadoop.mapreduce.lib.input.FileInputFormat
 
SPLIT_MINSIZE_PERNODE - Static variable in class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat
 
SPLIT_MINSIZE_PERRACK - Static variable in class org.apache.hadoop.mapreduce.lib.input.CombineFileInputFormat
 
splitKeyVal(byte[], int, int, Text, Text, int, int) - Static method in class org.apache.hadoop.streaming.StreamKeyValUtil
split a UTF-8 byte array into key and value assuming that the delimilator is at splitpos.
splitKeyVal(byte[], int, int, Text, Text, int) - Static method in class org.apache.hadoop.streaming.StreamKeyValUtil
split a UTF-8 byte array into key and value assuming that the delimilator is at splitpos.
splitKeyVal(byte[], Text, Text, int, int) - Static method in class org.apache.hadoop.streaming.StreamKeyValUtil
split a UTF-8 byte array into key and value assuming that the delimilator is at splitpos.
splitKeyVal(byte[], Text, Text, int) - Static method in class org.apache.hadoop.streaming.StreamKeyValUtil
split a UTF-8 byte array into key and value assuming that the delimilator is at splitpos.
splitKeyVal(byte[], int, int, Text, Text, int, int) - Static method in class org.apache.hadoop.streaming.UTF8ByteArrayUtils
Deprecated. use StreamKeyValUtil.splitKeyVal(byte[], int, int, Text, Text, int, int)
splitKeyVal(byte[], int, int, Text, Text, int) - Static method in class org.apache.hadoop.streaming.UTF8ByteArrayUtils
Deprecated. use StreamKeyValUtil.splitKeyVal(byte[], int, int, Text, Text, int)
splitKeyVal(byte[], Text, Text, int, int) - Static method in class org.apache.hadoop.streaming.UTF8ByteArrayUtils
Deprecated. use StreamKeyValUtil.splitKeyVal(byte[], Text, Text, int, int)
splitKeyVal(byte[], Text, Text, int) - Static method in class org.apache.hadoop.streaming.UTF8ByteArrayUtils
Deprecated. use StreamKeyValUtil.splitKeyVal(byte[], Text, Text, int)
start() - Method in class org.apache.hadoop.mapred.gridmix.SerialJobFactory
Start the reader thread, wait for latch if necessary.
start() - Method in class org.apache.hadoop.mapred.gridmix.Statistics
Attempt to start the service.
start() - Method in class org.apache.hadoop.mapred.gridmix.StressJobFactory
Start the reader thread, wait for latch if necessary.
startMap(String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
 
startMap(TreeMap, String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
 
startRecord(String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
 
startRecord(Record, String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
 
startVector(String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordInput
 
startVector(ArrayList, String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
 
State - Enum in org.apache.hadoop.mapreduce.server.jobtracker
Describes the state of JobTracker
statement - Variable in class org.apache.hadoop.mapreduce.lib.db.DBRecordReader
 
Statistics - Class in org.apache.hadoop.mapred.gridmix
Component collecting the stats required by other components to make decisions.
Statistics(Configuration, int, CountDownLatch) - Constructor for class org.apache.hadoop.mapred.gridmix.Statistics
 
stop() - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl
set the thread state to STOPPING so that the thread will stop when it wakes up.
StreamBackedIterator<X extends org.apache.hadoop.io.Writable> - Class in org.apache.hadoop.mapred.join
This class provides an implementation of ResetableIterator.
StreamBackedIterator() - Constructor for class org.apache.hadoop.mapred.join.StreamBackedIterator
 
StreamBackedIterator<X extends org.apache.hadoop.io.Writable> - Class in org.apache.hadoop.mapreduce.lib.join
This class provides an implementation of ResetableIterator.
StreamBackedIterator() - Constructor for class org.apache.hadoop.mapreduce.lib.join.StreamBackedIterator
 
StreamBaseRecordReader - Class in org.apache.hadoop.streaming
Shared functionality for hadoopStreaming formats.
StreamBaseRecordReader(FSDataInputStream, FileSplit, Reporter, JobConf, FileSystem) - Constructor for class org.apache.hadoop.streaming.StreamBaseRecordReader
 
StreamInputFormat - Class in org.apache.hadoop.streaming
An input format that selects a RecordReader based on a JobConf property.
StreamInputFormat() - Constructor for class org.apache.hadoop.streaming.StreamInputFormat
 
StreamJob - Class in org.apache.hadoop.streaming
All the client-side work happens here.
StreamJob(String[], boolean) - Constructor for class org.apache.hadoop.streaming.StreamJob
Deprecated. use StreamJob() with ToolRunner or set the Configuration using StreamJob.setConf(Configuration) and run with StreamJob.run(String[]).
StreamJob() - Constructor for class org.apache.hadoop.streaming.StreamJob
 
StreamKeyValUtil - Class in org.apache.hadoop.streaming
 
StreamKeyValUtil() - Constructor for class org.apache.hadoop.streaming.StreamKeyValUtil
 
StreamUtil - Class in org.apache.hadoop.streaming
Utilities not available elsewhere in Hadoop.
StreamUtil() - Constructor for class org.apache.hadoop.streaming.StreamUtil
 
StreamXmlRecordReader - Class in org.apache.hadoop.streaming
A way to interpret XML fragments as Mapper input records.
StreamXmlRecordReader(FSDataInputStream, FileSplit, Reporter, JobConf, FileSystem) - Constructor for class org.apache.hadoop.streaming.StreamXmlRecordReader
 
StressJobFactory - Class in org.apache.hadoop.mapred.gridmix
 
StressJobFactory(JobSubmitter, JobStoryProducer, Path, Configuration, CountDownLatch) - Constructor for class org.apache.hadoop.mapred.gridmix.StressJobFactory
Creating a new instance does not start the thread.
string2long(String) - Static method in class org.apache.hadoop.examples.pi.Util
Covert a String to a long.
string2TaskResult(String) - Static method in class org.apache.hadoop.examples.pi.DistSum
Convert a String to a (String, TaskResult) pair
STRING_VALUE_MAX - Static variable in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor
 
STRING_VALUE_MAX - Static variable in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorBaseDescriptor
 
STRING_VALUE_MIN - Static variable in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor
 
STRING_VALUE_MIN - Static variable in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorBaseDescriptor
 
stringifySolution(int, int, List<List<Pentomino.ColumnName>>) - Static method in class org.apache.hadoop.examples.dancing.Pentomino
Convert a solution to the puzzle returned by the model into a string that represents the placement of the pieces onto the board.
StringValueMax - Class in org.apache.hadoop.mapred.lib.aggregate
This class implements a value aggregator that maintain the biggest of a sequence of strings.
StringValueMax() - Constructor for class org.apache.hadoop.mapred.lib.aggregate.StringValueMax
 
StringValueMax - Class in org.apache.hadoop.mapreduce.lib.aggregate
This class implements a value aggregator that maintain the biggest of a sequence of strings.
StringValueMax() - Constructor for class org.apache.hadoop.mapreduce.lib.aggregate.StringValueMax
the default constructor
StringValueMin - Class in org.apache.hadoop.mapred.lib.aggregate
This class implements a value aggregator that maintain the smallest of a sequence of strings.
StringValueMin() - Constructor for class org.apache.hadoop.mapred.lib.aggregate.StringValueMin
 
StringValueMin - Class in org.apache.hadoop.mapreduce.lib.aggregate
This class implements a value aggregator that maintain the smallest of a sequence of strings.
StringValueMin() - Constructor for class org.apache.hadoop.mapreduce.lib.aggregate.StringValueMin
the default constructor
submit() - Method in class org.apache.hadoop.mapreduce.Job
Submit the job to the cluster and return immediately.
submit() - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
Submit this job to mapred.
SUBMIT_REPLICATION - Static variable in class org.apache.hadoop.mapreduce.Job
 
submitAndMonitorJob() - Method in class org.apache.hadoop.streaming.StreamJob
 
submitJob(String) - Method in class org.apache.hadoop.mapred.JobClient
Submit a job to the MR system.
submitJob(JobConf) - Method in class org.apache.hadoop.mapred.JobClient
Submit a job to the MR system.
submitJob(JobConf) - Static method in class org.apache.hadoop.mapred.pipes.Submitter
Deprecated. Use Submitter.runJob(JobConf)
Submitter - Class in org.apache.hadoop.mapred.pipes
The main entry point and job submitter.
Submitter() - Constructor for class org.apache.hadoop.mapred.pipes.Submitter
 
Submitter(Configuration) - Constructor for class org.apache.hadoop.mapred.pipes.Submitter
 
SUBSTITUTE_TOKEN - Static variable in class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat
If users are providing their own query, the following string is expected to appear in the WHERE clause, which will be substituted with a pair of conditions on the input to allow input splits to parallelise the import.
SUCCEEDED - Static variable in class org.apache.hadoop.mapred.JobStatus
 
SUCCEEDED_FILE_NAME - Static variable in class org.apache.hadoop.mapred.FileOutputCommitter
 
SUCCEEDED_FILE_NAME - Static variable in class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
 
SUCCESS - Static variable in class org.apache.hadoop.mapred.jobcontrol.Job
 
Sudoku - Class in org.apache.hadoop.examples.dancing
This class uses the dancing links algorithm from Knuth to solve sudoku puzzles.
Sudoku(InputStream) - Constructor for class org.apache.hadoop.examples.dancing.Sudoku
Set up a puzzle board to the given size.
Sudoku.ColumnName - Interface in org.apache.hadoop.examples.dancing
This interface is a marker class for the columns created for the Sudoku solver.
sum(Counters, Counters) - Static method in class org.apache.hadoop.mapred.Counters
Convenience method for computing the sum of two sets of counters.
Summation - Class in org.apache.hadoop.examples.pi.math
Represent the summation \sum \frac{2^e \mod n}{n}.
Summation(ArithmeticProgression, ArithmeticProgression) - Constructor for class org.apache.hadoop.examples.pi.math.Summation
Constructor
SummationWritable - Class in org.apache.hadoop.examples.pi
A Writable class for Summation
SummationWritable() - Constructor for class org.apache.hadoop.examples.pi.SummationWritable
 
suspend() - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl
suspend the running thread
symbol - Variable in class org.apache.hadoop.examples.pi.math.ArithmeticProgression
A symbol

T

tag - Variable in class org.apache.hadoop.contrib.utils.join.TaggedMapOutput
 
TaggedMapOutput - Class in org.apache.hadoop.contrib.utils.join
This abstract class serves as the base class for the values that flow from the mappers to the reducers in a data join job.
TaggedMapOutput() - Constructor for class org.apache.hadoop.contrib.utils.join.TaggedMapOutput
 
TASK - Static variable in class org.apache.hadoop.mapreduce.TaskID
 
TaskAttemptContext - Interface in org.apache.hadoop.mapred
 
TaskAttemptContext - Interface in org.apache.hadoop.mapreduce
The context for task attempts.
TaskAttemptID - Class in org.apache.hadoop.mapred
TaskAttemptID represents the immutable and unique identifier for a task attempt.
TaskAttemptID(TaskID, int) - Constructor for class org.apache.hadoop.mapred.TaskAttemptID
Constructs a TaskAttemptID object from given TaskID.
TaskAttemptID(String, int, boolean, int, int) - Constructor for class org.apache.hadoop.mapred.TaskAttemptID
Deprecated. Use TaskAttemptID.TaskAttemptID(String, int, TaskType, int, int).
TaskAttemptID(String, int, TaskType, int, int) - Constructor for class org.apache.hadoop.mapred.TaskAttemptID
Constructs a TaskId object from given parts.
TaskAttemptID() - Constructor for class org.apache.hadoop.mapred.TaskAttemptID
 
TaskAttemptID - Class in org.apache.hadoop.mapreduce
TaskAttemptID represents the immutable and unique identifier for a task attempt.
TaskAttemptID(TaskID, int) - Constructor for class org.apache.hadoop.mapreduce.TaskAttemptID
Constructs a TaskAttemptID object from given TaskID.
TaskAttemptID(String, int, TaskType, int, int) - Constructor for class org.apache.hadoop.mapreduce.TaskAttemptID
Constructs a TaskId object from given parts.
TaskAttemptID() - Constructor for class org.apache.hadoop.mapreduce.TaskAttemptID
 
TaskCompletionEvent - Class in org.apache.hadoop.mapred
This is used to track task completion events on job tracker.
TaskCompletionEvent() - Constructor for class org.apache.hadoop.mapred.TaskCompletionEvent
Default constructor for Writable.
TaskCompletionEvent(int, TaskAttemptID, int, boolean, TaskCompletionEvent.Status, String) - Constructor for class org.apache.hadoop.mapred.TaskCompletionEvent
Constructor.
TaskCompletionEvent - Class in org.apache.hadoop.mapreduce
This is used to track task completion events on job tracker.
TaskCompletionEvent() - Constructor for class org.apache.hadoop.mapreduce.TaskCompletionEvent
Default constructor for Writable.
TaskCompletionEvent(int, TaskAttemptID, int, boolean, TaskCompletionEvent.Status, String) - Constructor for class org.apache.hadoop.mapreduce.TaskCompletionEvent
Constructor.
TaskCompletionEvent.Status - Enum in org.apache.hadoop.mapred
 
TaskCompletionEvent.Status - Enum in org.apache.hadoop.mapreduce
 
TaskCounter - Enum in org.apache.hadoop.mapreduce
 
TaskID - Class in org.apache.hadoop.mapred
TaskID represents the immutable and unique identifier for a Map or Reduce Task.
TaskID(JobID, boolean, int) - Constructor for class org.apache.hadoop.mapred.TaskID
Deprecated. Use TaskID.TaskID(String, int, TaskType, int)
TaskID(String, int, boolean, int) - Constructor for class org.apache.hadoop.mapred.TaskID
Deprecated. Use TaskID.TaskID(org.apache.hadoop.mapreduce.JobID, TaskType, int)
TaskID(JobID, TaskType, int) - Constructor for class org.apache.hadoop.mapred.TaskID
Constructs a TaskID object from given JobID.
TaskID(String, int, TaskType, int) - Constructor for class org.apache.hadoop.mapred.TaskID
Constructs a TaskInProgressId object from given parts.
TaskID() - Constructor for class org.apache.hadoop.mapred.TaskID
 
TaskID - Class in org.apache.hadoop.mapreduce
TaskID represents the immutable and unique identifier for a Map or Reduce Task.
TaskID(JobID, TaskType, int) - Constructor for class org.apache.hadoop.mapreduce.TaskID
Constructs a TaskID object from given JobID.
TaskID(String, int, TaskType, int) - Constructor for class org.apache.hadoop.mapreduce.TaskID
Constructs a TaskInProgressId object from given parts.
TaskID() - Constructor for class org.apache.hadoop.mapreduce.TaskID
 
TaskInputOutputContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT> - Interface in org.apache.hadoop.mapreduce
A context object that allows input and output from the task.
TaskLog.Reader - Class in org.apache.hadoop.mapred
 
TaskLog.Reader(TaskAttemptID, TaskLog.LogName, long, long, boolean) - Constructor for class org.apache.hadoop.mapred.TaskLog.Reader
Read a log file from start to end positions.
TaskLogAppender - Class in org.apache.hadoop.mapred
A simple log4j-appender for the task child's map-reduce system logs.
TaskLogAppender() - Constructor for class org.apache.hadoop.mapred.TaskLogAppender
 
TaskReport - Class in org.apache.hadoop.mapred
A report on the state of a task.
TaskReport() - Constructor for class org.apache.hadoop.mapred.TaskReport
 
TaskResult - Class in org.apache.hadoop.examples.pi
A class for map task results or reduce task results.
TaskResult() - Constructor for class org.apache.hadoop.examples.pi.TaskResult
 
taskResult2string(String, TaskResult) - Static method in class org.apache.hadoop.examples.pi.DistSum
Convert a TaskResult to a String
TaskTrackerInfo - Class in org.apache.hadoop.mapreduce
Information about TaskTracker.
TaskTrackerInfo() - Constructor for class org.apache.hadoop.mapreduce.TaskTrackerInfo
 
TaskTrackerInfo(String) - Constructor for class org.apache.hadoop.mapreduce.TaskTrackerInfo
 
TaskTrackerInfo(String, String, String) - Constructor for class org.apache.hadoop.mapreduce.TaskTrackerInfo
 
TaskType - Enum in org.apache.hadoop.mapreduce
Enum for map, reduce, job-setup, job-cleanup, task-cleanup task types.
TEMP_DIR_NAME - Static variable in class org.apache.hadoop.mapred.FileOutputCommitter
Temporary directory name
TEMP_DIR_NAME - Static variable in class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
Temporary directory name
TeraChecksum - Class in org.apache.hadoop.examples.terasort
 
TeraChecksum() - Constructor for class org.apache.hadoop.examples.terasort.TeraChecksum
 
TeraGen - Class in org.apache.hadoop.examples.terasort
Generate the official GraySort input data set.
TeraGen() - Constructor for class org.apache.hadoop.examples.terasort.TeraGen
 
TeraGen.Counters - Enum in org.apache.hadoop.examples.terasort
 
TeraGen.SortGenMapper - Class in org.apache.hadoop.examples.terasort
The Mapper class that given a row number, will generate the appropriate output line.
TeraGen.SortGenMapper() - Constructor for class org.apache.hadoop.examples.terasort.TeraGen.SortGenMapper
 
TeraInputFormat - Class in org.apache.hadoop.examples.terasort
An input format that reads the first 10 characters of each line as the key and the rest of the line as the value.
TeraInputFormat() - Constructor for class org.apache.hadoop.examples.terasort.TeraInputFormat
 
TeraOutputFormat - Class in org.apache.hadoop.examples.terasort
An output format that writes the key and value appended together.
TeraOutputFormat() - Constructor for class org.apache.hadoop.examples.terasort.TeraOutputFormat
 
TeraOutputFormat.TeraOutputCommitter - Class in org.apache.hadoop.examples.terasort
 
TeraOutputFormat.TeraOutputCommitter(Path, TaskAttemptContext) - Constructor for class org.apache.hadoop.examples.terasort.TeraOutputFormat.TeraOutputCommitter
 
TeraSort - Class in org.apache.hadoop.examples.terasort
Generates the sampled split points, launches the job, and waits for it to finish.
TeraSort() - Constructor for class org.apache.hadoop.examples.terasort.TeraSort
 
TeraSort.SimplePartitioner - Class in org.apache.hadoop.examples.terasort
A total order partitioner that assigns keys based on their first PREFIX_LENGTH bytes, assuming a flat distribution.
TeraSort.SimplePartitioner() - Constructor for class org.apache.hadoop.examples.terasort.TeraSort.SimplePartitioner
 
TeraValidate - Class in org.apache.hadoop.examples.terasort
Generate 1 mapper per a file that checks to make sure the keys are sorted within each file.
TeraValidate() - Constructor for class org.apache.hadoop.examples.terasort.TeraValidate
 
TEXT_ID - Static variable in class org.apache.hadoop.streaming.io.IdentifierResolver
 
TextInputFormat - Class in org.apache.hadoop.mapred
An InputFormat for plain text files.
TextInputFormat() - Constructor for class org.apache.hadoop.mapred.TextInputFormat
 
TextInputFormat - Class in org.apache.hadoop.mapreduce.lib.input
An InputFormat for plain text files.
TextInputFormat() - Constructor for class org.apache.hadoop.mapreduce.lib.input.TextInputFormat
 
TextInputWriter - Class in org.apache.hadoop.streaming.io
InputWriter that writes the client's input as text.
TextInputWriter() - Constructor for class org.apache.hadoop.streaming.io.TextInputWriter
 
TextOutputFormat<K,V> - Class in org.apache.hadoop.mapred
An OutputFormat that writes plain text files.
TextOutputFormat() - Constructor for class org.apache.hadoop.mapred.TextOutputFormat
 
TextOutputFormat<K,V> - Class in org.apache.hadoop.mapreduce.lib.output
An OutputFormat that writes plain text files.
TextOutputFormat() - Constructor for class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat
 
TextOutputFormat.LineRecordWriter<K,V> - Class in org.apache.hadoop.mapred
 
TextOutputFormat.LineRecordWriter(DataOutputStream, String) - Constructor for class org.apache.hadoop.mapred.TextOutputFormat.LineRecordWriter
 
TextOutputFormat.LineRecordWriter(DataOutputStream) - Constructor for class org.apache.hadoop.mapred.TextOutputFormat.LineRecordWriter
 
TextOutputFormat.LineRecordWriter<K,V> - Class in org.apache.hadoop.mapreduce.lib.output
 
TextOutputFormat.LineRecordWriter(DataOutputStream, String) - Constructor for class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat.LineRecordWriter
 
TextOutputFormat.LineRecordWriter(DataOutputStream) - Constructor for class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat.LineRecordWriter
 
TextOutputReader - Class in org.apache.hadoop.streaming.io
OutputReader that reads the client's output as text.
TextOutputReader() - Constructor for class org.apache.hadoop.streaming.io.TextOutputReader
 
TextSplitter - Class in org.apache.hadoop.mapreduce.lib.db
Implement DBSplitter over text strings.
TextSplitter() - Constructor for class org.apache.hadoop.mapreduce.lib.db.TextSplitter
 
theAggregatorDescriptor - Variable in class org.apache.hadoop.mapreduce.lib.aggregate.UserDefinedValueAggregatorDescriptor
 
tick() - Method in class org.apache.hadoop.examples.pi.Util.Timer
Same as tick(null).
tick(String) - Method in class org.apache.hadoop.examples.pi.Util.Timer
Tick
TokenCache - Class in org.apache.hadoop.mapreduce.security
This class provides user facing APIs for transferring secrets from the job client to the tasks.
TokenCache() - Constructor for class org.apache.hadoop.mapreduce.security.TokenCache
 
TokenCounterMapper - Class in org.apache.hadoop.mapreduce.lib.map
Tokenize the input values and emit each word with a count of 1.
TokenCounterMapper() - Constructor for class org.apache.hadoop.mapreduce.lib.map.TokenCounterMapper
 
TokenCountMapper<K> - Class in org.apache.hadoop.mapred.lib
A Mapper that maps text values into pairs.
TokenCountMapper() - Constructor for class org.apache.hadoop.mapred.lib.TokenCountMapper
 
toMap() - Method in class org.apache.hadoop.streaming.Environment
 
toString() - Method in class org.apache.hadoop.contrib.index.example.LineDocTextAndOp
 
toString() - Method in class org.apache.hadoop.contrib.index.lucene.FileSystemDirectory
 
toString() - Method in class org.apache.hadoop.contrib.index.lucene.ShardWriter
 
toString() - Method in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp.Op
 
toString() - Method in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
 
toString() - Method in class org.apache.hadoop.contrib.index.mapred.DocumentID
 
toString() - Method in class org.apache.hadoop.contrib.index.mapred.IntermediateForm
 
toString() - Method in class org.apache.hadoop.contrib.index.mapred.Shard
 
toString() - Method in class org.apache.hadoop.examples.pi.DistSum.Machine
toString() - Method in class org.apache.hadoop.examples.pi.math.ArithmeticProgression
toString() - Method in class org.apache.hadoop.examples.pi.math.Bellard.Sum
toString() - Method in class org.apache.hadoop.examples.pi.math.Summation
toString() - Method in class org.apache.hadoop.examples.pi.SummationWritable
toString() - Method in class org.apache.hadoop.examples.pi.TaskResult
toString() - Method in class org.apache.hadoop.mapred.ClusterStatus.BlackListInfo
 
toString() - Method in class org.apache.hadoop.mapred.Counters
Return textual representation of the counter values.
toString() - Method in class org.apache.hadoop.mapred.FileSplit
 
toString() - Method in class org.apache.hadoop.mapred.MultiFileSplit
 
toString() - Method in class org.apache.hadoop.mapreduce.Counters
Return textual representation of the counter values.
toString() - Method in class org.apache.hadoop.mapreduce.ID
 
toString() - Method in class org.apache.hadoop.mapreduce.Job
Dump stats to screen.
toString() - Method in class org.apache.hadoop.mapreduce.JobID
 
toString() - Method in class org.apache.hadoop.mapreduce.JobStatus
 
toString() - Method in class org.apache.hadoop.mapreduce.lib.aggregate.UserDefinedValueAggregatorDescriptor
 
toString() - Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
 
toString() - Method in class org.apache.hadoop.mapreduce.lib.input.FileSplit
 
toString() - Method in class org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob
 
toString() - Method in class org.apache.hadoop.mapreduce.lib.join.TupleWritable
Convert Tuple to String as in the following.
toString() - Method in enum org.apache.hadoop.mapreduce.QueueState
 
toString() - Method in class org.apache.hadoop.mapreduce.split.JobSplit.SplitMetaInfo
 
toString() - Method in class org.apache.hadoop.mapreduce.TaskAttemptID
 
toString() - Method in class org.apache.hadoop.mapreduce.TaskCompletionEvent
 
toString() - Method in class org.apache.hadoop.mapreduce.TaskID
 
toString() - Method in class org.apache.hadoop.typedbytes.TypedBytesWritable
Generate a suitable string representation.
TOTAL_BYTES - Static variable in class org.apache.hadoop.examples.RandomTextWriter
 
TOTAL_BYTES - Static variable in class org.apache.hadoop.examples.RandomWriter
 
TotalOrderPartitioner<K extends org.apache.hadoop.io.WritableComparable<?>,V> - Class in org.apache.hadoop.mapred.lib
Partitioner effecting a total order by reading split points from an externally generated source.
TotalOrderPartitioner() - Constructor for class org.apache.hadoop.mapred.lib.TotalOrderPartitioner
 
TotalOrderPartitioner<K extends org.apache.hadoop.io.WritableComparable<?>,V> - Class in org.apache.hadoop.mapreduce.lib.partition
Partitioner effecting a total order by reading split points from an externally generated source.
TotalOrderPartitioner() - Constructor for class org.apache.hadoop.mapreduce.lib.partition.TotalOrderPartitioner
 
totalSizeInBytes() - Method in class org.apache.hadoop.contrib.index.mapred.IntermediateForm
The total size of files in the directory and ram used by the index writer.
touch(File) - Static method in class org.apache.hadoop.streaming.StreamUtil
 
touchFile(String) - Method in class org.apache.hadoop.contrib.index.lucene.FileSystemDirectory
 
tryDivide(BigDecimal, BigDecimal) - Method in class org.apache.hadoop.mapreduce.lib.db.BigDecimalSplitter
Divide numerator by denominator.
TupleWritable - Class in org.apache.hadoop.mapred.join
Writable type storing multiple Writables.
TupleWritable() - Constructor for class org.apache.hadoop.mapred.join.TupleWritable
Create an empty tuple with no allocated storage for writables.
TupleWritable(Writable[]) - Constructor for class org.apache.hadoop.mapred.join.TupleWritable
Initialize tuple with storage; unknown whether any of them contain "written" values.
TupleWritable - Class in org.apache.hadoop.mapreduce.lib.join
Writable type storing multiple Writables.
TupleWritable() - Constructor for class org.apache.hadoop.mapreduce.lib.join.TupleWritable
Create an empty tuple with no allocated storage for writables.
TupleWritable(Writable[]) - Constructor for class org.apache.hadoop.mapreduce.lib.join.TupleWritable
Initialize tuple with storage; unknown whether any of them contain "written" values.
twoRotations - Static variable in class org.apache.hadoop.examples.dancing.Pentomino
Is the piece identical if rotated 180 degrees?
Type - Enum in org.apache.hadoop.typedbytes
The possible type codes.
TYPE_SEPARATOR - Static variable in interface org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorDescriptor
 
TYPE_SEPARATOR - Static variable in interface org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorDescriptor
 
TYPED_BYTES_ID - Static variable in class org.apache.hadoop.streaming.io.IdentifierResolver
 
TypedBytesInput - Class in org.apache.hadoop.typedbytes
Provides functionality for reading typed bytes.
TypedBytesInput(DataInput) - Constructor for class org.apache.hadoop.typedbytes.TypedBytesInput
Creates a new instance of TypedBytesInput.
TypedBytesInputWriter - Class in org.apache.hadoop.streaming.io
InputWriter that writes the client's input as typed bytes.
TypedBytesInputWriter() - Constructor for class org.apache.hadoop.streaming.io.TypedBytesInputWriter
 
TypedBytesOutput - Class in org.apache.hadoop.typedbytes
Provides functionality for writing typed bytes.
TypedBytesOutput(DataOutput) - Constructor for class org.apache.hadoop.typedbytes.TypedBytesOutput
Creates a new instance of TypedBytesOutput.
TypedBytesOutputReader - Class in org.apache.hadoop.streaming.io
OutputReader that reads the client's output as typed bytes.
TypedBytesOutputReader() - Constructor for class org.apache.hadoop.streaming.io.TypedBytesOutputReader
 
TypedBytesRecordInput - Class in org.apache.hadoop.typedbytes
Serializer for records that writes typed bytes.
TypedBytesRecordInput(TypedBytesInput) - Constructor for class org.apache.hadoop.typedbytes.TypedBytesRecordInput
Creates a new instance of TypedBytesRecordInput.
TypedBytesRecordInput(DataInput) - Constructor for class org.apache.hadoop.typedbytes.TypedBytesRecordInput
Creates a new instance of TypedBytesRecordInput.
TypedBytesRecordOutput - Class in org.apache.hadoop.typedbytes
Deserialized for records that reads typed bytes.
TypedBytesRecordOutput(TypedBytesOutput) - Constructor for class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
Creates a new instance of TypedBytesRecordOutput.
TypedBytesRecordOutput(DataOutput) - Constructor for class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
Creates a new instance of TypedBytesRecordOutput.
TypedBytesWritable - Class in org.apache.hadoop.typedbytes
Writable for typed bytes.
TypedBytesWritable() - Constructor for class org.apache.hadoop.typedbytes.TypedBytesWritable
Create a TypedBytesWritable.
TypedBytesWritable(byte[]) - Constructor for class org.apache.hadoop.typedbytes.TypedBytesWritable
Create a TypedBytesWritable with a given byte array as initial value.
TypedBytesWritableInput - Class in org.apache.hadoop.typedbytes
Provides functionality for reading typed bytes as Writable objects.
TypedBytesWritableInput(TypedBytesInput) - Constructor for class org.apache.hadoop.typedbytes.TypedBytesWritableInput
Creates a new instance of TypedBytesWritableInput.
TypedBytesWritableInput(DataInput) - Constructor for class org.apache.hadoop.typedbytes.TypedBytesWritableInput
Creates a new instance of TypedBytesWritableInput.
TypedBytesWritableOutput - Class in org.apache.hadoop.typedbytes
Provides functionality for writing Writable objects as typed bytes.
TypedBytesWritableOutput(TypedBytesOutput) - Constructor for class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
Creates a new instance of TypedBytesWritableOutput.
TypedBytesWritableOutput(DataOutput) - Constructor for class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
Creates a new instance of TypedBytesWritableOutput.

U

UNIQ_VALUE_COUNT - Static variable in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor
 
UNIQ_VALUE_COUNT - Static variable in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorBaseDescriptor
 
UniqValueCount - Class in org.apache.hadoop.mapred.lib.aggregate
This class implements a value aggregator that dedupes a sequence of objects.
UniqValueCount() - Constructor for class org.apache.hadoop.mapred.lib.aggregate.UniqValueCount
the default constructor
UniqValueCount(long) - Constructor for class org.apache.hadoop.mapred.lib.aggregate.UniqValueCount
constructor
UniqValueCount - Class in org.apache.hadoop.mapreduce.lib.aggregate
This class implements a value aggregator that dedupes a sequence of objects.
UniqValueCount() - Constructor for class org.apache.hadoop.mapreduce.lib.aggregate.UniqValueCount
the default constructor
UniqValueCount(long) - Constructor for class org.apache.hadoop.mapreduce.lib.aggregate.UniqValueCount
constructor
UNPACK_JAR_PATTERN_DEFAULT - Static variable in class org.apache.hadoop.mapred.JobConf
Pattern for the default unpacking behavior for job jars
UPDATE - Static variable in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp.Op
 
update(Statistics.JobStats) - Method in class org.apache.hadoop.mapred.gridmix.SerialJobFactory
SERIAL.
update(Statistics.ClusterStats) - Method in class org.apache.hadoop.mapred.gridmix.StressJobFactory

STRESS Once you get the notification from StatsCollector.Collect the clustermetrics.

UpdateIndex - Class in org.apache.hadoop.contrib.index.main
A distributed "index" is partitioned into "shards".
UpdateIndex() - Constructor for class org.apache.hadoop.contrib.index.main.UpdateIndex
 
UPPER_LIMIT_ON_TASK_VMEM_PROPERTY - Static variable in class org.apache.hadoop.mapred.JobConf
Deprecated.  
URL_PROPERTY - Static variable in class org.apache.hadoop.mapred.lib.db.DBConfiguration
JDBC Database access URL
URL_PROPERTY - Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
JDBC Database access URL
USED_GENERIC_PARSER - Static variable in class org.apache.hadoop.mapreduce.Job
 
USER_JAR - Static variable in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorJobBase
 
UserDefinedValueAggregatorDescriptor - Class in org.apache.hadoop.mapred.lib.aggregate
This class implements a wrapper for a user defined value aggregator descriptor.
UserDefinedValueAggregatorDescriptor(String, JobConf) - Constructor for class org.apache.hadoop.mapred.lib.aggregate.UserDefinedValueAggregatorDescriptor
 
UserDefinedValueAggregatorDescriptor - Class in org.apache.hadoop.mapreduce.lib.aggregate
This class implements a wrapper for a user defined value aggregator descriptor.
UserDefinedValueAggregatorDescriptor(String, Configuration) - Constructor for class org.apache.hadoop.mapreduce.lib.aggregate.UserDefinedValueAggregatorDescriptor
 
USERNAME_PROPERTY - Static variable in class org.apache.hadoop.mapred.lib.db.DBConfiguration
User name to access the database
USERNAME_PROPERTY - Static variable in class org.apache.hadoop.mapreduce.lib.db.DBConfiguration
User name to access the database
UTF8ByteArrayUtils - Class in org.apache.hadoop.streaming
Deprecated. use UTF8ByteArrayUtils and StreamKeyValUtil instead
UTF8ByteArrayUtils() - Constructor for class org.apache.hadoop.streaming.UTF8ByteArrayUtils
Deprecated.  
Util - Class in org.apache.hadoop.examples.pi
Utility methods
Util() - Constructor for class org.apache.hadoop.examples.pi.Util
 
Util.Timer - Class in org.apache.hadoop.examples.pi
Timer
Util.Timer(boolean) - Constructor for class org.apache.hadoop.examples.pi.Util.Timer
Timer constructor
Utils - Class in org.apache.hadoop.mapred
A utility class.
Utils() - Constructor for class org.apache.hadoop.mapred.Utils
 
Utils.OutputFileUtils - Class in org.apache.hadoop.mapred
 
Utils.OutputFileUtils() - Constructor for class org.apache.hadoop.mapred.Utils.OutputFileUtils
 
Utils.OutputFileUtils.OutputFilesFilter - Class in org.apache.hadoop.mapred
This class filters output(part) files from the given directory It does not accept files with filenames _logs and _SUCCESS.
Utils.OutputFileUtils.OutputFilesFilter() - Constructor for class org.apache.hadoop.mapred.Utils.OutputFileUtils.OutputFilesFilter
 
Utils.OutputFileUtils.OutputLogFilter - Class in org.apache.hadoop.mapred
This class filters log files from directory given It doesnt accept paths having _logs.
Utils.OutputFileUtils.OutputLogFilter() - Constructor for class org.apache.hadoop.mapred.Utils.OutputFileUtils.OutputLogFilter
 

V

value - Variable in class org.apache.hadoop.examples.pi.math.ArithmeticProgression
Starting value
value - Variable in class org.apache.hadoop.mapreduce.lib.join.CompositeRecordReader
 
VALUE_CLASS - Static variable in class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat
 
VALUE_HISTOGRAM - Static variable in class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor
 
VALUE_HISTOGRAM - Static variable in class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorBaseDescriptor
 
ValueAggregator<E> - Interface in org.apache.hadoop.mapred.lib.aggregate
This interface defines the minimal protocol for value aggregators.
ValueAggregator<E> - Interface in org.apache.hadoop.mapreduce.lib.aggregate
This interface defines the minimal protocol for value aggregators.
ValueAggregatorBaseDescriptor - Class in org.apache.hadoop.mapred.lib.aggregate
This class implements the common functionalities of the subclasses of ValueAggregatorDescriptor class.
ValueAggregatorBaseDescriptor() - Constructor for class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorBaseDescriptor
 
ValueAggregatorBaseDescriptor - Class in org.apache.hadoop.mapreduce.lib.aggregate
This class implements the common functionalities of the subclasses of ValueAggregatorDescriptor class.
ValueAggregatorBaseDescriptor() - Constructor for class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorBaseDescriptor
 
ValueAggregatorCombiner<K1 extends org.apache.hadoop.io.WritableComparable,V1 extends org.apache.hadoop.io.Writable> - Class in org.apache.hadoop.mapred.lib.aggregate
This class implements the generic combiner of Aggregate.
ValueAggregatorCombiner() - Constructor for class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorCombiner
 
ValueAggregatorCombiner<K1 extends org.apache.hadoop.io.WritableComparable<?>,V1 extends org.apache.hadoop.io.Writable> - Class in org.apache.hadoop.mapreduce.lib.aggregate
This class implements the generic combiner of Aggregate.
ValueAggregatorCombiner() - Constructor for class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorCombiner
 
ValueAggregatorDescriptor - Interface in org.apache.hadoop.mapred.lib.aggregate
This interface defines the contract a value aggregator descriptor must support.
ValueAggregatorDescriptor - Interface in org.apache.hadoop.mapreduce.lib.aggregate
This interface defines the contract a value aggregator descriptor must support.
ValueAggregatorJob - Class in org.apache.hadoop.mapred.lib.aggregate
This is the main class for creating a map/reduce job using Aggregate framework.
ValueAggregatorJob() - Constructor for class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJob
 
ValueAggregatorJob - Class in org.apache.hadoop.mapreduce.lib.aggregate
This is the main class for creating a map/reduce job using Aggregate framework.
ValueAggregatorJob() - Constructor for class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorJob
 
ValueAggregatorJobBase<K1 extends org.apache.hadoop.io.WritableComparable,V1 extends org.apache.hadoop.io.Writable> - Class in org.apache.hadoop.mapred.lib.aggregate
This abstract class implements some common functionalities of the the generic mapper, reducer and combiner classes of Aggregate.
ValueAggregatorJobBase() - Constructor for class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorJobBase
 
ValueAggregatorJobBase<K1 extends org.apache.hadoop.io.WritableComparable<?>,V1 extends org.apache.hadoop.io.Writable> - Class in org.apache.hadoop.mapreduce.lib.aggregate
This abstract class implements some common functionalities of the the generic mapper, reducer and combiner classes of Aggregate.
ValueAggregatorJobBase() - Constructor for class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorJobBase
 
ValueAggregatorMapper<K1 extends org.apache.hadoop.io.WritableComparable,V1 extends org.apache.hadoop.io.Writable> - Class in org.apache.hadoop.mapred.lib.aggregate
This class implements the generic mapper of Aggregate.
ValueAggregatorMapper() - Constructor for class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorMapper
 
ValueAggregatorMapper<K1 extends org.apache.hadoop.io.WritableComparable<?>,V1 extends org.apache.hadoop.io.Writable> - Class in org.apache.hadoop.mapreduce.lib.aggregate
This class implements the generic mapper of Aggregate.
ValueAggregatorMapper() - Constructor for class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorMapper
 
ValueAggregatorReducer<K1 extends org.apache.hadoop.io.WritableComparable,V1 extends org.apache.hadoop.io.Writable> - Class in org.apache.hadoop.mapred.lib.aggregate
This class implements the generic reducer of Aggregate.
ValueAggregatorReducer() - Constructor for class org.apache.hadoop.mapred.lib.aggregate.ValueAggregatorReducer
 
ValueAggregatorReducer<K1 extends org.apache.hadoop.io.WritableComparable<?>,V1 extends org.apache.hadoop.io.Writable> - Class in org.apache.hadoop.mapreduce.lib.aggregate
This class implements the generic reducer of Aggregate.
ValueAggregatorReducer() - Constructor for class org.apache.hadoop.mapreduce.lib.aggregate.ValueAggregatorReducer
 
ValueHistogram - Class in org.apache.hadoop.mapred.lib.aggregate
This class implements a value aggregator that computes the histogram of a sequence of strings.
ValueHistogram() - Constructor for class org.apache.hadoop.mapred.lib.aggregate.ValueHistogram
 
ValueHistogram - Class in org.apache.hadoop.mapreduce.lib.aggregate
This class implements a value aggregator that computes the histogram of a sequence of strings.
ValueHistogram() - Constructor for class org.apache.hadoop.mapreduce.lib.aggregate.ValueHistogram
 
valueOf(String) - Static method in enum org.apache.hadoop.examples.dancing.Pentomino.SolutionCategory
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.hadoop.examples.pi.math.Bellard.Parameter
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in class org.apache.hadoop.examples.pi.math.Summation
Covert a String to a Summation.
valueOf(String) - Static method in class org.apache.hadoop.examples.pi.TaskResult
Covert a String to a TaskResult
valueOf(String) - Static method in enum org.apache.hadoop.examples.terasort.TeraGen.Counters
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.hadoop.mapred.JobClient.TaskStatusFilter
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.hadoop.mapred.JobPriority
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.hadoop.mapred.join.Parser.TType
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.hadoop.mapred.TaskCompletionEvent.Status
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.hadoop.mapreduce.Job.JobState
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.hadoop.mapreduce.Job.TaskStatusFilter
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.hadoop.mapreduce.JobCounter
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.hadoop.mapreduce.JobPriority
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.hadoop.mapreduce.JobStatus.State
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.State
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl.ThreadState
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.hadoop.mapreduce.lib.join.Parser.TType
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.hadoop.mapreduce.QueueState
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.hadoop.mapreduce.server.jobtracker.State
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.hadoop.mapreduce.TaskCompletionEvent.Status
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.hadoop.mapreduce.TaskCounter
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.hadoop.mapreduce.TaskType
Returns the enum constant of this type with the specified name.
valueOf(String) - Static method in enum org.apache.hadoop.typedbytes.Type
Returns the enum constant of this type with the specified name.
values() - Static method in enum org.apache.hadoop.examples.dancing.Pentomino.SolutionCategory
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.hadoop.examples.pi.math.Bellard.Parameter
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.hadoop.examples.terasort.TeraGen.Counters
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.hadoop.mapred.JobClient.TaskStatusFilter
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.hadoop.mapred.JobPriority
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.hadoop.mapred.join.Parser.TType
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.hadoop.mapred.TaskCompletionEvent.Status
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.hadoop.mapreduce.Job.JobState
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.hadoop.mapreduce.Job.TaskStatusFilter
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.hadoop.mapreduce.JobCounter
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.hadoop.mapreduce.JobPriority
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.hadoop.mapreduce.JobStatus.State
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.hadoop.mapreduce.lib.jobcontrol.ControlledJob.State
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.hadoop.mapreduce.lib.jobcontrol.JobControl.ThreadState
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.hadoop.mapreduce.lib.join.Parser.TType
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.hadoop.mapreduce.QueueState
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.hadoop.mapreduce.server.jobtracker.State
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.hadoop.mapreduce.TaskCompletionEvent.Status
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.hadoop.mapreduce.TaskCounter
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.hadoop.mapreduce.TaskType
Returns an array containing the constants of this enum type, in the order they are declared.
values() - Static method in enum org.apache.hadoop.typedbytes.Type
Returns an array containing the constants of this enum type, in the order they are declared.
verbose - Variable in class org.apache.hadoop.streaming.JarBuilder
 
verbose_ - Variable in class org.apache.hadoop.streaming.StreamJob
 

W

waitForCompletion() - Method in interface org.apache.hadoop.mapred.RunningJob
Blocks until the job is complete.
waitForCompletion(boolean) - Method in class org.apache.hadoop.mapreduce.Job
Submit the job to the cluster and wait for it to finish.
WAITING - Static variable in class org.apache.hadoop.mapred.jobcontrol.Job
 
WIDTH - Static variable in class org.apache.hadoop.examples.dancing.Pentomino
 
width - Variable in class org.apache.hadoop.examples.dancing.Pentomino
 
WordCount - Class in org.apache.hadoop.examples
 
WordCount() - Constructor for class org.apache.hadoop.examples.WordCount
 
WordCount.IntSumReducer - Class in org.apache.hadoop.examples
 
WordCount.IntSumReducer() - Constructor for class org.apache.hadoop.examples.WordCount.IntSumReducer
 
WordCount.TokenizerMapper - Class in org.apache.hadoop.examples
 
WordCount.TokenizerMapper() - Constructor for class org.apache.hadoop.examples.WordCount.TokenizerMapper
 
WrappedMapper<KEYIN,VALUEIN,KEYOUT,VALUEOUT> - Class in org.apache.hadoop.mapreduce.lib.map
A Mapper which wraps a given one to allow custom WrappedMapper.Context implementations.
WrappedMapper() - Constructor for class org.apache.hadoop.mapreduce.lib.map.WrappedMapper
 
WrappedMapper.Context - Class in org.apache.hadoop.mapreduce.lib.map
 
WrappedMapper.Context(MapContext<KEYIN, VALUEIN, KEYOUT, VALUEOUT>) - Constructor for class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
WrappedRecordReader<K extends org.apache.hadoop.io.WritableComparable,U extends org.apache.hadoop.io.Writable> - Class in org.apache.hadoop.mapred.join
Proxy class for a RecordReader participating in the join framework.
WrappedRecordReader<K extends org.apache.hadoop.io.WritableComparable<?>,U extends org.apache.hadoop.io.Writable> - Class in org.apache.hadoop.mapreduce.lib.join
Proxy class for a RecordReader participating in the join framework.
WrappedRecordReader(int) - Constructor for class org.apache.hadoop.mapreduce.lib.join.WrappedRecordReader
 
WrappedReducer<KEYIN,VALUEIN,KEYOUT,VALUEOUT> - Class in org.apache.hadoop.mapreduce.lib.reduce
A Reducer which wraps a given one to allow for custom WrappedReducer.Context implementations.
WrappedReducer() - Constructor for class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer
 
WrappedReducer.Context - Class in org.apache.hadoop.mapreduce.lib.reduce
 
WrappedReducer.Context(ReduceContext<KEYIN, VALUEIN, KEYOUT, VALUEOUT>) - Constructor for class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
write(DataOutput) - Method in class org.apache.hadoop.contrib.index.example.LineDocTextAndOp
 
write(DataOutput) - Method in class org.apache.hadoop.contrib.index.mapred.DocumentAndOp
 
write(DataOutput) - Method in class org.apache.hadoop.contrib.index.mapred.DocumentID
 
write(DataOutput) - Method in class org.apache.hadoop.contrib.index.mapred.IntermediateForm
 
write(DataOutput) - Method in class org.apache.hadoop.contrib.index.mapred.Shard
 
write(DataOutput) - Method in class org.apache.hadoop.examples.BaileyBorweinPlouffe.BbpSplit
write(DataOutput) - Method in class org.apache.hadoop.examples.MultiFileWordCount.WordOffset
 
write(DataOutput) - Method in class org.apache.hadoop.examples.pi.DistSum.Machine.SummationSplit
write(Summation, Class<?>, Configuration) - Static method in class org.apache.hadoop.examples.pi.SummationWritable
Write sigma to conf
write(Summation, DataOutput) - Static method in class org.apache.hadoop.examples.pi.SummationWritable
Write sigma to DataOutput
write(DataOutput) - Method in class org.apache.hadoop.examples.pi.SummationWritable
write(DataOutput) - Method in class org.apache.hadoop.examples.pi.TaskResult
write(DataOutput) - Method in class org.apache.hadoop.examples.SecondarySort.IntPair
 
write(DataOutput) - Method in class org.apache.hadoop.mapred.ClusterStatus.BlackListInfo
 
write(DataOutput) - Method in class org.apache.hadoop.mapred.ClusterStatus
 
write(DataOutput) - Method in class org.apache.hadoop.mapred.Counters.Group
 
write(DataOutput) - Method in class org.apache.hadoop.mapred.Counters
Write the set of groups.
write(DataOutput) - Method in class org.apache.hadoop.mapred.FileSplit
 
write(DataOutput) - Method in class org.apache.hadoop.mapred.join.CompositeInputSplit
Write splits in the following format.
write(K, V) - Method in class org.apache.hadoop.mapred.lib.FilterOutputFormat.FilterRecordWriter
 
write(K, V) - Method in interface org.apache.hadoop.mapred.RecordWriter
Writes a key/value pair.
write(K, V) - Method in class org.apache.hadoop.mapred.TextOutputFormat.LineRecordWriter
 
write(DataOutput) - Method in class org.apache.hadoop.mapreduce.ClusterMetrics
 
write(DataOutput) - Method in class org.apache.hadoop.mapreduce.Counter
Write the binary representation of the counter
write(DataOutput) - Method in class org.apache.hadoop.mapreduce.CounterGroup
 
write(DataOutput) - Method in class org.apache.hadoop.mapreduce.Counters
Write the set of groups.
write(DataOutput) - Method in class org.apache.hadoop.mapreduce.ID
 
write(DataOutput) - Method in class org.apache.hadoop.mapreduce.JobID
 
write(DataOutput) - Method in class org.apache.hadoop.mapreduce.JobStatus
 
write(DataOutput) - Method in class org.apache.hadoop.mapreduce.lib.db.DataDrivenDBInputFormat.DataDrivenDBInputSplit
write(DataOutput) - Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat.DBInputSplit
write(DataOutput) - Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat.NullDBWritable
 
write(PreparedStatement) - Method in class org.apache.hadoop.mapreduce.lib.db.DBInputFormat.NullDBWritable
 
write(K, V) - Method in class org.apache.hadoop.mapreduce.lib.db.DBOutputFormat.DBRecordWriter
Writes a key/value pair.
write(PreparedStatement) - Method in interface org.apache.hadoop.mapreduce.lib.db.DBWritable
Sets the fields of the object in the PreparedStatement.
write(DataOutput) - Method in class org.apache.hadoop.mapreduce.lib.input.CombineFileSplit
 
write(DataOutput) - Method in class org.apache.hadoop.mapreduce.lib.input.FileSplit
 
write(DataOutput) - Method in class org.apache.hadoop.mapreduce.lib.join.CompositeInputSplit
Write splits in the following format.
write(DataOutput) - Method in class org.apache.hadoop.mapreduce.lib.join.TupleWritable
Writes each Writable to out.
write(KEYOUT, VALUEOUT) - Method in class org.apache.hadoop.mapreduce.lib.map.WrappedMapper.Context
 
write(K, V) - Method in class org.apache.hadoop.mapreduce.lib.output.FilterOutputFormat.FilterRecordWriter
 
write(String, K, V) - Method in class org.apache.hadoop.mapreduce.lib.output.MultipleOutputs
Write key and value to the namedOutput.
write(String, K, V, String) - Method in class org.apache.hadoop.mapreduce.lib.output.MultipleOutputs
Write key and value to baseOutputPath using the namedOutput.
write(KEYOUT, VALUEOUT, String) - Method in class org.apache.hadoop.mapreduce.lib.output.MultipleOutputs
Write key value to an output file name.
write(K, V) - Method in class org.apache.hadoop.mapreduce.lib.output.TextOutputFormat.LineRecordWriter
 
write(KEYOUT, VALUEOUT) - Method in class org.apache.hadoop.mapreduce.lib.reduce.WrappedReducer.Context
 
write(DataOutput) - Method in class org.apache.hadoop.mapreduce.QueueAclsInfo
 
write(DataOutput) - Method in class org.apache.hadoop.mapreduce.QueueInfo
 
write(K, V) - Method in class org.apache.hadoop.mapreduce.RecordWriter
Writes a key/value pair.
write(DataOutput) - Method in class org.apache.hadoop.mapreduce.split.JobSplit.SplitMetaInfo
 
write(DataOutput) - Method in class org.apache.hadoop.mapreduce.split.JobSplit.TaskSplitIndex
 
write(DataOutput) - Method in class org.apache.hadoop.mapreduce.TaskAttemptID
 
write(DataOutput) - Method in class org.apache.hadoop.mapreduce.TaskCompletionEvent
 
write(DataOutput) - Method in class org.apache.hadoop.mapreduce.TaskID
 
write(KEYOUT, VALUEOUT) - Method in interface org.apache.hadoop.mapreduce.TaskInputOutputContext
Generate an output key/value pair.
write(DataOutput) - Method in class org.apache.hadoop.mapreduce.TaskTrackerInfo
 
write(Object) - Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
Writes a Java object as a typed bytes sequence.
write(Writable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
 
writeArray(ArrayWritable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
 
writeBool(boolean) - Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
Writes a boolean as a typed bytes sequence.
writeBool(boolean, String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
 
writeBoolean(BooleanWritable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
 
writeBuffer(Buffer, String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
 
writeByte(byte) - Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
Writes a byte as a typed bytes sequence.
writeByte(byte, String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
 
writeByte(ByteWritable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
 
writeBytes(byte[], int, int) - Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
Writes a bytes array as a typed bytes sequence, using a given typecode and length.
writeBytes(byte[], int) - Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
Writes a bytes array as a typed bytes sequence, using a given typecode.
writeBytes(byte[]) - Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
Writes a bytes array as a typed bytes sequence.
writeBytes(Buffer) - Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
Writes a bytes buffer as a typed bytes sequence.
writeBytes(BytesWritable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
 
writeCompressedBytes(DataOutputStream) - Method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat.WritableValueBytes
 
writeDouble(double) - Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
Writes a double as a typed bytes sequence.
writeDouble(double, String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
 
writeDouble(DoubleWritable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
 
writeFloat(float) - Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
Writes a float as a typed bytes sequence.
writeFloat(float, String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
 
writeFloat(FloatWritable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
 
writeInputData(long, Path) - Method in class org.apache.hadoop.mapred.gridmix.Gridmix
Write random bytes at the path provided.
writeInt(int) - Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
Writes an integer as a typed bytes sequence.
writeInt(int, String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
 
writeInt(IntWritable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
 
writeKey(K) - Method in class org.apache.hadoop.streaming.io.InputWriter
Writes an input key.
writeKey(Writable) - Method in class org.apache.hadoop.streaming.io.RawBytesInputWriter
 
writeKey(Object) - Method in class org.apache.hadoop.streaming.io.TextInputWriter
 
writeKey(Object) - Method in class org.apache.hadoop.streaming.io.TypedBytesInputWriter
 
writeList(List) - Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
Writes a list as a typed bytes sequence.
writeListFooter() - Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
Writes a list footer.
writeListHeader() - Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
Writes a list header.
writeLong(long) - Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
Writes a long as a typed bytes sequence.
writeLong(long, String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
 
writeLong(LongWritable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
 
writeMap(Map) - Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
Writes a map as a typed bytes sequence.
writeMap(MapWritable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
 
writeMapHeader(int) - Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
Writes a map header.
writePartitionFile(JobContext, Path) - Static method in class org.apache.hadoop.examples.terasort.TeraInputFormat
Use the input splits to take samples of the input and generate sample keys.
writePartitionFile(JobConf, InputSampler.Sampler<K, V>) - Static method in class org.apache.hadoop.mapred.lib.InputSampler
 
writePartitionFile(Job, InputSampler.Sampler<K, V>) - Static method in class org.apache.hadoop.mapreduce.lib.partition.InputSampler
Write a partition file for the given job, using the Sampler provided.
writeRAMFiles(DataOutput, RAMDirectory, String[]) - Static method in class org.apache.hadoop.contrib.index.lucene.RAMDirectoryUtil
Write a number of files from a ram directory to a data output.
writeRaw(byte[]) - Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
Writes a raw sequence of typed bytes.
writeRaw(byte[], int, int) - Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
Writes a raw sequence of typed bytes.
writeRecord(Record) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
 
writeSortedMap(SortedMapWritable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
 
writeString(String) - Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
Writes a string as a typed bytes sequence.
writeString(String, String) - Method in class org.apache.hadoop.typedbytes.TypedBytesRecordOutput
 
writeText(Text) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
 
writeTypedBytes(TypedBytesWritable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
 
writeUncompressedBytes(DataOutputStream) - Method in class org.apache.hadoop.mapreduce.lib.output.SequenceFileAsBinaryOutputFormat.WritableValueBytes
 
writeValue(V) - Method in class org.apache.hadoop.streaming.io.InputWriter
Writes an input value.
writeValue(Writable) - Method in class org.apache.hadoop.streaming.io.RawBytesInputWriter
 
writeValue(Object) - Method in class org.apache.hadoop.streaming.io.TextInputWriter
 
writeValue(Object) - Method in class org.apache.hadoop.streaming.io.TypedBytesInputWriter
 
writeVector(ArrayList) - Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
Writes a vector as a typed bytes sequence.
writeVectorHeader(int) - Method in class org.apache.hadoop.typedbytes.TypedBytesOutput
Writes a vector header.
writeVInt(VIntWritable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
 
writeVLong(VLongWritable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
 
writeWritable(Writable) - Method in class org.apache.hadoop.typedbytes.TypedBytesWritableOutput
 
written - Variable in class org.apache.hadoop.mapreduce.lib.join.TupleWritable
 

A B C D E F G H I J K L M N O P Q R S T U V W

Copyright © 2009 The Apache Software Foundation