Uses of Class
org.apache.hadoop.mapreduce.Job

Packages that use Job
org.apache.hadoop.examples Hadoop example code. 
org.apache.hadoop.examples.pi This package consists of a map/reduce application, distbbp, which computes exact binary digits of the mathematical constant π. 
org.apache.hadoop.examples.terasort This package consists of 3 map/reduce applications for Hadoop to compete in the annual terabyte sort competition. 
org.apache.hadoop.mapred.gridmix   
org.apache.hadoop.mapreduce   
org.apache.hadoop.mapreduce.lib.aggregate Classes for performing various counting and aggregations. 
org.apache.hadoop.mapreduce.lib.chain   
org.apache.hadoop.mapreduce.lib.db org.apache.hadoop.mapred.lib.db Package 
org.apache.hadoop.mapreduce.lib.input   
org.apache.hadoop.mapreduce.lib.jobcontrol Utilities for managing dependent jobs. 
org.apache.hadoop.mapreduce.lib.map   
org.apache.hadoop.mapreduce.lib.output   
org.apache.hadoop.mapreduce.lib.partition   
org.apache.hadoop.mapreduce.tools   
 

Uses of Job in org.apache.hadoop.examples
 

Methods in org.apache.hadoop.examples that return Job
 Job Sort.getResult()
          Get the last job that was run using this instance.
 

Uses of Job in org.apache.hadoop.examples.pi
 

Methods in org.apache.hadoop.examples.pi with parameters of type Job
 void DistSum.MapSide.init(Job job)
          Initialize a job
 void DistSum.ReduceSide.init(Job job)
          Initialize a job
 void DistSum.MixMachine.init(Job job)
          Initialize a job
 

Uses of Job in org.apache.hadoop.examples.terasort
 

Methods in org.apache.hadoop.examples.terasort with parameters of type Job
static void TeraSort.setOutputReplication(Job job, int value)
           
static void TeraSort.setUseSimplePartitioner(Job job, boolean value)
           
 

Uses of Job in org.apache.hadoop.mapred.gridmix
 

Methods in org.apache.hadoop.mapred.gridmix with parameters of type Job
 void Statistics.add(Job job)
          Used by JobMonitor to add the completed job.
 

Uses of Job in org.apache.hadoop.mapreduce
 

Methods in org.apache.hadoop.mapreduce that return Job
 Job[] Cluster.getAllJobs()
          Get all the jobs in cluster.
static Job Job.getInstance(Cluster cluster)
           
static Job Job.getInstance(Cluster cluster, org.apache.hadoop.conf.Configuration conf)
           
static Job Job.getInstance(Cluster cluster, JobStatus status, org.apache.hadoop.conf.Configuration conf)
           
 Job Cluster.getJob(JobID jobId)
          Get job corresponding to jobid.
 

Uses of Job in org.apache.hadoop.mapreduce.lib.aggregate
 

Methods in org.apache.hadoop.mapreduce.lib.aggregate that return Job
static Job ValueAggregatorJob.createValueAggregatorJob(org.apache.hadoop.conf.Configuration conf, String[] args)
          Create an Aggregate based map/reduce job.
static Job ValueAggregatorJob.createValueAggregatorJob(String[] args, Class<? extends ValueAggregatorDescriptor>[] descriptors)
           
 

Uses of Job in org.apache.hadoop.mapreduce.lib.chain
 

Methods in org.apache.hadoop.mapreduce.lib.chain with parameters of type Job
static void ChainReducer.addMapper(Job job, Class<? extends Mapper> klass, Class<?> inputKeyClass, Class<?> inputValueClass, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.conf.Configuration mapperConf)
          Adds a Mapper class to the chain reducer.
static void ChainMapper.addMapper(Job job, Class<? extends Mapper> klass, Class<?> inputKeyClass, Class<?> inputValueClass, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.conf.Configuration mapperConf)
          Adds a Mapper class to the chain mapper.
static void ChainReducer.setReducer(Job job, Class<? extends Reducer> klass, Class<?> inputKeyClass, Class<?> inputValueClass, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.conf.Configuration reducerConf)
          Sets the Reducer class to the chain job.
 

Uses of Job in org.apache.hadoop.mapreduce.lib.db
 

Methods in org.apache.hadoop.mapreduce.lib.db with parameters of type Job
static void DBInputFormat.setInput(Job job, Class<? extends DBWritable> inputClass, String inputQuery, String inputCountQuery)
          Initializes the map-part of the job with the appropriate input settings.
static void DataDrivenDBInputFormat.setInput(Job job, Class<? extends DBWritable> inputClass, String inputQuery, String inputBoundingQuery)
          setInput() takes a custom query and a separate "bounding query" to use instead of the custom "count query" used by DBInputFormat.
static void DBInputFormat.setInput(Job job, Class<? extends DBWritable> inputClass, String tableName, String conditions, String orderBy, String... fieldNames)
          Initializes the map-part of the job with the appropriate input settings.
static void DataDrivenDBInputFormat.setInput(Job job, Class<? extends DBWritable> inputClass, String tableName, String conditions, String splitBy, String... fieldNames)
          Note that the "orderBy" column is called the "splitBy" in this version.
static void DBOutputFormat.setOutput(Job job, String tableName, int fieldCount)
          Initializes the reduce-part of the job with the appropriate output settings
static void DBOutputFormat.setOutput(Job job, String tableName, String... fieldNames)
          Initializes the reduce-part of the job with the appropriate output settings
 

Uses of Job in org.apache.hadoop.mapreduce.lib.input
 

Methods in org.apache.hadoop.mapreduce.lib.input with parameters of type Job
static void FileInputFormat.addInputPath(Job job, org.apache.hadoop.fs.Path path)
          Add a Path to the list of inputs for the map-reduce job.
static void MultipleInputs.addInputPath(Job job, org.apache.hadoop.fs.Path path, Class<? extends InputFormat> inputFormatClass)
          Add a Path with a custom InputFormat to the list of inputs for the map-reduce job.
static void MultipleInputs.addInputPath(Job job, org.apache.hadoop.fs.Path path, Class<? extends InputFormat> inputFormatClass, Class<? extends Mapper> mapperClass)
          Add a Path with a custom InputFormat and Mapper to the list of inputs for the map-reduce job.
static void FileInputFormat.addInputPaths(Job job, String commaSeparatedPaths)
          Add the given comma separated paths to the list of inputs for the map-reduce job.
static void SequenceFileInputFilter.setFilterClass(Job job, Class<?> filterClass)
          set the filter class
static void FileInputFormat.setInputPathFilter(Job job, Class<? extends org.apache.hadoop.fs.PathFilter> filter)
          Set a PathFilter to be applied to the input paths for the map-reduce job.
static void FileInputFormat.setInputPaths(Job job, org.apache.hadoop.fs.Path... inputPaths)
          Set the array of Paths as the list of inputs for the map-reduce job.
static void FileInputFormat.setInputPaths(Job job, String commaSeparatedPaths)
          Sets the given comma separated paths as the list of inputs for the map-reduce job.
static void FileInputFormat.setMaxInputSplitSize(Job job, long size)
          Set the maximum split size
static void FileInputFormat.setMinInputSplitSize(Job job, long size)
          Set the minimum input split size
static void NLineInputFormat.setNumLinesPerSplit(Job job, int numLines)
          Set the number of lines per split
 

Uses of Job in org.apache.hadoop.mapreduce.lib.jobcontrol
 

Methods in org.apache.hadoop.mapreduce.lib.jobcontrol that return Job
 Job ControlledJob.getJob()
           
 

Methods in org.apache.hadoop.mapreduce.lib.jobcontrol with parameters of type Job
 void ControlledJob.setJob(Job job)
          Set the mapreduce job
 

Constructors in org.apache.hadoop.mapreduce.lib.jobcontrol with parameters of type Job
ControlledJob(Job job, List<ControlledJob> dependingJobs)
          Construct a job.
 

Uses of Job in org.apache.hadoop.mapreduce.lib.map
 

Methods in org.apache.hadoop.mapreduce.lib.map with parameters of type Job
static
<K1,V1,K2,V2>
void
MultithreadedMapper.setMapperClass(Job job, Class<? extends Mapper<K1,V1,K2,V2>> cls)
          Set the application's mapper class.
static void MultithreadedMapper.setNumberOfThreads(Job job, int threads)
          Set the number of threads in the pool for running maps.
 

Uses of Job in org.apache.hadoop.mapreduce.lib.output
 

Methods in org.apache.hadoop.mapreduce.lib.output with parameters of type Job
static void MultipleOutputs.addNamedOutput(Job job, String namedOutput, Class<? extends OutputFormat> outputFormatClass, Class<?> keyClass, Class<?> valueClass)
          Adds a named output for the job.
static void FileOutputFormat.setCompressOutput(Job job, boolean compress)
          Set whether the output of the job is compressed.
static void MultipleOutputs.setCountersEnabled(Job job, boolean enabled)
          Enables or disables counters for the named outputs.
static void SequenceFileOutputFormat.setOutputCompressionType(Job job, org.apache.hadoop.io.SequenceFile.CompressionType style)
          Set the SequenceFile.CompressionType for the output SequenceFile.
static void FileOutputFormat.setOutputCompressorClass(Job job, Class<? extends org.apache.hadoop.io.compress.CompressionCodec> codecClass)
          Set the CompressionCodec to be used to compress job outputs.
static void LazyOutputFormat.setOutputFormatClass(Job job, Class<? extends OutputFormat> theClass)
          Set the underlying output format for LazyOutputFormat.
static void FileOutputFormat.setOutputPath(Job job, org.apache.hadoop.fs.Path outputDir)
          Set the Path of the output directory for the map-reduce job.
static void SequenceFileAsBinaryOutputFormat.setSequenceFileOutputKeyClass(Job job, Class<?> theClass)
          Set the key class for the SequenceFile
static void SequenceFileAsBinaryOutputFormat.setSequenceFileOutputValueClass(Job job, Class<?> theClass)
          Set the value class for the SequenceFile
 

Uses of Job in org.apache.hadoop.mapreduce.lib.partition
 

Methods in org.apache.hadoop.mapreduce.lib.partition with parameters of type Job
 K[] InputSampler.Sampler.getSample(InputFormat<K,V> inf, Job job)
          For a given job, collect and return a subset of the keys from the input data.
 K[] InputSampler.SplitSampler.getSample(InputFormat<K,V> inf, Job job)
          From each split sampled, take the first numSamples / numSplits records.
 K[] InputSampler.RandomSampler.getSample(InputFormat<K,V> inf, Job job)
          Randomize the split order, then take the specified number of keys from each split sampled, where each key is selected with the specified probability and possibly replaced by a subsequently selected key when the quota of keys from that split is satisfied.
 K[] InputSampler.IntervalSampler.getSample(InputFormat<K,V> inf, Job job)
          For each split sampled, emit when the ratio of the number of records retained to the total record count is less than the specified frequency.
static void KeyFieldBasedComparator.setKeyFieldComparatorOptions(Job job, String keySpec)
          Set the KeyFieldBasedComparator options used to compare keys.
 void KeyFieldBasedPartitioner.setKeyFieldPartitionerOptions(Job job, String keySpec)
          Set the KeyFieldBasedPartitioner options used for Partitioner
static
<K,V> void
InputSampler.writePartitionFile(Job job, InputSampler.Sampler<K,V> sampler)
          Write a partition file for the given job, using the Sampler provided.
 

Uses of Job in org.apache.hadoop.mapreduce.tools
 

Methods in org.apache.hadoop.mapreduce.tools with parameters of type Job
protected  void CLI.displayJobList(Job[] jobs)
           
protected  void CLI.displayTasks(Job job, String type, String state)
          Display the information about a job's tasks, of a particular type and in a particular state
 



Copyright © 2009 The Apache Software Foundation