Uses of Interface
org.apache.hadoop.mapreduce.TaskAttemptContext

Packages that use TaskAttemptContext
org.apache.hadoop.mapred   
org.apache.hadoop.mapred.lib   
org.apache.hadoop.mapreduce   
org.apache.hadoop.mapreduce.lib.db   
org.apache.hadoop.mapreduce.lib.input   
org.apache.hadoop.mapreduce.lib.join   
org.apache.hadoop.mapreduce.lib.output   
 

Uses of TaskAttemptContext in org.apache.hadoop.mapred
 

Subinterfaces of TaskAttemptContext in org.apache.hadoop.mapred
 interface TaskAttemptContext
           
 

Methods in org.apache.hadoop.mapred with parameters of type TaskAttemptContext
 void OutputCommitter.abortTask(TaskAttemptContext taskContext)
          This method implements the new interface by calling the old method.
 void OutputCommitter.commitTask(TaskAttemptContext taskContext)
          This method implements the new interface by calling the old method.
 boolean OutputCommitter.needsTaskCommit(TaskAttemptContext taskContext)
          This method implements the new interface by calling the old method.
 void OutputCommitter.recoverTask(TaskAttemptContext taskContext)
          This method implements the new interface by calling the old method.
 void OutputCommitter.setupTask(TaskAttemptContext taskContext)
          This method implements the new interface by calling the old method.
 

Uses of TaskAttemptContext in org.apache.hadoop.mapred.lib
 

Methods in org.apache.hadoop.mapred.lib with parameters of type TaskAttemptContext
 RecordReader<K,V> CombineFileInputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
           
 

Uses of TaskAttemptContext in org.apache.hadoop.mapreduce
 

Subinterfaces of TaskAttemptContext in org.apache.hadoop.mapreduce
 interface MapContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT>
          The context that is given to the Mapper.
 interface ReduceContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT>
          The context passed to the Reducer.
 interface TaskInputOutputContext<KEYIN,VALUEIN,KEYOUT,VALUEOUT>
          A context object that allows input and output from the task.
 

Methods in org.apache.hadoop.mapreduce with parameters of type TaskAttemptContext
abstract  void OutputCommitter.abortTask(TaskAttemptContext taskContext)
          Discard the task output.
abstract  void RecordWriter.close(TaskAttemptContext context)
          Close this RecordWriter to future operations.
abstract  void OutputCommitter.commitTask(TaskAttemptContext taskContext)
          To promote the task's temporary output to final output location.
abstract  RecordReader<K,V> InputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
          Create a record reader for a given split.
abstract  OutputCommitter OutputFormat.getOutputCommitter(TaskAttemptContext context)
          Get the output committer for this output format.
abstract  RecordWriter<K,V> OutputFormat.getRecordWriter(TaskAttemptContext context)
          Get the RecordWriter for the given task.
abstract  void RecordReader.initialize(InputSplit split, TaskAttemptContext context)
          Called once at initialization.
abstract  boolean OutputCommitter.needsTaskCommit(TaskAttemptContext taskContext)
          Check whether task needs a commit.
 void OutputCommitter.recoverTask(TaskAttemptContext taskContext)
          Recover the task output.
abstract  void OutputCommitter.setupTask(TaskAttemptContext taskContext)
          Sets up output for the task.
 

Uses of TaskAttemptContext in org.apache.hadoop.mapreduce.lib.db
 

Methods in org.apache.hadoop.mapreduce.lib.db with parameters of type TaskAttemptContext
 RecordReader<LongWritable,T> DBInputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
          Create a record reader for a given split.
 OutputCommitter DBOutputFormat.getOutputCommitter(TaskAttemptContext context)
           
 RecordWriter<K,V> DBOutputFormat.getRecordWriter(TaskAttemptContext context)
          Get the RecordWriter for the given task.
 void DBRecordReader.initialize(InputSplit split, TaskAttemptContext context)
           
 

Uses of TaskAttemptContext in org.apache.hadoop.mapreduce.lib.input
 

Fields in org.apache.hadoop.mapreduce.lib.input declared as TaskAttemptContext
protected  TaskAttemptContext CombineFileRecordReader.context
           
 

Methods in org.apache.hadoop.mapreduce.lib.input with parameters of type TaskAttemptContext
 RecordReader<LongWritable,Text> NLineInputFormat.createRecordReader(InputSplit genericSplit, TaskAttemptContext context)
           
 RecordReader<K,V> SequenceFileInputFilter.createRecordReader(InputSplit split, TaskAttemptContext context)
          Create a record reader for the given split
 RecordReader<Text,Text> KeyValueTextInputFormat.createRecordReader(InputSplit genericSplit, TaskAttemptContext context)
           
 RecordReader<K,V> SequenceFileInputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
           
 RecordReader<LongWritable,Text> TextInputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
           
 RecordReader<Text,Text> SequenceFileAsTextInputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
           
abstract  RecordReader<K,V> CombineFileInputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
          This is not implemented yet.
 RecordReader<BytesWritable,BytesWritable> SequenceFileAsBinaryInputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
           
 void KeyValueLineRecordReader.initialize(InputSplit genericSplit, TaskAttemptContext context)
           
 void SequenceFileAsTextRecordReader.initialize(InputSplit split, TaskAttemptContext context)
           
 void CombineFileRecordReader.initialize(InputSplit split, TaskAttemptContext context)
           
 void SequenceFileRecordReader.initialize(InputSplit split, TaskAttemptContext context)
           
 

Constructors in org.apache.hadoop.mapreduce.lib.input with parameters of type TaskAttemptContext
CombineFileRecordReader(CombineFileSplit split, TaskAttemptContext context, Class<? extends RecordReader<K,V>> rrClass)
          A generic RecordReader that can hand out different recordReaders for each chunk in the CombineFileSplit.
 

Uses of TaskAttemptContext in org.apache.hadoop.mapreduce.lib.join
 

Methods in org.apache.hadoop.mapreduce.lib.join with parameters of type TaskAttemptContext
abstract  ComposableRecordReader<K,V> ComposableInputFormat.createRecordReader(InputSplit split, TaskAttemptContext context)
           
 RecordReader<K,TupleWritable> CompositeInputFormat.createRecordReader(InputSplit split, TaskAttemptContext taskContext)
          Construct a CompositeRecordReader for the children of this InputFormat as defined in the init expression.
 void WrappedRecordReader.initialize(InputSplit split, TaskAttemptContext context)
           
 void CompositeRecordReader.initialize(InputSplit split, TaskAttemptContext context)
           
 void MultiFilterRecordReader.initialize(InputSplit split, TaskAttemptContext context)
           
 

Uses of TaskAttemptContext in org.apache.hadoop.mapreduce.lib.output
 

Methods in org.apache.hadoop.mapreduce.lib.output with parameters of type TaskAttemptContext
 void FileOutputCommitter.abortTask(TaskAttemptContext context)
          Delete the work directory
 void FileOutputCommitter.abortTask(TaskAttemptContext context, Path taskAttemptPath)
           
 void FileOutputCommitter.commitTask(TaskAttemptContext context)
          Move the files from the work directory to the job output directory
 void FileOutputCommitter.commitTask(TaskAttemptContext context, Path taskAttemptPath)
           
 Path FileOutputCommitter.getCommittedTaskPath(TaskAttemptContext context)
          Compute the path where the output of a committed task is stored until the entire job is committed.
static Path FileOutputCommitter.getCommittedTaskPath(TaskAttemptContext context, Path out)
           
 Path FileOutputFormat.getDefaultWorkFile(TaskAttemptContext context, String extension)
          Get the default path and filename for the output format.
 OutputCommitter FileOutputFormat.getOutputCommitter(TaskAttemptContext context)
           
 OutputCommitter FilterOutputFormat.getOutputCommitter(TaskAttemptContext context)
           
 OutputCommitter NullOutputFormat.getOutputCommitter(TaskAttemptContext context)
           
 OutputCommitter LazyOutputFormat.getOutputCommitter(TaskAttemptContext context)
           
abstract  RecordWriter<K,V> FileOutputFormat.getRecordWriter(TaskAttemptContext job)
           
 RecordWriter<K,V> FilterOutputFormat.getRecordWriter(TaskAttemptContext context)
           
 RecordWriter<K,V> NullOutputFormat.getRecordWriter(TaskAttemptContext context)
           
 RecordWriter<K,V> SequenceFileOutputFormat.getRecordWriter(TaskAttemptContext context)
           
 RecordWriter<K,V> TextOutputFormat.getRecordWriter(TaskAttemptContext job)
           
 RecordWriter<BytesWritable,BytesWritable> SequenceFileAsBinaryOutputFormat.getRecordWriter(TaskAttemptContext context)
           
 RecordWriter<WritableComparable<?>,Writable> MapFileOutputFormat.getRecordWriter(TaskAttemptContext context)
           
 RecordWriter<K,V> LazyOutputFormat.getRecordWriter(TaskAttemptContext context)
           
protected  org.apache.hadoop.io.SequenceFile.Writer SequenceFileOutputFormat.getSequenceWriter(TaskAttemptContext context, Class<?> keyClass, Class<?> valueClass)
           
 Path FileOutputCommitter.getTaskAttemptPath(TaskAttemptContext context)
          Compute the path where the output of a task attempt is stored until that task is committed.
static Path FileOutputCommitter.getTaskAttemptPath(TaskAttemptContext context, Path out)
          Compute the path where the output of a task attempt is stored until that task is committed.
static String FileOutputFormat.getUniqueFile(TaskAttemptContext context, String name, String extension)
          Generate a unique filename, based on the task id, name, and extension
 boolean FileOutputCommitter.needsTaskCommit(TaskAttemptContext context)
          Did this task write any files in the work directory?
 boolean FileOutputCommitter.needsTaskCommit(TaskAttemptContext context, Path taskAttemptPath)
           
 void FileOutputCommitter.recoverTask(TaskAttemptContext context)
           
 void FileOutputCommitter.setupTask(TaskAttemptContext context)
          No task setup required.
 

Constructors in org.apache.hadoop.mapreduce.lib.output with parameters of type TaskAttemptContext
FileOutputCommitter(Path outputPath, TaskAttemptContext context)
          Create a file output committer
 



Copyright © 2013 Apache Software Foundation. All Rights Reserved.