Uses of Package
org.apache.hadoop.mapreduce

Packages that use org.apache.hadoop.mapreduce
org.apache.hadoop.examples Hadoop example code. 
org.apache.hadoop.filecache   
org.apache.hadoop.mapred A software framework for easily writing applications which process vast amounts of data (multi-terabyte data-sets) parallelly on large clusters (thousands of nodes) built of commodity hardware in a reliable, fault-tolerant manner. 
org.apache.hadoop.mapred.lib.db org.apache.hadoop.mapred.lib.db Package 
org.apache.hadoop.mapreduce   
org.apache.hadoop.mapreduce.counters This package contains the implementations of different types of map-reduce counters. 
org.apache.hadoop.mapreduce.lib.db org.apache.hadoop.mapred.lib.db Package 
org.apache.hadoop.mapreduce.lib.fieldsel   
org.apache.hadoop.mapreduce.lib.input   
org.apache.hadoop.mapreduce.lib.jobcontrol Utilities for managing dependent jobs. 
org.apache.hadoop.mapreduce.lib.map   
org.apache.hadoop.mapreduce.lib.output   
org.apache.hadoop.mapreduce.lib.partition   
org.apache.hadoop.mapreduce.lib.reduce   
org.apache.hadoop.mapreduce.security.token   
org.apache.hadoop.mapreduce.server.jobtracker   
org.apache.hadoop.mapreduce.server.tasktracker   
org.apache.hadoop.mapreduce.server.tasktracker.userlogs   
org.apache.hadoop.mapreduce.split   
org.apache.hadoop.mapreduce.task   
org.apache.hadoop.mapreduce.util   
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.examples
InputFormat
          InputFormat describes the input-specification for a Map-Reduce job.
InputSplit
          InputSplit represents the data to be processed by an individual Mapper.
Mapper
          Maps input key/value pairs to a set of intermediate key/value pairs.
Mapper.Context
          The Context passed on to the Mapper implementations.
Partitioner
          Partitions the key space.
RecordReader
          The record reader breaks the data into key/value pairs for input to the Mapper.
Reducer
          Reduces a set of intermediate values which share a key to a smaller set of values.
Reducer.Context
          The Context passed on to the Reducer implementations.
TaskAttemptContext
          The context for task attempts.
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.filecache
JobID
          JobID represents the immutable and unique identifier for the job.
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.mapred
Cluster.JobTrackerStatus
           
ClusterMetrics
          Status information on the current state of the Map-Reduce cluster.
Counter
          A named counter that tracks the progress of a map/reduce job.
Counters
          Counters holds per job/task counters, defined either by the Map-Reduce framework or applications.
ID
          A general identifier, which internally stores the id as an integer.
InputSplit
          InputSplit represents the data to be processed by an individual Mapper.
JobACL
          Job related ACLs
JobContext
          A read-only view of the job that is provided to the tasks while they are running.
JobID
          JobID represents the immutable and unique identifier for the job.
JobStatus.State
          Current state of the job
OutputCommitter
          OutputCommitter describes the commit of task output for a Map-Reduce job.
OutputFormat
          OutputFormat describes the output-specification for a Map-Reduce job.
RecordWriter
          RecordWriter writes the output <key, value> pairs to an output file.
Reducer
          Reduces a set of intermediate values which share a key to a smaller set of values.
Reducer.Context
          The Context passed on to the Reducer implementations.
StatusReporter
           
TaskAttemptContext
          The context for task attempts.
TaskAttemptID
          TaskAttemptID represents the immutable and unique identifier for a task attempt.
TaskID
          TaskID represents the immutable and unique identifier for a Map or Reduce Task.
TaskType
          Enum for map, reduce, job-setup, job-cleanup, task-cleanup task types.
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.mapred.lib.db
RecordReader
          The record reader breaks the data into key/value pairs for input to the Mapper.
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.mapreduce
Cluster.JobTrackerStatus
           
Counter
          A named counter that tracks the progress of a map/reduce job.
Counters
          Counters holds per job/task counters, defined either by the Map-Reduce framework or applications.
FileSystemCounter
           
ID
          A general identifier, which internally stores the id as an integer.
InputFormat
          InputFormat describes the input-specification for a Map-Reduce job.
InputSplit
          InputSplit represents the data to be processed by an individual Mapper.
Job
          The job submitter's view of the Job.
Job.JobState
           
JobACL
          Job related ACLs
JobContext
          A read-only view of the job that is provided to the tasks while they are running.
JobCounter
           
JobID
          JobID represents the immutable and unique identifier for the job.
JobStatus.State
          Current state of the job
MapContext
          The context that is given to the Mapper.
Mapper
          Maps input key/value pairs to a set of intermediate key/value pairs.
Mapper.Context
          The Context passed on to the Mapper implementations.
OutputCommitter
          OutputCommitter describes the commit of task output for a Map-Reduce job.
OutputFormat
          OutputFormat describes the output-specification for a Map-Reduce job.
Partitioner
          Partitions the key space.
RecordReader
          The record reader breaks the data into key/value pairs for input to the Mapper.
RecordWriter
          RecordWriter writes the output <key, value> pairs to an output file.
ReduceContext
          The context passed to the Reducer.
Reducer
          Reduces a set of intermediate values which share a key to a smaller set of values.
Reducer.Context
          The Context passed on to the Reducer implementations.
TaskAttemptContext
          The context for task attempts.
TaskAttemptID
          TaskAttemptID represents the immutable and unique identifier for a task attempt.
TaskCounter
           
TaskID
          TaskID represents the immutable and unique identifier for a Map or Reduce Task.
TaskInputOutputContext
          A context object that allows input and output from the task.
TaskType
          Enum for map, reduce, job-setup, job-cleanup, task-cleanup task types.
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.mapreduce.counters
Counter
          A named counter that tracks the progress of a map/reduce job.
FileSystemCounter
           
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.mapreduce.lib.db
InputFormat
          InputFormat describes the input-specification for a Map-Reduce job.
InputSplit
          InputSplit represents the data to be processed by an individual Mapper.
Job
          The job submitter's view of the Job.
JobContext
          A read-only view of the job that is provided to the tasks while they are running.
OutputCommitter
          OutputCommitter describes the commit of task output for a Map-Reduce job.
OutputFormat
          OutputFormat describes the output-specification for a Map-Reduce job.
RecordReader
          The record reader breaks the data into key/value pairs for input to the Mapper.
RecordWriter
          RecordWriter writes the output <key, value> pairs to an output file.
TaskAttemptContext
          The context for task attempts.
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.mapreduce.lib.fieldsel
Mapper
          Maps input key/value pairs to a set of intermediate key/value pairs.
Mapper.Context
          The Context passed on to the Mapper implementations.
Reducer
          Reduces a set of intermediate values which share a key to a smaller set of values.
Reducer.Context
          The Context passed on to the Reducer implementations.
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.mapreduce.lib.input
InputFormat
          InputFormat describes the input-specification for a Map-Reduce job.
InputSplit
          InputSplit represents the data to be processed by an individual Mapper.
Job
          The job submitter's view of the Job.
JobContext
          A read-only view of the job that is provided to the tasks while they are running.
Mapper
          Maps input key/value pairs to a set of intermediate key/value pairs.
Mapper.Context
          The Context passed on to the Mapper implementations.
RecordReader
          The record reader breaks the data into key/value pairs for input to the Mapper.
TaskAttemptContext
          The context for task attempts.
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.mapreduce.lib.jobcontrol
Job
          The job submitter's view of the Job.
JobID
          JobID represents the immutable and unique identifier for the job.
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.mapreduce.lib.map
Counter
          A named counter that tracks the progress of a map/reduce job.
InputFormat
          InputFormat describes the input-specification for a Map-Reduce job.
InputSplit
          InputSplit represents the data to be processed by an individual Mapper.
Job
          The job submitter's view of the Job.
JobContext
          A read-only view of the job that is provided to the tasks while they are running.
JobID
          JobID represents the immutable and unique identifier for the job.
MapContext
          The context that is given to the Mapper.
Mapper
          Maps input key/value pairs to a set of intermediate key/value pairs.
Mapper.Context
          The Context passed on to the Mapper implementations.
OutputCommitter
          OutputCommitter describes the commit of task output for a Map-Reduce job.
OutputFormat
          OutputFormat describes the output-specification for a Map-Reduce job.
Partitioner
          Partitions the key space.
Reducer
          Reduces a set of intermediate values which share a key to a smaller set of values.
TaskAttemptContext
          The context for task attempts.
TaskAttemptID
          TaskAttemptID represents the immutable and unique identifier for a task attempt.
TaskInputOutputContext
          A context object that allows input and output from the task.
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.mapreduce.lib.output
Job
          The job submitter's view of the Job.
JobContext
          A read-only view of the job that is provided to the tasks while they are running.
JobStatus.State
          Current state of the job
OutputCommitter
          OutputCommitter describes the commit of task output for a Map-Reduce job.
OutputFormat
          OutputFormat describes the output-specification for a Map-Reduce job.
Partitioner
          Partitions the key space.
RecordWriter
          RecordWriter writes the output <key, value> pairs to an output file.
TaskAttemptContext
          The context for task attempts.
TaskInputOutputContext
          A context object that allows input and output from the task.
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.mapreduce.lib.partition
InputFormat
          InputFormat describes the input-specification for a Map-Reduce job.
Job
          The job submitter's view of the Job.
JobContext
          A read-only view of the job that is provided to the tasks while they are running.
Partitioner
          Partitions the key space.
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.mapreduce.lib.reduce
Counter
          A named counter that tracks the progress of a map/reduce job.
InputFormat
          InputFormat describes the input-specification for a Map-Reduce job.
JobContext
          A read-only view of the job that is provided to the tasks while they are running.
JobID
          JobID represents the immutable and unique identifier for the job.
Mapper
          Maps input key/value pairs to a set of intermediate key/value pairs.
OutputCommitter
          OutputCommitter describes the commit of task output for a Map-Reduce job.
OutputFormat
          OutputFormat describes the output-specification for a Map-Reduce job.
Partitioner
          Partitions the key space.
ReduceContext
          The context passed to the Reducer.
Reducer
          Reduces a set of intermediate values which share a key to a smaller set of values.
Reducer.Context
          The Context passed on to the Reducer implementations.
TaskAttemptContext
          The context for task attempts.
TaskAttemptID
          TaskAttemptID represents the immutable and unique identifier for a task attempt.
TaskInputOutputContext
          A context object that allows input and output from the task.
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.mapreduce.security.token
JobID
          JobID represents the immutable and unique identifier for the job.
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.mapreduce.server.jobtracker
TaskType
          Enum for map, reduce, job-setup, job-cleanup, task-cleanup task types.
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.mapreduce.server.tasktracker
JobID
          JobID represents the immutable and unique identifier for the job.
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.mapreduce.server.tasktracker.userlogs
JobID
          JobID represents the immutable and unique identifier for the job.
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.mapreduce.split
InputSplit
          InputSplit represents the data to be processed by an individual Mapper.
JobID
          JobID represents the immutable and unique identifier for the job.
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.mapreduce.task
Counter
          A named counter that tracks the progress of a map/reduce job.
InputFormat
          InputFormat describes the input-specification for a Map-Reduce job.
InputSplit
          InputSplit represents the data to be processed by an individual Mapper.
JobContext
          A read-only view of the job that is provided to the tasks while they are running.
JobID
          JobID represents the immutable and unique identifier for the job.
MapContext
          The context that is given to the Mapper.
Mapper
          Maps input key/value pairs to a set of intermediate key/value pairs.
OutputCommitter
          OutputCommitter describes the commit of task output for a Map-Reduce job.
OutputFormat
          OutputFormat describes the output-specification for a Map-Reduce job.
Partitioner
          Partitions the key space.
RecordReader
          The record reader breaks the data into key/value pairs for input to the Mapper.
RecordWriter
          RecordWriter writes the output <key, value> pairs to an output file.
ReduceContext
          The context passed to the Reducer.
ReduceContext.ValueIterator
          Iterator to iterate over values for a given group of records.
Reducer
          Reduces a set of intermediate values which share a key to a smaller set of values.
StatusReporter
           
TaskAttemptContext
          The context for task attempts.
TaskAttemptID
          TaskAttemptID represents the immutable and unique identifier for a task attempt.
TaskInputOutputContext
          A context object that allows input and output from the task.
 

Classes in org.apache.hadoop.mapreduce used by org.apache.hadoop.mapreduce.util
Counter
          A named counter that tracks the progress of a map/reduce job.
 



Copyright © 2009 The Apache Software Foundation