Deprecated API


Contents
Deprecated Interfaces
org.apache.hadoop.mapred.TaskAttemptContext
          Use TaskAttemptContext instead. 
 

Deprecated Classes
org.apache.hadoop.mapred.Counters
          Use Counters instead. 
org.apache.hadoop.mapred.jobcontrol.Job
          Use ControlledJob instead. 
org.apache.hadoop.mapred.JobContextImpl
          Use JobContext instead. 
org.apache.hadoop.mapred.jobcontrol.JobControl
          Use JobControl instead 
org.apache.hadoop.mapred.LineRecordReader.LineReader
          Use LineReader instead. 
org.apache.hadoop.mapred.MultiFileInputFormat
          Use CombineFileInputFormat instead 
org.apache.hadoop.mapred.MultiFileSplit
          Use CombineFileSplit instead 
org.apache.hadoop.mapred.OutputLogFilter
          Use Utils.OutputFileUtils.OutputLogFilter instead. 
org.apache.hadoop.mapred.TaskAttemptContextImpl
          Use TaskAttemptContextImpl instead. 
org.apache.hadoop.streaming.UTF8ByteArrayUtils
          use UTF8ByteArrayUtils and StreamKeyValUtil instead 
 

Deprecated Fields
org.apache.hadoop.mapred.JobConf.MAPRED_TASK_DEFAULT_MAXVMEM_PROPERTY
            
org.apache.hadoop.mapred.JobConf.MAPRED_TASK_ENV
          Use JobConf.MAPRED_MAP_TASK_ENV or JobConf.MAPRED_REDUCE_TASK_ENV 
org.apache.hadoop.mapred.JobConf.MAPRED_TASK_JAVA_OPTS
          Use JobConf.MAPRED_MAP_TASK_JAVA_OPTS or JobConf.MAPRED_REDUCE_TASK_JAVA_OPTS 
org.apache.hadoop.mapred.JobConf.MAPRED_TASK_MAXPMEM_PROPERTY
            
org.apache.hadoop.mapred.JobConf.MAPRED_TASK_MAXVMEM_PROPERTY
          Use JobConf.MAPRED_JOB_MAP_MEMORY_MB_PROPERTY and JobConf.MAPRED_JOB_REDUCE_MEMORY_MB_PROPERTY 
org.apache.hadoop.mapred.JobConf.MAPRED_TASK_ULIMIT
          Use JobConf.MAPRED_MAP_TASK_ULIMIT or JobConf.MAPRED_REDUCE_TASK_ULIMIT 
org.apache.hadoop.mapred.JobConf.UPPER_LIMIT_ON_TASK_VMEM_PROPERTY
            
 

Deprecated Methods
org.apache.hadoop.mapred.TaskLog.captureOutAndError(List, List, File, File, long, boolean, String)
          pidFiles are no more used. Instead pid is exported to env variable JVM_PID. 
org.apache.hadoop.mapred.TaskLog.captureOutAndError(List, List, File, File, long, String)
          pidFiles are no more used. Instead pid is exported to env variable JVM_PID. 
org.apache.hadoop.mapred.FileOutputCommitter.cleanupJob(JobContext)
           
org.apache.hadoop.mapred.OutputCommitter.cleanupJob(JobContext)
          use OutputCommitter.commitJob(JobContext) or OutputCommitter.abortJob(JobContext, int) instead 
org.apache.hadoop.mapred.OutputCommitter.cleanupJob(JobContext)
           
org.apache.hadoop.mapreduce.OutputCommitter.cleanupJob(JobContext)
          use OutputCommitter.commitJob(JobContext) or OutputCommitter.abortJob(JobContext, JobStatus.State) instead 
org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter.cleanupJob(JobContext)
           
org.apache.hadoop.mapred.TaskTracker.cleanupStorage()
           
org.apache.hadoop.mapred.Counters.Counter.contentEquals(Counters.Counter)
            
org.apache.hadoop.filecache.DistributedCache.createAllSymlink(Configuration, File, File)
          Internal to MapReduce framework. Use DistributedCacheManager instead. 
org.apache.hadoop.mapreduce.lib.db.DBRecordReader.createValue()
            
org.apache.hadoop.mapred.JobConf.deleteLocalFiles()
           
org.apache.hadoop.streaming.UTF8ByteArrayUtils.findByte(byte[], int, int, byte)
          use UTF8ByteArrayUtils.findByte(byte[], int, int, byte) 
org.apache.hadoop.streaming.UTF8ByteArrayUtils.findBytes(byte[], int, int, byte[])
          use UTF8ByteArrayUtils.findBytes(byte[], int, int, byte[]) 
org.apache.hadoop.mapred.Counters.findCounter(String, int, String)
          use Counters.findCounter(String, String) instead 
org.apache.hadoop.streaming.UTF8ByteArrayUtils.findNthByte(byte[], byte, int)
          use UTF8ByteArrayUtils.findNthByte(byte[], byte, int) 
org.apache.hadoop.streaming.UTF8ByteArrayUtils.findNthByte(byte[], int, int, byte, int)
          use UTF8ByteArrayUtils.findNthByte(byte[], int, int, byte, int) 
org.apache.hadoop.streaming.UTF8ByteArrayUtils.findTab(byte[])
          use StreamKeyValUtil.findTab(byte[]) 
org.apache.hadoop.streaming.UTF8ByteArrayUtils.findTab(byte[], int, int)
          use StreamKeyValUtil.findTab(byte[], int, int) 
org.apache.hadoop.streaming.StreamJob.getClusterNick()
           
org.apache.hadoop.mapred.JobTracker.getClusterStatus()
          use JobTracker.getClusterStatus(boolean) 
org.apache.hadoop.mapred.Counters.Group.getCounter(int, String)
          use Counters.Group.findCounter(String) instead 
org.apache.hadoop.mapred.JobClient.getJob(String)
          Applications should rather use JobClient.getJob(JobID). 
org.apache.hadoop.mapred.JobProfile.getJobId()
          use getJobID() instead 
org.apache.hadoop.mapred.JobStatus.getJobId()
          use getJobID instead 
org.apache.hadoop.mapred.RunningJob.getJobID()
          This method is deprecated and will be removed. Applications should rather use RunningJob.getID(). 
org.apache.hadoop.mapred.JobID.getJobIDsPattern(String, Integer)
           
org.apache.hadoop.mapred.JobClient.getMapTaskReports(String)
          Applications should rather use JobClient.getMapTaskReports(JobID) 
org.apache.hadoop.mapred.ClusterStatus.getMaxMemory()
           
org.apache.hadoop.mapred.JobConf.getMaxPhysicalMemoryForTask()
          this variable is deprecated and nolonger in use. 
org.apache.hadoop.mapred.JobConf.getMaxVirtualMemoryForTask()
          Use JobConf.getMemoryForMapTask() and JobConf.getMemoryForReduceTask() 
org.apache.hadoop.mapreduce.lib.db.DBRecordReader.getPos()
            
org.apache.hadoop.mapred.JobClient.getReduceTaskReports(String)
          Applications should rather use JobClient.getReduceTaskReports(JobID) 
org.apache.hadoop.mapred.NetUtils2.getServerAddress(Configuration, String, String, String)
           
org.apache.hadoop.mapred.TaskAttemptID.getTaskAttemptIDsPattern(String, Integer, Boolean, Integer, Integer)
           
org.apache.hadoop.mapred.TaskCompletionEvent.getTaskId()
          use TaskCompletionEvent.getTaskAttemptId() instead. 
org.apache.hadoop.mapred.TaskReport.getTaskId()
          use TaskReport.getTaskID() instead 
org.apache.hadoop.mapred.TaskID.getTaskIDsPattern(String, Integer, Boolean, Integer)
           
org.apache.hadoop.mapred.JobClient.getTaskOutputFilter()
           
org.apache.hadoop.mapred.ClusterStatus.getUsedMemory()
           
org.apache.hadoop.streaming.StreamJob.go()
          use StreamJob.run(String[]) instead. 
org.apache.hadoop.mapred.RunningJob.killTask(String, boolean)
          Applications should rather use RunningJob.killTask(TaskAttemptID, boolean) 
org.apache.hadoop.mapred.JobHistory.MapAttempt.logFailed(TaskAttemptID, long, String, String)
          Use JobHistory.MapAttempt.logFailed(TaskAttemptID, long, String, String, String) 
org.apache.hadoop.mapred.JobHistory.ReduceAttempt.logFailed(TaskAttemptID, long, String, String)
          Use JobHistory.ReduceAttempt.logFailed(TaskAttemptID, long, String, String, String) 
org.apache.hadoop.mapred.JobHistory.ReduceAttempt.logFinished(TaskAttemptID, long, long, long, String)
          Use JobHistory.ReduceAttempt.logFinished(TaskAttemptID, long, long, long, String, String, String, Counters) 
org.apache.hadoop.mapred.JobHistory.MapAttempt.logFinished(TaskAttemptID, long, String)
          Use JobHistory.MapAttempt.logFinished(TaskAttemptID, long, String, String, String, Counters) 
org.apache.hadoop.mapred.JobHistory.JobInfo.logJobInfo(JobID, long, long, int)
          Use JobHistory.JobInfo.logJobInfo(JobID, long, long) instead. 
org.apache.hadoop.mapred.JobHistory.MapAttempt.logKilled(TaskAttemptID, long, String, String)
          Use JobHistory.MapAttempt.logKilled(TaskAttemptID, long, String, String, String) 
org.apache.hadoop.mapred.JobHistory.ReduceAttempt.logKilled(TaskAttemptID, long, String, String)
          Use JobHistory.ReduceAttempt.logKilled(TaskAttemptID, long, String, String, String) 
org.apache.hadoop.mapred.JobHistory.JobInfo.logStarted(JobID, long, int, int)
          Use JobHistory.JobInfo.logInited(JobID, long, int, int) and JobHistory.JobInfo.logStarted(JobID) 
org.apache.hadoop.mapred.JobHistory.MapAttempt.logStarted(TaskAttemptID, long, String)
          Use JobHistory.MapAttempt.logStarted(TaskAttemptID, long, String, int, String) 
org.apache.hadoop.mapred.JobHistory.ReduceAttempt.logStarted(TaskAttemptID, long, String)
          Use JobHistory.ReduceAttempt.logStarted(TaskAttemptID, long, String, int, String) 
org.apache.hadoop.mapred.JobHistory.JobInfo.logSubmitted(JobID, JobConf, String, long)
          Use JobHistory.JobInfo.logSubmitted(JobID, JobConf, String, long, boolean) instead. 
org.apache.hadoop.mapreduce.lib.db.DBRecordReader.next(LongWritable, T)
          Use DBRecordReader.nextKeyValue() 
org.apache.hadoop.mapred.JobID.read(DataInput)
           
org.apache.hadoop.mapred.TaskID.read(DataInput)
           
org.apache.hadoop.mapred.TaskAttemptID.read(DataInput)
           
org.apache.hadoop.streaming.UTF8ByteArrayUtils.readLine(LineReader, Text)
          use StreamKeyValUtil.readLine(LineReader, Text) 
org.apache.hadoop.mapred.jobcontrol.Job.setAssignedJobID(JobID)
          setAssignedJobID should not be called. JOBID is set by the framework. 
org.apache.hadoop.mapreduce.Counter.setDisplayName(String)
          (and no-op by default) 
org.apache.hadoop.mapreduce.counters.AbstractCounter.setDisplayName(String)
           
org.apache.hadoop.mapreduce.counters.GenericCounter.setDisplayName(String)
           
org.apache.hadoop.mapred.JobConf.setMaxPhysicalMemoryForTask(long)
           
org.apache.hadoop.mapred.JobConf.setMaxVirtualMemoryForTask(long)
          Use JobConf.setMemoryForMapTask(long mem) and Use JobConf.setMemoryForReduceTask(long mem) 
org.apache.hadoop.util.ProcfsBasedProcessTree.setSigKillInterval(long)
          Use ProcfsBasedProcessTree.ProcfsBasedProcessTree( String, boolean, long) instead 
org.apache.hadoop.mapred.TaskCompletionEvent.setTaskId(String)
          use TaskCompletionEvent.setTaskID(TaskAttemptID) instead. 
org.apache.hadoop.mapred.JobClient.setTaskOutputFilter(JobClient.TaskStatusFilter)
           
org.apache.hadoop.mapred.Counters.size()
          use AbstractCounters.countCounters() instead 
org.apache.hadoop.streaming.UTF8ByteArrayUtils.splitKeyVal(byte[], int, int, Text, Text, int)
          use StreamKeyValUtil.splitKeyVal(byte[], int, int, Text, Text, int) 
org.apache.hadoop.streaming.UTF8ByteArrayUtils.splitKeyVal(byte[], int, int, Text, Text, int, int)
          use StreamKeyValUtil.splitKeyVal(byte[], int, int, Text, Text, int, int) 
org.apache.hadoop.streaming.UTF8ByteArrayUtils.splitKeyVal(byte[], Text, Text, int)
          use StreamKeyValUtil.splitKeyVal(byte[], Text, Text, int) 
org.apache.hadoop.streaming.UTF8ByteArrayUtils.splitKeyVal(byte[], Text, Text, int, int)
          use StreamKeyValUtil.splitKeyVal(byte[], Text, Text, int, int) 
org.apache.hadoop.mapred.pipes.Submitter.submitJob(JobConf)
          Use Submitter.runJob(JobConf) 
 

Deprecated Constructors
org.apache.hadoop.mapred.FileSplit(Path, long, long, JobConf)
            
org.apache.hadoop.mapred.JobProfile(String, String, String, String, String)
          use JobProfile(String, JobID, String, String, String) instead 
org.apache.hadoop.streaming.StreamJob(String[], boolean)
          use StreamJob() with ToolRunner or set the Configuration using StreamJob.setConf(Configuration) and run with StreamJob.run(String[]). 
 



Copyright © 2009 The Apache Software Foundation