[ https://issues.apache.org/jira/browse/HADOOP-9324?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Andras Bokor resolved HADOOP-9324. ---------------------------------- Resolution: Duplicate I have raised HADOOP-15021 which is the root cause of most of the issues above. Others are ok. 1. Covered by HADOOP-15021 2. Covered by HADOOP-15021 3. Covered by HADOOP-15021 4. JoinCollector is not deleted 5. No longer an issue 6. Covered by HADOOP-15021 7. Covered by HADOOP-15021 8. Covered by HADOOP-15021 9. Covered by HADOOP-15021 10. JobContextImpl is not deleted. It will covered by HADOOP-15021 11. It is correct as it is 12. Covered by HADOOP-15021 13. Covered by HADOOP-15021 14. Covered by HADOOP-15021 15. Covered by HADOOP-15021 16. Package exists 17. Covered by HADOOP-15021 18. Covered by HADOOP-15021 19. No longer valid 20. Covered by HADOOP-15021 > Out of date API document > ------------------------ > > Key: HADOOP-9324 > URL: https://issues.apache.org/jira/browse/HADOOP-9324 > Project: Hadoop Common > Issue Type: Bug > Components: documentation > Affects Versions: 2.0.3-alpha > Reporter: Hao Zhong > > The documentation is out of date. Some code references are broken: > 1. > http://hadoop.apache.org/docs/current/api/org/apache/hadoop/fs/FSDataInputStream.html > "All Implemented Interfaces: > Closeable, DataInput, *org.apache.hadoop.fs.ByteBufferReadable*, > *org.apache.hadoop.fs.HasFileDescriptor*, PositionedReadable, Seekable " > 2.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/Cluster.html > renewDelegationToken(*org.apache.hadoop.security.token.Token<org.apache.hadoop.mapreduce.security.token.delegation.DelegationTokenIdentifier>* > token) > Deprecated. Use Token.renew(*org.apache.hadoop.conf.Configuration*) > instead > 3.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/JobConf.html > "Use MRAsyncDiskService.moveAndDeleteAllVolumes instead. " > I cannot find the MRAsyncDiskService class in the documentation of 2.0.3. > 4.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/join/CompositeRecordReader.html > "protected > *org.apache.hadoop.mapred.join.CompositeRecordReader.JoinCollector* jc" > Please globally search JoinCollector. It is deleted, but mentioned many times > in the current documentation. > 5.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/OutputCommitter.html > "abortJob(JobContext context, *org.apache.hadoop.mapreduce.JobStatus.State > runState*)" > http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/Job.html > "public *org.apache.hadoop.mapreduce.JobStatus.State* getJobState()" > 4.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/SequenceFileOutputFormat.html > " static *org.apache.hadoop.io.SequenceFile.CompressionType* > getOutputCompressionType" > " static *org.apache.hadoop.io.SequenceFile.Reader[]* getReaders" > 5.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/TaskCompletionEvent.html > "Returns enum Status.SUCESS or Status.FAILURE."->Status.SUCCEEDED? > 6.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/Job.html > " static *org.apache.hadoop.mapreduce.Job.TaskStatusFilter* > getTaskOutputFilter" > " org.apache.hadoop.mapreduce.TaskReport[] getTaskReports(TaskType type) " > 7.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/Reducer.html > "cleanup(*org.apache.hadoop.mapreduce.Reducer.Context* context) " > 8.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/SequenceFileOutputFormat.html > "static *org.apache.hadoop.io.SequenceFile.CompressionType* > getOutputCompressionType(JobConf conf) > Get the *SequenceFile.CompressionType* for the output SequenceFile." > " static *org.apache.hadoop.io.SequenceFile.Reader[]* getReaders" > 9.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/lib/partition/InputSampler.html > "writePartitionFile(Job job, > *org.apache.hadoop.mapreduce.lib.partition.InputSampler.Sampler<K,V>* > sampler) " > 10.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/lib/partition/TotalOrderPartitioner.html > contain JobContextImpl.getNumReduceTasks() - 1 keys. > The JobContextImpl class is already deleted. > 11. > http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/OutputCommitter.html > "Note that this is invoked for jobs with final runstate as > JobStatus.State.FAILED or JobStatus.State.KILLED."->JobStatus.FAILED > JobStatus.KILLED? > 12.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapred/TaskAttemptContext.html > "All Superinterfaces: > JobContext, *org.apache.hadoop.mapreduce.MRJobConfig*, Progressable, > TaskAttemptContext " > 13.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/metrics/file/FileContext.html > "All Implemented Interfaces: > *org.apache.hadoop.metrics.MetricsContext*" > 14.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/metrics/spi/AbstractMetricsContext.html > > "*org.apache.hadoop.metrics.MetricsRecord* createRecord(String recordName)" > 15. > http://hadoop.apache.org/docs/current/api/org/apache/hadoop/net/DNSToSwitchMapping.html > "If a name cannot be resolved to a rack, the implementation should return > NetworkTopology.DEFAULT_RACK." > NetworkTopology is deleted. > 16.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/metrics2/package-summary.html > "myprefix.sink.file.class=org.hadoop.metrics2.sink.FileSink" -> > org.apache.hadoop.metrics2.sink.FileSink? > "org.apache.hadoop.metrics2.impl" -> The package is not found. > 17.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/ha/HAServiceTarget.html > " abstract *org.apache.hadoop.ha.NodeFencer* getFencer() " > 18.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/mapreduce/MarkableIterator.html > "MarkableIterator is a wrapper iterator class that implements the > MarkableIteratorInterface. " > MarkableIteratorInterface is deleted. > 19.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/metrics/spi/NoEmitMetricsContext.html > "A MetricsContext that does not emit data, but, unlike NullContextWithUpdate" > NullContextWithUpdate is deleted. > 20.http://hadoop.apache.org/docs/current/api/org/apache/hadoop/net/ConnectTimeoutException.html > "Thrown by NetUtils.connect(java.net.Socket, java.net.SocketAddress, int) " > The NetUtils class is deleted. > Please revise the documentation. -- This message was sent by Atlassian JIRA (v6.4.14#64029) --------------------------------------------------------------------- To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org