MAPREDUCE-5657. Fix Javadoc errors caused by incorrect or illegal tags in doc comments. Contributed by Akira AJISAKA.
Project: http://git-wip-us.apache.org/repos/asf/hadoop/repo Commit: http://git-wip-us.apache.org/repos/asf/hadoop/commit/9ae7f9eb Tree: http://git-wip-us.apache.org/repos/asf/hadoop/tree/9ae7f9eb Diff: http://git-wip-us.apache.org/repos/asf/hadoop/diff/9ae7f9eb Branch: refs/heads/YARN-2928 Commit: 9ae7f9eb7baeb244e1b95aabc93ad8124870b9a9 Parents: 742f9d9 Author: Tsuyoshi Ozawa <oz...@apache.org> Authored: Tue Mar 3 18:06:26 2015 +0900 Committer: Tsuyoshi Ozawa <oz...@apache.org> Committed: Tue Mar 3 18:06:26 2015 +0900 ---------------------------------------------------------------------- hadoop-mapreduce-project/CHANGES.txt | 3 ++ .../hadoop/mapred/TaskAttemptListenerImpl.java | 4 +- .../hadoop/mapreduce/v2/app/JobEndNotifier.java | 1 - .../apache/hadoop/mapreduce/v2/util/MRApps.java | 2 +- .../hadoop/filecache/DistributedCache.java | 2 +- .../org/apache/hadoop/mapred/ClusterStatus.java | 4 +- .../apache/hadoop/mapred/FileOutputFormat.java | 2 +- .../java/org/apache/hadoop/mapred/IFile.java | 2 +- .../apache/hadoop/mapred/JobACLsManager.java | 1 - .../org/apache/hadoop/mapred/JobClient.java | 8 ++-- .../java/org/apache/hadoop/mapred/JobConf.java | 49 +++++++++----------- .../java/org/apache/hadoop/mapred/Mapper.java | 2 +- .../org/apache/hadoop/mapred/QueueManager.java | 30 ++++++------ .../org/apache/hadoop/mapred/RecordReader.java | 2 +- .../java/org/apache/hadoop/mapred/Reducer.java | 14 +++--- .../hadoop/mapred/TaskUmbilicalProtocol.java | 1 - .../apache/hadoop/mapred/lib/ChainMapper.java | 40 ++++++++-------- .../apache/hadoop/mapred/lib/ChainReducer.java | 44 +++++++++--------- .../hadoop/mapred/lib/MultipleOutputs.java | 29 +++++------- .../hadoop/mapred/lib/TokenCountMapper.java | 2 +- .../lib/aggregate/ValueAggregatorJob.java | 2 +- .../lib/aggregate/ValueAggregatorReducer.java | 3 +- .../hadoop/mapred/lib/db/DBInputFormat.java | 4 +- .../org/apache/hadoop/mapreduce/Cluster.java | 1 + .../apache/hadoop/mapreduce/ClusterMetrics.java | 6 +-- .../apache/hadoop/mapreduce/CryptoUtils.java | 10 ++-- .../java/org/apache/hadoop/mapreduce/Job.java | 2 +- .../org/apache/hadoop/mapreduce/JobContext.java | 2 - .../hadoop/mapreduce/JobSubmissionFiles.java | 2 +- .../org/apache/hadoop/mapreduce/Mapper.java | 9 ++-- .../org/apache/hadoop/mapreduce/Reducer.java | 12 ++--- .../mapreduce/filecache/DistributedCache.java | 5 +- .../lib/aggregate/ValueAggregatorJob.java | 2 +- .../hadoop/mapreduce/lib/chain/Chain.java | 4 +- .../hadoop/mapreduce/lib/chain/ChainMapper.java | 10 ++-- .../mapreduce/lib/chain/ChainReducer.java | 14 +++--- .../hadoop/mapreduce/lib/db/DBInputFormat.java | 2 +- .../hadoop/mapreduce/lib/db/DBWritable.java | 2 +- .../mapreduce/lib/join/TupleWritable.java | 2 +- .../mapreduce/lib/map/MultithreadedMapper.java | 6 +-- .../mapreduce/lib/output/FileOutputFormat.java | 2 +- .../mapreduce/lib/output/MultipleOutputs.java | 11 ++--- .../lib/partition/BinaryPartitioner.java | 2 +- .../hadoop/mapreduce/task/JobContextImpl.java | 2 - .../hadoop/mapreduce/RandomTextWriter.java | 4 +- .../apache/hadoop/mapreduce/RandomWriter.java | 5 +- .../hadoop/examples/MultiFileWordCount.java | 2 +- .../apache/hadoop/examples/QuasiMonteCarlo.java | 4 +- .../hadoop/examples/RandomTextWriter.java | 4 +- .../apache/hadoop/examples/RandomWriter.java | 5 +- .../apache/hadoop/examples/SecondarySort.java | 2 +- .../org/apache/hadoop/examples/pi/DistBbp.java | 2 +- .../apache/hadoop/examples/pi/math/Modular.java | 2 +- .../hadoop/examples/terasort/GenSort.java | 2 +- .../org/apache/hadoop/tools/CopyListing.java | 14 +++--- .../java/org/apache/hadoop/tools/DistCp.java | 4 +- .../apache/hadoop/tools/DistCpOptionSwitch.java | 2 +- .../org/apache/hadoop/tools/OptionsParser.java | 2 +- .../hadoop/tools/mapred/CopyCommitter.java | 4 +- .../apache/hadoop/tools/mapred/CopyMapper.java | 5 +- .../hadoop/tools/mapred/CopyOutputFormat.java | 4 +- .../tools/mapred/RetriableFileCopyCommand.java | 3 +- .../tools/mapred/UniformSizeInputFormat.java | 4 +- .../tools/mapred/lib/DynamicInputFormat.java | 4 +- .../tools/mapred/lib/DynamicRecordReader.java | 12 ++--- .../apache/hadoop/tools/util/DistCpUtils.java | 2 +- .../hadoop/tools/util/RetriableCommand.java | 2 +- .../hadoop/tools/util/ThrottledInputStream.java | 8 ++-- .../java/org/apache/hadoop/tools/Logalyzer.java | 4 +- .../ResourceUsageEmulatorPlugin.java | 2 +- .../fs/swift/http/RestClientBindings.java | 6 +-- .../hadoop/fs/swift/http/SwiftRestClient.java | 6 +-- .../fs/swift/snative/SwiftNativeFileSystem.java | 6 +-- .../snative/SwiftNativeFileSystemStore.java | 6 +-- .../hadoop/fs/swift/util/SwiftTestUtils.java | 2 +- .../apache/hadoop/tools/rumen/InputDemuxer.java | 4 +- .../util/MapReduceJobPropertiesParser.java | 5 +- .../apache/hadoop/tools/rumen/package-info.java | 8 ++-- 78 files changed, 249 insertions(+), 261 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/CHANGES.txt ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/CHANGES.txt b/hadoop-mapreduce-project/CHANGES.txt index 5fd7d30..5524b14 100644 --- a/hadoop-mapreduce-project/CHANGES.txt +++ b/hadoop-mapreduce-project/CHANGES.txt @@ -181,6 +181,9 @@ Trunk (Unreleased) MAPREDUCE-6234. TestHighRamJob fails due to the change in MAPREDUCE-5785. (Masatake Iwasaki via kasha) + MAPREDUCE-5657. Fix Javadoc errors caused by incorrect or illegal tags in doc + comments. (Akira AJISAKA via ozawa) + BREAKDOWN OF MAPREDUCE-2841 (NATIVE TASK) SUBTASKS MAPREDUCE-5985. native-task: Fix build on macosx. Contributed by http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapred/TaskAttemptListenerImpl.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapred/TaskAttemptListenerImpl.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapred/TaskAttemptListenerImpl.java index 5f39edd..c8f2427 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapred/TaskAttemptListenerImpl.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapred/TaskAttemptListenerImpl.java @@ -174,7 +174,7 @@ public class TaskAttemptListenerImpl extends CompositeService /** * Child checking whether it can commit. * - * <br/> + * <br> * Commit is a two-phased protocol. First the attempt informs the * ApplicationMaster that it is * {@link #commitPending(TaskAttemptID, TaskStatus)}. Then it repeatedly polls @@ -208,7 +208,7 @@ public class TaskAttemptListenerImpl extends CompositeService * TaskAttempt is reporting that it is in commit_pending and it is waiting for * the commit Response * - * <br/> + * <br> * Commit it a two-phased protocol. First the attempt informs the * ApplicationMaster that it is * {@link #commitPending(TaskAttemptID, TaskStatus)}. Then it repeatedly polls http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/JobEndNotifier.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/JobEndNotifier.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/JobEndNotifier.java index 981e6ff..05bb40b 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/JobEndNotifier.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/JobEndNotifier.java @@ -44,7 +44,6 @@ import org.mortbay.log.Log; * proxy if needed</li><li> * The URL may contain sentinels which will be replaced by jobId and jobStatus * (eg. SUCCEEDED/KILLED/FAILED) </li> </ul> - * </p> */ public class JobEndNotifier implements Configurable { private static final String JOB_ID = "$jobId"; http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/java/org/apache/hadoop/mapreduce/v2/util/MRApps.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/java/org/apache/hadoop/mapreduce/v2/util/MRApps.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/java/org/apache/hadoop/mapreduce/v2/util/MRApps.java index 1520fc8..e4b43b5 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/java/org/apache/hadoop/mapreduce/v2/util/MRApps.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-common/src/main/java/org/apache/hadoop/mapreduce/v2/util/MRApps.java @@ -345,7 +345,7 @@ public class MRApps extends Apps { * {@link MRJobConfig#MAPREDUCE_JOB_CLASSLOADER} is set to true, and * the APP_CLASSPATH environment variable is set. * @param conf - * @returns the created job classloader, or null if the job classloader is not + * @return the created job classloader, or null if the job classloader is not * enabled or the APP_CLASSPATH environment variable is not set * @throws IOException */ http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/filecache/DistributedCache.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/filecache/DistributedCache.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/filecache/DistributedCache.java index 370d67d..0783eb5 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/filecache/DistributedCache.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/filecache/DistributedCache.java @@ -113,7 +113,7 @@ import org.apache.hadoop.mapreduce.MRJobConfig; * } * } * - * </pre></blockquote></p> + * </pre></blockquote> * * It is also very common to use the DistributedCache by using * {@link org.apache.hadoop.util.GenericOptionsParser}. http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/ClusterStatus.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/ClusterStatus.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/ClusterStatus.java index 8b56787..904897b 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/ClusterStatus.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/ClusterStatus.java @@ -48,7 +48,7 @@ import org.apache.hadoop.util.StringInterner; * Task capacity of the cluster. * </li> * <li> - * The number of currently running map & reduce tasks. + * The number of currently running map and reduce tasks. * </li> * <li> * State of the <code>JobTracker</code>. @@ -56,7 +56,7 @@ import org.apache.hadoop.util.StringInterner; * <li> * Details regarding black listed trackers. * </li> - * </ol></p> + * </ol> * * <p>Clients can query for the latest <code>ClusterStatus</code>, via * {@link JobClient#getClusterStatus()}.</p> http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/FileOutputFormat.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/FileOutputFormat.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/FileOutputFormat.java index 721c8a8..821c1e8 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/FileOutputFormat.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/FileOutputFormat.java @@ -179,7 +179,7 @@ public abstract class FileOutputFormat<K, V> implements OutputFormat<K, V> { * Get the {@link Path} to the task's temporary output directory * for the map-reduce job * - * <h4 id="SideEffectFiles">Tasks' Side-Effect Files</h4> + * <b id="SideEffectFiles">Tasks' Side-Effect Files</b> * * <p><i>Note:</i> The following is valid only if the {@link OutputCommitter} * is {@link FileOutputCommitter}. If <code>OutputCommitter</code> is not http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/IFile.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/IFile.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/IFile.java index 30ebd6b..32e07e7 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/IFile.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/IFile.java @@ -47,7 +47,7 @@ import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; /** - * <code>IFile</code> is the simple <key-len, value-len, key, value> format + * <code>IFile</code> is the simple <key-len, value-len, key, value> format * for the intermediate map-outputs in Map-Reduce. * * There is a <code>Writer</code> to write out map-outputs in this format and http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/JobACLsManager.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/JobACLsManager.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/JobACLsManager.java index 37633ab..0dbbe5a 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/JobACLsManager.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/JobACLsManager.java @@ -101,7 +101,6 @@ public class JobACLsManager { * @param jobOperation * @param jobOwner * @param jobACL - * @throws AccessControlException */ public boolean checkAccess(UserGroupInformation callerUGI, JobACL jobOperation, String jobOwner, AccessControlList jobACL) { http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/JobClient.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/JobClient.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/JobClient.java index 89a966e..e91fbfe 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/JobClient.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/JobClient.java @@ -79,7 +79,7 @@ import org.apache.hadoop.util.ToolRunner; * Submitting the job to the cluster and optionally monitoring * it's status. * </li> - * </ol></p> + * </ol> * * Normally the user creates the application, describes various facets of the * job via {@link JobConf} and then uses the <code>JobClient</code> to submit @@ -101,9 +101,9 @@ import org.apache.hadoop.util.ToolRunner; * * // Submit the job, then poll for progress until the job is complete * JobClient.runJob(job); - * </pre></blockquote></p> + * </pre></blockquote> * - * <h4 id="JobControl">Job Control</h4> + * <b id="JobControl">Job Control</b> * * <p>At times clients would chain map-reduce jobs to accomplish complex tasks * which cannot be done via a single map-reduce job. This is fairly easy since @@ -127,7 +127,7 @@ import org.apache.hadoop.util.ToolRunner; * {@link JobConf#setJobEndNotificationURI(String)} : setup a notification * on job-completion, thus avoiding polling. * </li> - * </ol></p> + * </ol> * * @see JobConf * @see ClusterStatus http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/JobConf.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/JobConf.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/JobConf.java index 315c829..c388bda 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/JobConf.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/JobConf.java @@ -74,7 +74,7 @@ import org.apache.log4j.Level; * more complex for the user to control finely * (e.g. {@link #setNumMapTasks(int)}). * </li> - * </ol></p> + * </ol> * * <p><code>JobConf</code> typically specifies the {@link Mapper}, combiner * (if any), {@link Partitioner}, {@link Reducer}, {@link InputFormat} and @@ -105,7 +105,7 @@ import org.apache.log4j.Level; * * job.setInputFormat(SequenceFileInputFormat.class); * job.setOutputFormat(SequenceFileOutputFormat.class); - * </pre></blockquote></p> + * </pre></blockquote> * * @see JobClient * @see ClusterStatus @@ -486,7 +486,7 @@ public class JobConf extends Configuration { /** A new map/reduce configuration where the behavior of reading from the * default resources can be turned off. - * <p/> + * <p> * If the parameter {@code loadDefaults} is false, the new instance * will not load resources from the default files. * @@ -993,19 +993,19 @@ public class JobConf extends Configuration { /** * Set the user defined {@link RawComparator} comparator for * grouping keys in the input to the combiner. - * <p/> + * * <p>This comparator should be provided if the equivalence rules for keys * for sorting the intermediates are different from those for grouping keys * before each call to * {@link Reducer#reduce(Object, java.util.Iterator, OutputCollector, Reporter)}.</p> - * <p/> + * * <p>For key-value pairs (K1,V1) and (K2,V2), the values (V1, V2) are passed * in a single call to the reduce function if K1 and K2 compare as equal.</p> - * <p/> + * * <p>Since {@link #setOutputKeyComparatorClass(Class)} can be used to control * how keys are sorted, this can be used in conjunction to simulate * <i>secondary sort on values</i>.</p> - * <p/> + * * <p><i>Note</i>: This is not a guarantee of the combiner sort being * <i>stable</i> in any sense. (In any case, with the order of available * map-outputs to the combiner being non-deterministic, it wouldn't make @@ -1210,7 +1210,7 @@ public class JobConf extends Configuration { * <li> be side-effect free</li> * <li> have the same input and output key types and the same input and * output value types</li> - * </ul></p> + * </ul> * * <p>Typically the combiner is same as the <code>Reducer</code> for the * job i.e. {@link #setReducerClass(Class)}.</p> @@ -1309,7 +1309,7 @@ public class JobConf extends Configuration { * A custom {@link InputFormat} is typically used to accurately control * the number of map tasks for the job.</p> * - * <h4 id="NoOfMaps">How many maps?</h4> + * <b id="NoOfMaps">How many maps?</b> * * <p>The number of maps is usually driven by the total size of the inputs * i.e. total number of blocks of the input files.</p> @@ -1350,7 +1350,7 @@ public class JobConf extends Configuration { /** * Set the requisite number of reduce tasks for this job. * - * <h4 id="NoOfReduces">How many reduces?</h4> + * <b id="NoOfReduces">How many reduces?</b> * * <p>The right number of reduces seems to be <code>0.95</code> or * <code>1.75</code> multiplied by (<<i>no. of nodes</i>> * @@ -1370,7 +1370,7 @@ public class JobConf extends Configuration { * reserve a few reduce slots in the framework for speculative-tasks, failures * etc.</p> * - * <h4 id="ReducerNone">Reducer NONE</h4> + * <b id="ReducerNone">Reducer NONE</b> * * <p>It is legal to set the number of reduce-tasks to <code>zero</code>.</p> * @@ -1693,9 +1693,9 @@ public class JobConf extends Configuration { * given task's stdout, stderr, syslog, jobconf files as arguments.</p> * * <p>The debug command, run on the node where the map failed, is:</p> - * <p><pre><blockquote> + * <p><blockquote><pre> * $script $stdout $stderr $syslog $jobconf. - * </blockquote></pre></p> + * </pre></blockquote> * * <p> The script file is distributed through {@link DistributedCache} * APIs. The script needs to be symlinked. </p> @@ -1705,7 +1705,7 @@ public class JobConf extends Configuration { * job.setMapDebugScript("./myscript"); * DistributedCache.createSymlink(job); * DistributedCache.addCacheFile("/debug/scripts/myscript#myscript"); - * </pre></blockquote></p> + * </pre></blockquote> * * @param mDbgScript the script name */ @@ -1730,9 +1730,9 @@ public class JobConf extends Configuration { * is given task's stdout, stderr, syslog, jobconf files as arguments.</p> * * <p>The debug command, run on the node where the map failed, is:</p> - * <p><pre><blockquote> + * <p><blockquote><pre> * $script $stdout $stderr $syslog $jobconf. - * </blockquote></pre></p> + * </pre></blockquote> * * <p> The script file is distributed through {@link DistributedCache} * APIs. The script file needs to be symlinked </p> @@ -1742,7 +1742,7 @@ public class JobConf extends Configuration { * job.setReduceDebugScript("./myscript"); * DistributedCache.createSymlink(job); * DistributedCache.addCacheFile("/debug/scripts/myscript#myscript"); - * </pre></blockquote></p> + * </pre></blockquote> * * @param rDbgScript the script name */ @@ -1785,8 +1785,6 @@ public class JobConf extends Configuration { * * @param uri the job end notification uri * @see JobStatus - * @see <a href="{@docRoot}/org/apache/hadoop/mapred/JobClient.html# - * JobCompletionAndChaining">Job Completion and Chaining</a> */ public void setJobEndNotificationURI(String uri) { set(JobContext.MR_JOB_END_NOTIFICATION_URL, uri); @@ -1816,7 +1814,7 @@ public class JobConf extends Configuration { * * If a value is specified in the configuration, it is returned. * Else, it returns {@link JobContext#DEFAULT_MAP_MEMORY_MB}. - * <p/> + * <p> * For backward compatibility, if the job configuration sets the * key {@link #MAPRED_TASK_MAXVMEM_PROPERTY} to a value different * from {@link #DISABLED_MEMORY_LIMIT}, that value will be used @@ -1842,7 +1840,7 @@ public class JobConf extends Configuration { * * If a value is specified in the configuration, it is returned. * Else, it returns {@link JobContext#DEFAULT_REDUCE_MEMORY_MB}. - * <p/> + * <p> * For backward compatibility, if the job configuration sets the * key {@link #MAPRED_TASK_MAXVMEM_PROPERTY} to a value different * from {@link #DISABLED_MEMORY_LIMIT}, that value will be used @@ -1915,7 +1913,6 @@ public class JobConf extends Configuration { * * @param my_class the class to find. * @return a jar file that contains the class, or null. - * @throws IOException */ public static String findContainingJar(Class my_class) { return ClassUtil.findContainingJar(my_class); @@ -1924,10 +1921,10 @@ public class JobConf extends Configuration { /** * Get the memory required to run a task of this job, in bytes. See * {@link #MAPRED_TASK_MAXVMEM_PROPERTY} - * <p/> + * <p> * This method is deprecated. Now, different memory limits can be * set for map and reduce tasks of a job, in MB. - * <p/> + * <p> * For backward compatibility, if the job configuration sets the * key {@link #MAPRED_TASK_MAXVMEM_PROPERTY}, that value is returned. * Otherwise, this method will return the larger of the values returned by @@ -1953,7 +1950,7 @@ public class JobConf extends Configuration { /** * Set the maximum amount of memory any task of this job can use. See * {@link #MAPRED_TASK_MAXVMEM_PROPERTY} - * <p/> + * <p> * mapred.task.maxvmem is split into * mapreduce.map.memory.mb * and mapreduce.map.memory.mb,mapred @@ -2073,7 +2070,7 @@ public class JobConf extends Configuration { /** * Parse the Maximum heap size from the java opts as specified by the -Xmx option - * Format: -Xmx<size>[g|G|m|M|k|K] + * Format: -Xmx<size>[g|G|m|M|k|K] * @param javaOpts String to parse to read maximum heap size * @return Maximum heap size in MB or -1 if not specified */ http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/Mapper.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/Mapper.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/Mapper.java index eaa6c2b..ac2c96d 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/Mapper.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/Mapper.java @@ -117,7 +117,7 @@ import org.apache.hadoop.io.compress.CompressionCodec; * output.collect(key, val); * } * } - * </pre></blockquote></p> + * </pre></blockquote> * * <p>Applications may write a custom {@link MapRunnable} to exert greater * control on map processing e.g. multi-threaded <code>Mapper</code>s etc.</p> http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/QueueManager.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/QueueManager.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/QueueManager.java index 39fae2a..794c55d 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/QueueManager.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/QueueManager.java @@ -46,20 +46,20 @@ import java.net.URL; /** * Class that exposes information about queues maintained by the Hadoop * Map/Reduce framework. - * <p/> + * <p> * The Map/Reduce framework can be configured with one or more queues, * depending on the scheduler it is configured with. While some * schedulers work only with one queue, some schedulers support multiple * queues. Some schedulers also support the notion of queues within * queues - a feature called hierarchical queues. - * <p/> + * <p> * Queue names are unique, and used as a key to lookup queues. Hierarchical * queues are named by a 'fully qualified name' such as q1:q2:q3, where * q2 is a child queue of q1 and q3 is a child queue of q2. - * <p/> + * <p> * Leaf level queues are queues that contain no queues within them. Jobs * can be submitted only to leaf level queues. - * <p/> + * <p> * Queues can be configured with various properties. Some of these * properties are common to all schedulers, and those are handled by this * class. Schedulers might also associate several custom properties with @@ -69,11 +69,11 @@ import java.net.URL; * provided by the framework, but define their own mechanisms. In such cases, * it is likely that the name of the queue will be used to relate the * common properties of a queue with scheduler specific properties. - * <p/> + * <p> * Information related to a queue, such as its name, properties, scheduling * information and children are exposed by this class via a serializable * class called {@link JobQueueInfo}. - * <p/> + * <p> * Queues are configured in the configuration file mapred-queues.xml. * To support backwards compatibility, queues can also be configured * in mapred-site.xml. However, when configured in the latter, there is @@ -102,7 +102,7 @@ public class QueueManager { /** * Factory method to create an appropriate instance of a queue * configuration parser. - * <p/> + * <p> * Returns a parser that can parse either the deprecated property * style queue configuration in mapred-site.xml, or one that can * parse hierarchical queues in mapred-queues.xml. First preference @@ -157,7 +157,7 @@ public class QueueManager { /** * Construct a new QueueManager using configuration specified in the passed * in {@link org.apache.hadoop.conf.Configuration} object. - * <p/> + * <p> * This instance supports queue configuration specified in mapred-site.xml, * but without support for hierarchical queues. If no queue configuration * is found in mapred-site.xml, it will then look for site configuration @@ -173,7 +173,7 @@ public class QueueManager { /** * Create an instance that supports hierarchical queues, defined in * the passed in configuration file. - * <p/> + * <p> * This is mainly used for testing purposes and should not called from * production code. * @@ -208,7 +208,7 @@ public class QueueManager { /** * Return the set of leaf level queues configured in the system to * which jobs are submitted. - * <p/> + * <p> * The number of queues configured should be dependent on the Scheduler * configured. Note that some schedulers work with only one queue, whereas * others can support multiple queues. @@ -222,7 +222,7 @@ public class QueueManager { /** * Return true if the given user is part of the ACL for the given * {@link QueueACL} name for the given queue. - * <p/> + * <p> * An operation is allowed if all users are provided access for this * operation, or if either the user or any of the groups specified is * provided access. @@ -283,7 +283,7 @@ public class QueueManager { /** * Set a generic Object that represents scheduling information relevant * to a queue. - * <p/> + * <p> * A string representation of this Object will be used by the framework * to display in user facing applications like the JobTracker web UI and * the hadoop CLI. @@ -323,7 +323,7 @@ public class QueueManager { /** * Refresh acls, state and scheduler properties for the configured queues. - * <p/> + * <p> * This method reloads configuration related to queues, but does not * support changes to the list of queues or hierarchy. The expected usage * is that an administrator can modify the queue configuration file and @@ -431,7 +431,7 @@ public class QueueManager { /** * JobQueueInfo for all the queues. - * <p/> + * <p> * Contribs can use this data structure to either create a hierarchy or for * traversing. * They can also use this to refresh properties in case of refreshQueues @@ -450,7 +450,7 @@ public class QueueManager { /** * Generates the array of QueueAclsInfo object. - * <p/> + * <p> * The array consists of only those queues for which user has acls. * * @return QueueAclsInfo[] http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/RecordReader.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/RecordReader.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/RecordReader.java index 0c95a14..6e2c89f 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/RecordReader.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/RecordReader.java @@ -29,7 +29,7 @@ import org.apache.hadoop.classification.InterfaceStability; * * <p><code>RecordReader</code>, typically, converts the byte-oriented view of * the input, provided by the <code>InputSplit</code>, and presents a - * record-oriented view for the {@link Mapper} & {@link Reducer} tasks for + * record-oriented view for the {@link Mapper} and {@link Reducer} tasks for * processing. It thus assumes the responsibility of processing record * boundaries and presenting the tasks with keys and values.</p> * http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/Reducer.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/Reducer.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/Reducer.java index 3fefa4b..962e195 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/Reducer.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/Reducer.java @@ -42,7 +42,7 @@ import org.apache.hadoop.io.Closeable; * <ol> * <li> * - * <h4 id="Shuffle">Shuffle</h4> + * <b id="Shuffle">Shuffle</b> * * <p><code>Reducer</code> is input the grouped output of a {@link Mapper}. * In the phase the framework, for each <code>Reducer</code>, fetches the @@ -51,7 +51,7 @@ import org.apache.hadoop.io.Closeable; * </li> * * <li> - * <h4 id="Sort">Sort</h4> + * <b id="Sort">Sort</b> * * <p>The framework groups <code>Reducer</code> inputs by <code>key</code>s * (since different <code>Mapper</code>s may have output the same key) in this @@ -60,7 +60,7 @@ import org.apache.hadoop.io.Closeable; * <p>The shuffle and sort phases occur simultaneously i.e. while outputs are * being fetched they are merged.</p> * - * <h5 id="SecondarySort">SecondarySort</h5> + * <b id="SecondarySort">SecondarySort</b> * * <p>If equivalence rules for keys while grouping the intermediates are * different from those for grouping keys before reduction, then one may @@ -86,11 +86,11 @@ import org.apache.hadoop.io.Closeable; * </li> * * <li> - * <h4 id="Reduce">Reduce</h4> + * <b id="Reduce">Reduce</b> * * <p>In this phase the * {@link #reduce(Object, Iterator, OutputCollector, Reporter)} - * method is called for each <code><key, (list of values)></code> pair in + * method is called for each <code><key, (list of values)></code> pair in * the grouped inputs.</p> * <p>The output of the reduce task is typically written to the * {@link FileSystem} via @@ -156,7 +156,7 @@ import org.apache.hadoop.io.Closeable; * } * } * } - * </pre></blockquote></p> + * </pre></blockquote> * * @see Mapper * @see Partitioner @@ -171,7 +171,7 @@ public interface Reducer<K2, V2, K3, V3> extends JobConfigurable, Closeable { * <i>Reduces</i> values for a given key. * * <p>The framework calls this method for each - * <code><key, (list of values)></code> pair in the grouped inputs. + * <code><key, (list of values)></code> pair in the grouped inputs. * Output values must be of the same type as input values. Input keys must * not be altered. The framework will <b>reuse</b> the key and value objects * that are passed into the reduce, therefore the application should clone http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/TaskUmbilicalProtocol.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/TaskUmbilicalProtocol.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/TaskUmbilicalProtocol.java index 5df02c7..c3678d6 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/TaskUmbilicalProtocol.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/TaskUmbilicalProtocol.java @@ -178,7 +178,6 @@ public interface TaskUmbilicalProtocol extends VersionedProtocol { * * @param taskID task's id * @return the most recent checkpoint (if any) for this task - * @throws IOException */ TaskCheckpointID getCheckpointID(TaskID taskID); http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/ChainMapper.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/ChainMapper.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/ChainMapper.java index 14f040a..723a234 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/ChainMapper.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/ChainMapper.java @@ -29,61 +29,61 @@ import java.io.IOException; /** * The ChainMapper class allows to use multiple Mapper classes within a single * Map task. - * <p/> + * <p> * The Mapper classes are invoked in a chained (or piped) fashion, the output of * the first becomes the input of the second, and so on until the last Mapper, * the output of the last Mapper will be written to the task's output. - * <p/> + * <p> * The key functionality of this feature is that the Mappers in the chain do not * need to be aware that they are executed in a chain. This enables having * reusable specialized Mappers that can be combined to perform composite * operations within a single task. - * <p/> + * <p> * Special care has to be taken when creating chains that the key/values output * by a Mapper are valid for the following Mapper in the chain. It is assumed * all Mappers and the Reduce in the chain use maching output and input key and * value classes as no conversion is done by the chaining code. - * <p/> + * <p> * Using the ChainMapper and the ChainReducer classes is possible to compose * Map/Reduce jobs that look like <code>[MAP+ / REDUCE MAP*]</code>. And * immediate benefit of this pattern is a dramatic reduction in disk IO. - * <p/> + * <p> * IMPORTANT: There is no need to specify the output key/value classes for the * ChainMapper, this is done by the addMapper for the last mapper in the chain. - * <p/> + * <p> * ChainMapper usage pattern: - * <p/> + * <p> * <pre> * ... * conf.setJobName("chain"); * conf.setInputFormat(TextInputFormat.class); * conf.setOutputFormat(TextOutputFormat.class); - * <p/> + * * JobConf mapAConf = new JobConf(false); * ... * ChainMapper.addMapper(conf, AMap.class, LongWritable.class, Text.class, * Text.class, Text.class, true, mapAConf); - * <p/> + * * JobConf mapBConf = new JobConf(false); * ... * ChainMapper.addMapper(conf, BMap.class, Text.class, Text.class, * LongWritable.class, Text.class, false, mapBConf); - * <p/> + * * JobConf reduceConf = new JobConf(false); * ... * ChainReducer.setReducer(conf, XReduce.class, LongWritable.class, Text.class, * Text.class, Text.class, true, reduceConf); - * <p/> + * * ChainReducer.addMapper(conf, CMap.class, Text.class, Text.class, * LongWritable.class, Text.class, false, null); - * <p/> + * * ChainReducer.addMapper(conf, DMap.class, LongWritable.class, Text.class, * LongWritable.class, LongWritable.class, true, null); - * <p/> + * * FileInputFormat.setInputPaths(conf, inDir); * FileOutputFormat.setOutputPath(conf, outDir); * ... - * <p/> + * * JobClient jc = new JobClient(conf); * RunningJob job = jc.submitJob(conf); * ... @@ -95,21 +95,21 @@ public class ChainMapper implements Mapper { /** * Adds a Mapper class to the chain job's JobConf. - * <p/> + * <p> * It has to be specified how key and values are passed from one element of * the chain to the next, by value or by reference. If a Mapper leverages the * assumed semantics that the key and values are not modified by the collector * 'by value' must be used. If the Mapper does not expect this semantics, as * an optimization to avoid serialization and deserialization 'by reference' * can be used. - * <p/> + * <p> * For the added Mapper the configuration given for it, * <code>mapperConf</code>, have precedence over the job's JobConf. This * precedence is in effect when the task is running. - * <p/> + * <p> * IMPORTANT: There is no need to specify the output key/value classes for the * ChainMapper, this is done by the addMapper for the last mapper in the chain - * <p/> + * <p> * * @param job job's JobConf to add the Mapper class. * @param klass the Mapper class to add. @@ -148,7 +148,7 @@ public class ChainMapper implements Mapper { /** * Configures the ChainMapper and all the Mappers in the chain. - * <p/> + * <p> * If this method is overriden <code>super.configure(...)</code> should be * invoked at the beginning of the overwriter method. */ @@ -171,7 +171,7 @@ public class ChainMapper implements Mapper { /** * Closes the ChainMapper and all the Mappers in the chain. - * <p/> + * <p> * If this method is overriden <code>super.close()</code> should be * invoked at the end of the overwriter method. */ http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/ChainReducer.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/ChainReducer.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/ChainReducer.java index 641d82c..6f5b7cd 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/ChainReducer.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/ChainReducer.java @@ -27,63 +27,63 @@ import java.util.Iterator; /** * The ChainReducer class allows to chain multiple Mapper classes after a * Reducer within the Reducer task. - * <p/> + * <p> * For each record output by the Reducer, the Mapper classes are invoked in a * chained (or piped) fashion, the output of the first becomes the input of the * second, and so on until the last Mapper, the output of the last Mapper will * be written to the task's output. - * <p/> + * <p> * The key functionality of this feature is that the Mappers in the chain do not * need to be aware that they are executed after the Reducer or in a chain. * This enables having reusable specialized Mappers that can be combined to * perform composite operations within a single task. - * <p/> + * <p> * Special care has to be taken when creating chains that the key/values output * by a Mapper are valid for the following Mapper in the chain. It is assumed * all Mappers and the Reduce in the chain use maching output and input key and * value classes as no conversion is done by the chaining code. - * <p/> + * <p> * Using the ChainMapper and the ChainReducer classes is possible to compose * Map/Reduce jobs that look like <code>[MAP+ / REDUCE MAP*]</code>. And * immediate benefit of this pattern is a dramatic reduction in disk IO. - * <p/> + * <p> * IMPORTANT: There is no need to specify the output key/value classes for the * ChainReducer, this is done by the setReducer or the addMapper for the last * element in the chain. - * <p/> + * <p> * ChainReducer usage pattern: - * <p/> + * <p> * <pre> * ... * conf.setJobName("chain"); * conf.setInputFormat(TextInputFormat.class); * conf.setOutputFormat(TextOutputFormat.class); - * <p/> + * * JobConf mapAConf = new JobConf(false); * ... * ChainMapper.addMapper(conf, AMap.class, LongWritable.class, Text.class, * Text.class, Text.class, true, mapAConf); - * <p/> + * * JobConf mapBConf = new JobConf(false); * ... * ChainMapper.addMapper(conf, BMap.class, Text.class, Text.class, * LongWritable.class, Text.class, false, mapBConf); - * <p/> + * * JobConf reduceConf = new JobConf(false); * ... * ChainReducer.setReducer(conf, XReduce.class, LongWritable.class, Text.class, * Text.class, Text.class, true, reduceConf); - * <p/> + * * ChainReducer.addMapper(conf, CMap.class, Text.class, Text.class, * LongWritable.class, Text.class, false, null); - * <p/> + * * ChainReducer.addMapper(conf, DMap.class, LongWritable.class, Text.class, * LongWritable.class, LongWritable.class, true, null); - * <p/> + * * FileInputFormat.setInputPaths(conf, inDir); * FileOutputFormat.setOutputPath(conf, outDir); * ... - * <p/> + * * JobClient jc = new JobClient(conf); * RunningJob job = jc.submitJob(conf); * ... @@ -95,18 +95,18 @@ public class ChainReducer implements Reducer { /** * Sets the Reducer class to the chain job's JobConf. - * <p/> + * <p> * It has to be specified how key and values are passed from one element of * the chain to the next, by value or by reference. If a Reducer leverages the * assumed semantics that the key and values are not modified by the collector * 'by value' must be used. If the Reducer does not expect this semantics, as * an optimization to avoid serialization and deserialization 'by reference' * can be used. - * <p/> + * <p> * For the added Reducer the configuration given for it, * <code>reducerConf</code>, have precedence over the job's JobConf. This * precedence is in effect when the task is running. - * <p/> + * <p> * IMPORTANT: There is no need to specify the output key/value classes for the * ChainReducer, this is done by the setReducer or the addMapper for the last * element in the chain. @@ -139,18 +139,18 @@ public class ChainReducer implements Reducer { /** * Adds a Mapper class to the chain job's JobConf. - * <p/> + * <p> * It has to be specified how key and values are passed from one element of * the chain to the next, by value or by reference. If a Mapper leverages the * assumed semantics that the key and values are not modified by the collector * 'by value' must be used. If the Mapper does not expect this semantics, as * an optimization to avoid serialization and deserialization 'by reference' * can be used. - * <p/> + * <p> * For the added Mapper the configuration given for it, * <code>mapperConf</code>, have precedence over the job's JobConf. This * precedence is in effect when the task is running. - * <p/> + * <p> * IMPORTANT: There is no need to specify the output key/value classes for the * ChainMapper, this is done by the addMapper for the last mapper in the chain * . @@ -191,7 +191,7 @@ public class ChainReducer implements Reducer { /** * Configures the ChainReducer, the Reducer and all the Mappers in the chain. - * <p/> + * <p> * If this method is overriden <code>super.configure(...)</code> should be * invoked at the beginning of the overwriter method. */ @@ -215,7 +215,7 @@ public class ChainReducer implements Reducer { /** * Closes the ChainReducer, the Reducer and all the Mappers in the chain. - * <p/> + * <p> * If this method is overriden <code>super.close()</code> should be * invoked at the end of the overwriter method. */ http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/MultipleOutputs.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/MultipleOutputs.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/MultipleOutputs.java index 39e80f9..f0f3652 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/MultipleOutputs.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/MultipleOutputs.java @@ -31,29 +31,29 @@ import java.util.*; * than the job default output via the <code>OutputCollector</code> passed to * the <code>map()</code> and <code>reduce()</code> methods of the * <code>Mapper</code> and <code>Reducer</code> implementations. - * <p/> + * <p> * Each additional output, or named output, may be configured with its own * <code>OutputFormat</code>, with its own key class and with its own value * class. - * <p/> + * <p> * A named output can be a single file or a multi file. The later is refered as * a multi named output. - * <p/> + * <p> * A multi named output is an unbound set of files all sharing the same * <code>OutputFormat</code>, key class and value class configuration. - * <p/> + * <p> * When named outputs are used within a <code>Mapper</code> implementation, * key/values written to a name output are not part of the reduce phase, only * key/values written to the job <code>OutputCollector</code> are part of the * reduce phase. - * <p/> + * <p> * MultipleOutputs supports counters, by default the are disabled. The counters * group is the {@link MultipleOutputs} class name. * </p> * The names of the counters are the same as the named outputs. For multi * named outputs the name of the counter is the concatenation of the named * output, and underscore '_' and the multiname. - * <p/> + * <p> * Job configuration usage pattern is: * <pre> * @@ -82,7 +82,7 @@ import java.util.*; * * ... * </pre> - * <p/> + * <p> * Job configuration usage pattern is: * <pre> * @@ -271,7 +271,6 @@ public class MultipleOutputs { /** * Adds a named output for the job. - * <p/> * * @param conf job conf to add the named output * @param namedOutput named output name, it has to be a word, letters @@ -291,7 +290,6 @@ public class MultipleOutputs { /** * Adds a multi named output for the job. - * <p/> * * @param conf job conf to add the named output * @param namedOutput named output name, it has to be a word, letters @@ -311,7 +309,6 @@ public class MultipleOutputs { /** * Adds a named output for the job. - * <p/> * * @param conf job conf to add the named output * @param namedOutput named output name, it has to be a word, letters @@ -339,9 +336,9 @@ public class MultipleOutputs { /** * Enables or disables counters for the named outputs. - * <p/> + * <p> * By default these counters are disabled. - * <p/> + * <p> * MultipleOutputs supports counters, by default the are disabled. * The counters group is the {@link MultipleOutputs} class name. * </p> @@ -358,9 +355,9 @@ public class MultipleOutputs { /** * Returns if the counters for the named outputs are enabled or not. - * <p/> + * <p> * By default these counters are disabled. - * <p/> + * <p> * MultipleOutputs supports counters, by default the are disabled. * The counters group is the {@link MultipleOutputs} class name. * </p> @@ -465,7 +462,6 @@ public class MultipleOutputs { /** * Gets the output collector for a named output. - * <p/> * * @param namedOutput the named output name * @param reporter the reporter @@ -480,7 +476,6 @@ public class MultipleOutputs { /** * Gets the output collector for a multi named output. - * <p/> * * @param namedOutput the named output name * @param multiName the multi name part @@ -525,7 +520,7 @@ public class MultipleOutputs { /** * Closes all the opened named outputs. - * <p/> + * <p> * If overriden subclasses must invoke <code>super.close()</code> at the * end of their <code>close()</code> * http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/TokenCountMapper.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/TokenCountMapper.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/TokenCountMapper.java index 8e884ce..75179e1 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/TokenCountMapper.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/TokenCountMapper.java @@ -32,7 +32,7 @@ import org.apache.hadoop.mapred.Reporter; /** - * A {@link Mapper} that maps text values into <token,freq> pairs. Uses + * A {@link Mapper} that maps text values into <token,freq> pairs. Uses * {@link StringTokenizer} to break text into tokens. */ @InterfaceAudience.Public http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/aggregate/ValueAggregatorJob.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/aggregate/ValueAggregatorJob.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/aggregate/ValueAggregatorJob.java index 8c20723..6251925 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/aggregate/ValueAggregatorJob.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/aggregate/ValueAggregatorJob.java @@ -60,7 +60,7 @@ import org.apache.hadoop.util.GenericOptionsParser; * The developer using Aggregate will need only to provide a plugin class * conforming to the following interface: * - * public interface ValueAggregatorDescriptor { public ArrayList<Entry> + * public interface ValueAggregatorDescriptor { public ArrayList<Entry> * generateKeyValPairs(Object key, Object value); public void * configure(JobConfjob); } * http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/aggregate/ValueAggregatorReducer.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/aggregate/ValueAggregatorReducer.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/aggregate/ValueAggregatorReducer.java index a6b3573..2738968 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/aggregate/ValueAggregatorReducer.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/aggregate/ValueAggregatorReducer.java @@ -45,7 +45,8 @@ public class ValueAggregatorReducer<K1 extends WritableComparable, * driven computing is achieved. It is assumed that each aggregator's * getReport method emits appropriate output for the aggregator. This * may be further customiized. - * @value the values to be aggregated + * @param values + * the values to be aggregated */ public void reduce(Text key, Iterator<Text> values, OutputCollector<Text, Text> output, Reporter reporter) throws IOException { http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/db/DBInputFormat.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/db/DBInputFormat.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/db/DBInputFormat.java index 2715705..159919f 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/db/DBInputFormat.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/lib/db/DBInputFormat.java @@ -195,8 +195,8 @@ public class DBInputFormat<T extends DBWritable> * @param inputClass the class object implementing DBWritable, which is the * Java object holding tuple fields. * @param tableName The table to read data from - * @param conditions The condition which to select data with, eg. '(updated > - * 20070101 AND length > 0)' + * @param conditions The condition which to select data with, eg. '(updated > + * 20070101 AND length > 0)' * @param orderBy the fieldNames in the orderBy clause. * @param fieldNames The field names in the table * @see #setInput(JobConf, Class, String, String) http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/Cluster.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/Cluster.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/Cluster.java index 60ff715..34353ac 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/Cluster.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/Cluster.java @@ -134,6 +134,7 @@ public class Cluster { /** * Close the <code>Cluster</code>. + * @throws IOException */ public synchronized void close() throws IOException { clientProtocolProvider.close(client); http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/ClusterMetrics.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/ClusterMetrics.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/ClusterMetrics.java index c4c2778..b5e54b5 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/ClusterMetrics.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/ClusterMetrics.java @@ -40,15 +40,15 @@ import org.apache.hadoop.io.Writable; * Slot capacity of the cluster. * </li> * <li> - * The number of currently occupied/reserved map & reduce slots. + * The number of currently occupied/reserved map and reduce slots. * </li> * <li> - * The number of currently running map & reduce tasks. + * The number of currently running map and reduce tasks. * </li> * <li> * The number of job submissions. * </li> - * </ol></p> + * </ol> * * <p>Clients can query for the latest <code>ClusterMetrics</code>, via * {@link Cluster#getClusterStatus()}.</p> http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/CryptoUtils.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/CryptoUtils.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/CryptoUtils.java index 184cdf0..ef06176 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/CryptoUtils.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/CryptoUtils.java @@ -123,11 +123,11 @@ public class CryptoUtils { * "mapreduce.job.encrypted-intermediate-data.buffer.kb" Job configuration * variable. * - * If the value of 'length' is > -1, The InputStream is additionally wrapped - * in a LimitInputStream. CryptoStreams are late buffering in nature. This - * means they will always try to read ahead if they can. The LimitInputStream - * will ensure that the CryptoStream does not read past the provided length - * from the given Input Stream. + * If the value of 'length' is > -1, The InputStream is additionally + * wrapped in a LimitInputStream. CryptoStreams are late buffering in nature. + * This means they will always try to read ahead if they can. The + * LimitInputStream will ensure that the CryptoStream does not read past the + * provided length from the given Input Stream. * * @param conf * @param in http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/Job.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/Job.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/Job.java index 470290c..f404175 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/Job.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/Job.java @@ -69,7 +69,7 @@ import org.apache.hadoop.yarn.api.records.ReservationId; * * // Submit the job, then poll for progress until the job is complete * job.waitForCompletion(true); - * </pre></blockquote></p> + * </pre></blockquote> * * */ http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/JobContext.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/JobContext.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/JobContext.java index 836f182..6bd2d1f 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/JobContext.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/JobContext.java @@ -289,7 +289,6 @@ public interface JobContext extends MRJobConfig { * Get the timestamps of the archives. Used by internal * DistributedCache and MapReduce code. * @return a string array of timestamps - * @throws IOException */ public String[] getArchiveTimestamps(); @@ -297,7 +296,6 @@ public interface JobContext extends MRJobConfig { * Get the timestamps of the files. Used by internal * DistributedCache and MapReduce code. * @return a string array of timestamps - * @throws IOException */ public String[] getFileTimestamps(); http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/JobSubmissionFiles.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/JobSubmissionFiles.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/JobSubmissionFiles.java index 516e661..7125077 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/JobSubmissionFiles.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/JobSubmissionFiles.java @@ -100,7 +100,7 @@ public class JobSubmissionFiles { /** * Initializes the staging directory and returns the path. It also - * keeps track of all necessary ownership & permissions + * keeps track of all necessary ownership and permissions * @param cluster * @param conf */ http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/Mapper.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/Mapper.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/Mapper.java index 3a6186b..6b4147b 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/Mapper.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/Mapper.java @@ -42,9 +42,9 @@ import org.apache.hadoop.mapreduce.task.MapContextImpl; * * <p>The framework first calls * {@link #setup(org.apache.hadoop.mapreduce.Mapper.Context)}, followed by - * {@link #map(Object, Object, Context)} + * {@link #map(Object, Object, org.apache.hadoop.mapreduce.Mapper.Context)} * for each key/value pair in the <code>InputSplit</code>. Finally - * {@link #cleanup(Context)} is called.</p> + * {@link #cleanup(org.apache.hadoop.mapreduce.Mapper.Context)} is called.</p> * * <p>All intermediate values associated with a given output key are * subsequently grouped by the framework, and passed to a {@link Reducer} to @@ -84,9 +84,10 @@ import org.apache.hadoop.mapreduce.task.MapContextImpl; * } * } * } - * </pre></blockquote></p> + * </pre></blockquote> * - * <p>Applications may override the {@link #run(Context)} method to exert + * <p>Applications may override the + * {@link #run(org.apache.hadoop.mapreduce.Mapper.Context)} method to exert * greater control on map processing e.g. multi-threaded <code>Mapper</code>s * etc.</p> * http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/Reducer.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/Reducer.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/Reducer.java index ddf67e1..ab67ab0 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/Reducer.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/Reducer.java @@ -39,14 +39,14 @@ import java.util.Iterator; * <ol> * <li> * - * <h4 id="Shuffle">Shuffle</h4> + * <b id="Shuffle">Shuffle</b> * * <p>The <code>Reducer</code> copies the sorted output from each * {@link Mapper} using HTTP across the network.</p> * </li> * * <li> - * <h4 id="Sort">Sort</h4> + * <b id="Sort">Sort</b> * * <p>The framework merge sorts <code>Reducer</code> inputs by * <code>key</code>s @@ -55,7 +55,7 @@ import java.util.Iterator; * <p>The shuffle and sort phases occur simultaneously i.e. while outputs are * being fetched they are merged.</p> * - * <h5 id="SecondarySort">SecondarySort</h5> + * <b id="SecondarySort">SecondarySort</b> * * <p>To achieve a secondary sort on the values returned by the value * iterator, the application should extend the key with the secondary @@ -83,10 +83,10 @@ import java.util.Iterator; * </li> * * <li> - * <h4 id="Reduce">Reduce</h4> + * <b id="Reduce">Reduce</b> * * <p>In this phase the - * {@link #reduce(Object, Iterable, Context)} + * {@link #reduce(Object, Iterable, org.apache.hadoop.mapreduce.Reducer.Context)} * method is called for each <code><key, (collection of values)></code> in * the sorted inputs.</p> * <p>The output of the reduce task is typically written to a @@ -113,7 +113,7 @@ import java.util.Iterator; * context.write(key, result); * } * } - * </pre></blockquote></p> + * </pre></blockquote> * * @see Mapper * @see Partitioner http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/filecache/DistributedCache.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/filecache/DistributedCache.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/filecache/DistributedCache.java index 06737c9..51fe69a 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/filecache/DistributedCache.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/filecache/DistributedCache.java @@ -115,7 +115,7 @@ import java.net.URI; * } * } * - * </pre></blockquote></p> + * </pre></blockquote> * * It is also very common to use the DistributedCache by using * {@link org.apache.hadoop.util.GenericOptionsParser}. @@ -235,7 +235,6 @@ public class DistributedCache { * DistributedCache and MapReduce code. * @param conf The configuration which stored the timestamps * @return a long array of timestamps - * @throws IOException * @deprecated Use {@link JobContext#getArchiveTimestamps()} instead */ @Deprecated @@ -250,7 +249,6 @@ public class DistributedCache { * DistributedCache and MapReduce code. * @param conf The configuration which stored the timestamps * @return a long array of timestamps - * @throws IOException * @deprecated Use {@link JobContext#getFileTimestamps()} instead */ @Deprecated @@ -434,7 +432,6 @@ public class DistributedCache { * internal DistributedCache and MapReduce code. * @param conf The configuration which stored the timestamps * @return a string array of booleans - * @throws IOException */ public static boolean[] getFileVisibilities(Configuration conf) { return parseBooleans(conf.getStrings(MRJobConfig.CACHE_FILE_VISIBILITIES)); http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/aggregate/ValueAggregatorJob.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/aggregate/ValueAggregatorJob.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/aggregate/ValueAggregatorJob.java index d8833da..de25f64 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/aggregate/ValueAggregatorJob.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/aggregate/ValueAggregatorJob.java @@ -60,7 +60,7 @@ import org.apache.hadoop.util.GenericOptionsParser; * The developer using Aggregate will need only to provide a plugin class * conforming to the following interface: * - * public interface ValueAggregatorDescriptor { public ArrayList<Entry> + * public interface ValueAggregatorDescriptor { public ArrayList<Entry> * generateKeyValPairs(Object key, Object value); public void * configure(Configuration conf); } * http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/chain/Chain.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/chain/Chain.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/chain/Chain.java index 208616b..1dad13e 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/chain/Chain.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/chain/Chain.java @@ -600,7 +600,7 @@ public class Chain { /** * Adds a Mapper class to the chain job. * - * <p/> + * <p> * The configuration properties of the chain job have precedence over the * configuration properties of the Mapper. * @@ -738,7 +738,7 @@ public class Chain { /** * Sets the Reducer class to the chain job. * - * <p/> + * <p> * The configuration properties of the chain job have precedence over the * configuration properties of the Reducer. * http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/chain/ChainMapper.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/chain/ChainMapper.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/chain/ChainMapper.java index c042ff0..c3bf012 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/chain/ChainMapper.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/chain/ChainMapper.java @@ -57,24 +57,24 @@ import org.apache.hadoop.mapreduce.lib.chain.Chain.ChainBlockingQueue; * ChainMapper, this is done by the addMapper for the last mapper in the chain. * </p> * ChainMapper usage pattern: - * <p/> + * <p> * * <pre> * ... * Job = new Job(conf); - * <p/> + * * Configuration mapAConf = new Configuration(false); * ... * ChainMapper.addMapper(job, AMap.class, LongWritable.class, Text.class, * Text.class, Text.class, true, mapAConf); - * <p/> + * * Configuration mapBConf = new Configuration(false); * ... * ChainMapper.addMapper(job, BMap.class, Text.class, Text.class, * LongWritable.class, Text.class, false, mapBConf); - * <p/> + * * ... - * <p/> + * * job.waitForComplettion(true); * ... * </pre> http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/chain/ChainReducer.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/chain/ChainReducer.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/chain/ChainReducer.java index dc03d5d..1c37587 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/chain/ChainReducer.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/chain/ChainReducer.java @@ -50,7 +50,7 @@ import java.io.IOException; * all Mappers and the Reduce in the chain use matching output and input key and * value classes as no conversion is done by the chaining code. * </p> - * </p> Using the ChainMapper and the ChainReducer classes is possible to + * <p> Using the ChainMapper and the ChainReducer classes is possible to * compose Map/Reduce jobs that look like <code>[MAP+ / REDUCE MAP*]</code>. And * immediate benefit of this pattern is a dramatic reduction in disk IO. </p> * <p> @@ -59,26 +59,26 @@ import java.io.IOException; * element in the chain. * </p> * ChainReducer usage pattern: - * <p/> + * <p> * * <pre> * ... * Job = new Job(conf); * .... - * <p/> + * * Configuration reduceConf = new Configuration(false); * ... * ChainReducer.setReducer(job, XReduce.class, LongWritable.class, Text.class, * Text.class, Text.class, true, reduceConf); - * <p/> + * * ChainReducer.addMapper(job, CMap.class, Text.class, Text.class, * LongWritable.class, Text.class, false, null); - * <p/> + * * ChainReducer.addMapper(job, DMap.class, LongWritable.class, Text.class, * LongWritable.class, LongWritable.class, true, null); - * <p/> + * * ... - * <p/> + * * job.waitForCompletion(true); * ... * </pre> http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/db/DBInputFormat.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/db/DBInputFormat.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/db/DBInputFormat.java index a6953b7..78c3a0f 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/db/DBInputFormat.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/db/DBInputFormat.java @@ -319,7 +319,7 @@ public class DBInputFormat<T extends DBWritable> * Java object holding tuple fields. * @param tableName The table to read data from * @param conditions The condition which to select data with, - * eg. '(updated > 20070101 AND length > 0)' + * eg. '(updated > 20070101 AND length > 0)' * @param orderBy the fieldNames in the orderBy clause. * @param fieldNames The field names in the table * @see #setInput(Job, Class, String, String) http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/db/DBWritable.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/db/DBWritable.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/db/DBWritable.java index cc0d30a..5753a3b 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/db/DBWritable.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/db/DBWritable.java @@ -73,7 +73,7 @@ import org.apache.hadoop.io.Writable; * timestamp = resultSet.getLong(2); * } * } - * </pre></p> + * </pre> */ @InterfaceAudience.Public @InterfaceStability.Stable http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/join/TupleWritable.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/join/TupleWritable.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/join/TupleWritable.java index af6b3f0..2990ca9 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/join/TupleWritable.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/join/TupleWritable.java @@ -144,7 +144,7 @@ public class TupleWritable implements Writable, Iterable<Writable> { /** * Convert Tuple to String as in the following. - * <tt>[<child1>,<child2>,...,<childn>]</tt> + * <tt>[<child1>,<child2>,...,<childn>]</tt> */ public String toString() { StringBuffer buf = new StringBuffer("["); http://git-wip-us.apache.org/repos/asf/hadoop/blob/9ae7f9eb/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/map/MultithreadedMapper.java ---------------------------------------------------------------------- diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/map/MultithreadedMapper.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/map/MultithreadedMapper.java index 814e494..733b18c 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/map/MultithreadedMapper.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/map/MultithreadedMapper.java @@ -44,15 +44,15 @@ import java.util.List; * Multithreaded implementation for @link org.apache.hadoop.mapreduce.Mapper. * <p> * It can be used instead of the default implementation, - * @link org.apache.hadoop.mapred.MapRunner, when the Map operation is not CPU + * {@link org.apache.hadoop.mapred.MapRunner}, when the Map operation is not CPU * bound in order to improve throughput. * <p> * Mapper implementations using this MapRunnable must be thread-safe. * <p> * The Map-Reduce job has to be configured with the mapper to use via - * {@link #setMapperClass(Configuration, Class)} and + * {@link #setMapperClass(Job, Class)} and * the number of thread the thread-pool can use with the - * {@link #getNumberOfThreads(Configuration) method. The default + * {@link #getNumberOfThreads(JobContext)} method. The default * value is 10 threads. * <p> */