http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-java/src/main/java/org/apache/flink/api/java/operators/join/JoinOperatorSetsBase.java ---------------------------------------------------------------------- diff --git a/flink-java/src/main/java/org/apache/flink/api/java/operators/join/JoinOperatorSetsBase.java b/flink-java/src/main/java/org/apache/flink/api/java/operators/join/JoinOperatorSetsBase.java index 6f00cdb..e14e06d 100644 --- a/flink-java/src/main/java/org/apache/flink/api/java/operators/join/JoinOperatorSetsBase.java +++ b/flink-java/src/main/java/org/apache/flink/api/java/operators/join/JoinOperatorSetsBase.java @@ -33,7 +33,7 @@ import org.apache.flink.api.java.tuple.Tuple; import org.apache.flink.api.java.typeutils.TypeExtractor; /** - * Intermediate step of an Outer Join transformation. <br/> + * Intermediate step of an Outer Join transformation. <br> * To continue the Join transformation, select the join key of the first input {@link DataSet} by calling * {@link JoinOperatorSetsBase#where(int...)} or * {@link JoinOperatorSetsBase#where(KeySelector)}. @@ -69,9 +69,9 @@ public class JoinOperatorSetsBase<I1, I2> { } /** - * Continues a Join transformation. <br/> - * Defines the {@link Tuple} fields of the first join {@link DataSet} that should be used as join keys.<br/> - * <b>Note: Fields can only be selected as join keys on Tuple DataSets.</b><br/> + * Continues a Join transformation. <br> + * Defines the {@link Tuple} fields of the first join {@link DataSet} that should be used as join keys.<br> + * <b>Note: Fields can only be selected as join keys on Tuple DataSets.</b><br> * * @param fields The indexes of the other Tuple fields of the first join DataSets that should be used as keys. * @return An incomplete Join transformation. @@ -87,7 +87,7 @@ public class JoinOperatorSetsBase<I1, I2> { } /** - * Continues a Join transformation. <br/> + * Continues a Join transformation. <br> * Defines the fields of the first join {@link DataSet} that should be used as grouping keys. Fields * are the names of member fields of the underlying type of the data set. * @@ -105,9 +105,9 @@ public class JoinOperatorSetsBase<I1, I2> { } /** - * Continues a Join transformation and defines a {@link KeySelector} function for the first join {@link DataSet}.</br> + * Continues a Join transformation and defines a {@link KeySelector} function for the first join {@link DataSet}.<br> * The KeySelector function is called for each element of the first DataSet and extracts a single - * key value on which the DataSet is joined. </br> + * key value on which the DataSet is joined. <br> * * @param keySelector The KeySelector function which extracts the key values from the DataSet on which it is joined. * @return An incomplete Join transformation. @@ -125,7 +125,7 @@ public class JoinOperatorSetsBase<I1, I2> { /** - * Intermediate step of a Join transformation. <br/> + * Intermediate step of a Join transformation. <br> * To continue the Join transformation, select the join key of the second input {@link DataSet} by calling * {@link org.apache.flink.api.java.operators.join.JoinOperatorSetsBase.JoinOperatorSetsPredicateBase#equalTo(int...)} or * {@link org.apache.flink.api.java.operators.join.JoinOperatorSetsBase.JoinOperatorSetsPredicateBase#equalTo(KeySelector)}. @@ -149,8 +149,8 @@ public class JoinOperatorSetsBase<I1, I2> { /** * Continues a Join transformation and defines the {@link Tuple} fields of the second join - * {@link DataSet} that should be used as join keys.<br/> - * <b>Note: Fields can only be selected as join keys on Tuple DataSets.</b><br/> + * {@link DataSet} that should be used as join keys.<br> + * <b>Note: Fields can only be selected as join keys on Tuple DataSets.</b><br> * * The resulting {@link JoinFunctionAssigner} needs to be finished by providing a * {@link JoinFunction} by calling {@link JoinFunctionAssigner#with(JoinFunction)} @@ -164,7 +164,7 @@ public class JoinOperatorSetsBase<I1, I2> { /** * Continues a Join transformation and defines the fields of the second join - * {@link DataSet} that should be used as join keys.<br/> + * {@link DataSet} that should be used as join keys.<br> * * The resulting {@link JoinFunctionAssigner} needs to be finished by providing a * {@link JoinFunction} by calling {@link JoinFunctionAssigner#with(JoinFunction)} @@ -177,9 +177,9 @@ public class JoinOperatorSetsBase<I1, I2> { } /** - * Continues a Join transformation and defines a {@link KeySelector} function for the second join {@link DataSet}.</br> + * Continues a Join transformation and defines a {@link KeySelector} function for the second join {@link DataSet}.<br> * The KeySelector function is called for each element of the second DataSet and extracts a single - * key value on which the DataSet is joined. </br> + * key value on which the DataSet is joined. <br> * * The resulting {@link JoinFunctionAssigner} needs to be finished by providing a * {@link JoinFunction} by calling {@link JoinFunctionAssigner#with(JoinFunction)}
http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-java/src/main/java/org/apache/flink/api/java/operators/translation/Tuple3WrappingCollector.java ---------------------------------------------------------------------- diff --git a/flink-java/src/main/java/org/apache/flink/api/java/operators/translation/Tuple3WrappingCollector.java b/flink-java/src/main/java/org/apache/flink/api/java/operators/translation/Tuple3WrappingCollector.java index 66592f5..29d68f1 100644 --- a/flink-java/src/main/java/org/apache/flink/api/java/operators/translation/Tuple3WrappingCollector.java +++ b/flink-java/src/main/java/org/apache/flink/api/java/operators/translation/Tuple3WrappingCollector.java @@ -22,7 +22,7 @@ import org.apache.flink.api.java.tuple.Tuple3; import org.apache.flink.util.Collector; /** - * Needed to wrap tuples to Tuple3<groupKey, sortKey, value> for combine method of group reduce with key selector sorting + * Needed to wrap tuples to {@code Tuple3<groupKey, sortKey, value>} for combine method of group reduce with key selector sorting */ public class Tuple3WrappingCollector<IN, K1, K2> implements Collector<IN>, java.io.Serializable { http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-java/src/main/java/org/apache/flink/api/java/operators/translation/TupleWrappingCollector.java ---------------------------------------------------------------------- diff --git a/flink-java/src/main/java/org/apache/flink/api/java/operators/translation/TupleWrappingCollector.java b/flink-java/src/main/java/org/apache/flink/api/java/operators/translation/TupleWrappingCollector.java index 43f667f..8f0b4fa 100644 --- a/flink-java/src/main/java/org/apache/flink/api/java/operators/translation/TupleWrappingCollector.java +++ b/flink-java/src/main/java/org/apache/flink/api/java/operators/translation/TupleWrappingCollector.java @@ -22,7 +22,7 @@ import org.apache.flink.api.java.tuple.Tuple2; import org.apache.flink.util.Collector; /** - * Needed to wrap tuples to Tuple2<key, value> pairs for combine method of group reduce with key selector function + * Needed to wrap tuples to {@code Tuple2<key, value>} pairs for combine method of group reduce with key selector function */ public class TupleWrappingCollector<IN, K> implements Collector<IN>, java.io.Serializable { http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-java/src/main/java/org/apache/flink/api/java/record/functions/FunctionAnnotation.java ---------------------------------------------------------------------- diff --git a/flink-java/src/main/java/org/apache/flink/api/java/record/functions/FunctionAnnotation.java b/flink-java/src/main/java/org/apache/flink/api/java/record/functions/FunctionAnnotation.java index d7a17b8..71d2a62 100644 --- a/flink-java/src/main/java/org/apache/flink/api/java/record/functions/FunctionAnnotation.java +++ b/flink-java/src/main/java/org/apache/flink/api/java/record/functions/FunctionAnnotation.java @@ -44,8 +44,8 @@ import org.apache.flink.api.common.operators.util.UserCodeWrapper; * annotation for a map-type function that realizes a simple absolute function, * use it the following way: * - * <pre><blockquote> - * \@ConstantFieldsExcept(fields={2}) + * <pre>{@code + * {@literal @}ConstantFieldsExcept(fields={2}) * public class MyMapper extends MapFunction * { * public void map(Record record, Collector out) @@ -56,7 +56,7 @@ import org.apache.flink.api.common.operators.util.UserCodeWrapper; out.collect(record); * } * } - * </blockquote></pre> + * }</pre> * * Be aware that some annotations should only be used for functions with as single input * ({@link MapFunction}, {@link ReduceFunction}) and some only for stubs with two inputs http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-java/src/main/java/org/apache/flink/api/java/record/io/ExternalProcessInputFormat.java ---------------------------------------------------------------------- diff --git a/flink-java/src/main/java/org/apache/flink/api/java/record/io/ExternalProcessInputFormat.java b/flink-java/src/main/java/org/apache/flink/api/java/record/io/ExternalProcessInputFormat.java index b1f7ffc..cbf16b5 100644 --- a/flink-java/src/main/java/org/apache/flink/api/java/record/io/ExternalProcessInputFormat.java +++ b/flink-java/src/main/java/org/apache/flink/api/java/record/io/ExternalProcessInputFormat.java @@ -32,7 +32,7 @@ import org.apache.flink.core.io.GenericInputSplit; * The input format checks the exit code of the process to validate whether the process terminated correctly. A list of allowed exit codes can be provided. * The input format requires ({@link ExternalProcessInputSplit} objects that hold the command to execute. * - * <b>Attention! </b><br/> + * <b>Attention! </b><br> * You must take care to read from (and process) both output streams of the process, standard out (stdout) and standard error (stderr). * Otherwise, the input format might get deadlocked! * http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-java/src/main/java/org/apache/flink/api/java/typeutils/AvroTypeInfo.java ---------------------------------------------------------------------- diff --git a/flink-java/src/main/java/org/apache/flink/api/java/typeutils/AvroTypeInfo.java b/flink-java/src/main/java/org/apache/flink/api/java/typeutils/AvroTypeInfo.java index 59c7eb0..5f74513 100644 --- a/flink-java/src/main/java/org/apache/flink/api/java/typeutils/AvroTypeInfo.java +++ b/flink-java/src/main/java/org/apache/flink/api/java/typeutils/AvroTypeInfo.java @@ -29,8 +29,8 @@ import java.util.List; /** * Special type information to generate a special AvroTypeInfo for Avro POJOs (implementing SpecificRecordBase, the typed Avro POJOs) * - * Proceeding: It uses a regular pojo type analysis and replaces all GenericType<CharSequence> - * with a GenericType<avro.Utf8>. + * Proceeding: It uses a regular pojo type analysis and replaces all {@code GenericType<CharSequence>} + * with a {@code GenericType<avro.Utf8>}. * All other types used by Avro are standard Java types. * Only strings are represented as CharSequence fields and represented as Utf8 classes at runtime. * CharSequence is not comparable. To make them nicely usable with field expressions, we replace them here http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-java/src/main/java/org/apache/flink/api/java/typeutils/runtime/kryo/Serializers.java ---------------------------------------------------------------------- diff --git a/flink-java/src/main/java/org/apache/flink/api/java/typeutils/runtime/kryo/Serializers.java b/flink-java/src/main/java/org/apache/flink/api/java/typeutils/runtime/kryo/Serializers.java index 9e0da25..76f8eb4 100644 --- a/flink-java/src/main/java/org/apache/flink/api/java/typeutils/runtime/kryo/Serializers.java +++ b/flink-java/src/main/java/org/apache/flink/api/java/typeutils/runtime/kryo/Serializers.java @@ -107,8 +107,8 @@ public class Serializers { } /** - * Register these serializers for using Avro's {@see GenericData.Record} and classes - * implementing {@see org.apache.avro.specific.SpecificRecordBase} + * Register these serializers for using Avro's {@link GenericData.Record} and classes + * implementing {@link org.apache.avro.specific.SpecificRecordBase} */ public static void registerGenericAvro(ExecutionConfig reg) { // Avro POJOs contain java.util.List which have GenericData.Array as their runtime type http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-java/src/main/java/org/apache/flink/api/java/utils/DataSetUtils.java ---------------------------------------------------------------------- diff --git a/flink-java/src/main/java/org/apache/flink/api/java/utils/DataSetUtils.java b/flink-java/src/main/java/org/apache/flink/api/java/utils/DataSetUtils.java index 8f9fa3f..b91dc82 100644 --- a/flink-java/src/main/java/org/apache/flink/api/java/utils/DataSetUtils.java +++ b/flink-java/src/main/java/org/apache/flink/api/java/utils/DataSetUtils.java @@ -206,7 +206,7 @@ public final class DataSetUtils { * <p> * <strong>NOTE:</strong> Sample with fixed size is not as efficient as sample with fraction, use sample with * fraction unless you need exact precision. - * <p/> + * </p> * * @param withReplacement Whether element can be selected more than once. * @param numSample The expected sample size. @@ -225,7 +225,7 @@ public final class DataSetUtils { * <p> * <strong>NOTE:</strong> Sample with fixed size is not as efficient as sample with fraction, use sample with * fraction unless you need exact precision. - * <p/> + * </p> * * @param withReplacement Whether element can be selected more than once. * @param numSample The expected sample size. http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-java8/src/main/java/org/apache/flink/examples/java8/relational/TPCHQuery10.java ---------------------------------------------------------------------- diff --git a/flink-java8/src/main/java/org/apache/flink/examples/java8/relational/TPCHQuery10.java b/flink-java8/src/main/java/org/apache/flink/examples/java8/relational/TPCHQuery10.java index 4dae630..1ad4f41 100644 --- a/flink-java8/src/main/java/org/apache/flink/examples/java8/relational/TPCHQuery10.java +++ b/flink-java8/src/main/java/org/apache/flink/examples/java8/relational/TPCHQuery10.java @@ -35,7 +35,7 @@ import org.apache.flink.api.java.tuple.Tuple6; * This program implements the following SQL equivalent: * * <p> - * <code><pre> + * <pre>{@code * SELECT * c_custkey, * c_name, @@ -60,7 +60,7 @@ import org.apache.flink.api.java.tuple.Tuple6; * c_acctbal, * n_name, * c_address - * </pre></code> + * }</pre> * * <p> * Compared to the original TPC-H query this version does not print http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/Graph.java ---------------------------------------------------------------------- diff --git a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/Graph.java b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/Graph.java index 6015be4..bab2911 100755 --- a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/Graph.java +++ b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/Graph.java @@ -352,10 +352,10 @@ public class Graph<K, VV, EV> { * @return An instance of {@link org.apache.flink.graph.GraphCsvReader}, * on which calling methods to specify types of the Vertex ID, Vertex value and Edge value returns a Graph. * - * @see {@link org.apache.flink.graph.GraphCsvReader#types(Class, Class, Class)}, - * {@link org.apache.flink.graph.GraphCsvReader#vertexTypes(Class, Class)}, - * {@link org.apache.flink.graph.GraphCsvReader#edgeTypes(Class, Class)} and - * {@link org.apache.flink.graph.GraphCsvReader#keyType(Class)}. + * @see org.apache.flink.graph.GraphCsvReader#types(Class, Class, Class) + * @see org.apache.flink.graph.GraphCsvReader#vertexTypes(Class, Class) + * @see org.apache.flink.graph.GraphCsvReader#edgeTypes(Class, Class) + * @see org.apache.flink.graph.GraphCsvReader#keyType(Class) */ public static GraphCsvReader fromCsvReader(String verticesPath, String edgesPath, ExecutionEnvironment context) { return new GraphCsvReader(verticesPath, edgesPath, context); @@ -369,10 +369,10 @@ public class Graph<K, VV, EV> { * @return An instance of {@link org.apache.flink.graph.GraphCsvReader}, * on which calling methods to specify types of the Vertex ID, Vertex value and Edge value returns a Graph. * - * @see {@link org.apache.flink.graph.GraphCsvReader#types(Class, Class, Class)}, - * {@link org.apache.flink.graph.GraphCsvReader#vertexTypes(Class, Class)}, - * {@link org.apache.flink.graph.GraphCsvReader#edgeTypes(Class, Class)} and - * {@link org.apache.flink.graph.GraphCsvReader#keyType(Class)}. + * @see org.apache.flink.graph.GraphCsvReader#types(Class, Class, Class) + * @see org.apache.flink.graph.GraphCsvReader#vertexTypes(Class, Class) + * @see org.apache.flink.graph.GraphCsvReader#edgeTypes(Class, Class) + * @see org.apache.flink.graph.GraphCsvReader#keyType(Class) */ public static GraphCsvReader fromCsvReader(String edgesPath, ExecutionEnvironment context) { return new GraphCsvReader(edgesPath, context); @@ -389,10 +389,10 @@ public class Graph<K, VV, EV> { * @return An instance of {@link org.apache.flink.graph.GraphCsvReader}, * on which calling methods to specify types of the Vertex ID, Vertex Value and Edge value returns a Graph. * - * @see {@link org.apache.flink.graph.GraphCsvReader#types(Class, Class, Class)}, - * {@link org.apache.flink.graph.GraphCsvReader#vertexTypes(Class, Class)}, - * {@link org.apache.flink.graph.GraphCsvReader#edgeTypes(Class, Class)} and - * {@link org.apache.flink.graph.GraphCsvReader#keyType(Class)}. + * @see org.apache.flink.graph.GraphCsvReader#types(Class, Class, Class) + * @see org.apache.flink.graph.GraphCsvReader#vertexTypes(Class, Class) + * @see org.apache.flink.graph.GraphCsvReader#edgeTypes(Class, Class) + * @see org.apache.flink.graph.GraphCsvReader#keyType(Class) */ public static <K, VV> GraphCsvReader fromCsvReader(String edgesPath, final MapFunction<K, VV> vertexValueInitializer, ExecutionEnvironment context) { @@ -821,7 +821,7 @@ public class Graph<K, VV, EV> { /** * Return the out-degree of all vertices in the graph * - * @return A DataSet of Tuple2<vertexId, outDegree> + * @return A DataSet of {@code Tuple2<vertexId, outDegree>} */ public DataSet<Tuple2<K, Long>> outDegrees() { @@ -851,7 +851,7 @@ public class Graph<K, VV, EV> { /** * Return the in-degree of all vertices in the graph * - * @return A DataSet of Tuple2<vertexId, inDegree> + * @return A DataSet of {@code Tuple2<vertexId, inDegree>} */ public DataSet<Tuple2<K, Long>> inDegrees() { @@ -861,7 +861,7 @@ public class Graph<K, VV, EV> { /** * Return the degree of all vertices in the graph * - * @return A DataSet of Tuple2<vertexId, degree> + * @return A DataSet of {@code Tuple2<vertexId, degree>} */ public DataSet<Tuple2<K, Long>> getDegrees() { return outDegrees().union(inDegrees()).groupBy(0).sum(1); http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/example/EuclideanGraphWeighing.java ---------------------------------------------------------------------- diff --git a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/example/EuclideanGraphWeighing.java b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/example/EuclideanGraphWeighing.java index 2a10bd1..712be3e 100644 --- a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/example/EuclideanGraphWeighing.java +++ b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/example/EuclideanGraphWeighing.java @@ -35,7 +35,7 @@ import java.io.Serializable; /** * This example shows how to use Gelly's {@link Graph#getTriplets()} and - * {@link Graph#joinWithEdges(DataSet, MapFunction)} methods. + * {@link Graph#joinWithEdges(DataSet, EdgeJoinFunction)} methods. * * Given a directed, unweighted graph, with vertex values representing points in a plan, * return a weighted graph where the edge weights are equal to the Euclidean distance between the @@ -51,7 +51,6 @@ import java.io.Serializable; * Edges themselves are separated by newlines. * For example: <code>1,2\n1,3\n</code> defines two edges 1-2 and 1-3. * </ul> - * </p> * * Usage <code>EuclideanGraphWeighing <vertex path> <edge path> <result path></code><br> * If no parameters are provided, the program is run with default data from http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/example/GraphMetrics.java ---------------------------------------------------------------------- diff --git a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/example/GraphMetrics.java b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/example/GraphMetrics.java index b808e76..117f7d1 100644 --- a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/example/GraphMetrics.java +++ b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/example/GraphMetrics.java @@ -42,7 +42,7 @@ import org.apache.flink.types.NullValue; * * The input file is expected to contain one edge per line, * with long IDs and no values, in the following format: - * "<sourceVertexID>\t<targetVertexID>". + * "<sourceVertexID>\t<targetVertexID>". * If no arguments are provided, the example runs with a random graph of 100 vertices. * */ http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/example/IncrementalSSSP.java ---------------------------------------------------------------------- diff --git a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/example/IncrementalSSSP.java b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/example/IncrementalSSSP.java index c03937d..947f343 100644 --- a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/example/IncrementalSSSP.java +++ b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/example/IncrementalSSSP.java @@ -52,7 +52,7 @@ import org.apache.flink.graph.spargel.VertexUpdateFunction; * The edge is simply removed from the graph. * - If the removed edge is an SP-edge, then all nodes, whose shortest path contains the removed edge, * potentially require re-computation. - * When the edge <u, v> is removed, v checks if it has another out-going SP-edge. + * When the edge {@code <u, v>} is removed, v checks if it has another out-going SP-edge. * If yes, no further computation is required. * If v has no other out-going SP-edge, it invalidates its current value, by setting it to INF. * Then, it informs all its SP-in-neighbors by sending them an INVALIDATE message. http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/example/MusicProfiles.java ---------------------------------------------------------------------- diff --git a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/example/MusicProfiles.java b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/example/MusicProfiles.java index 297dce2..24244c8 100644 --- a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/example/MusicProfiles.java +++ b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/example/MusicProfiles.java @@ -45,10 +45,10 @@ import org.apache.flink.util.Collector; /** * This example demonstrates how to mix the DataSet Flink API with the Gelly API. - * The input is a set <userId - songId - playCount> triplets and + * The input is a set <userId - songId - playCount> triplets and * a set of bad records, i.e. song ids that should not be trusted. * Initially, we use the DataSet API to filter out the bad records. - * Then, we use Gelly to create a user -> song weighted bipartite graph and compute + * Then, we use Gelly to create a user -> song weighted bipartite graph and compute * the top song (most listened) per user. * Then, we use the DataSet API again, to create a user-user similarity graph, * based on common songs, where users that are listeners of the same song @@ -58,11 +58,11 @@ import org.apache.flink.util.Collector; * the similarity graph. * * The triplets input is expected to be given as one triplet per line, - * in the following format: "<userID>\t<songID>\t<playcount>". + * in the following format: "<userID>\t<songID>\t<playcount>". * * The mismatches input file is expected to contain one mismatch record per line, * in the following format: - * "ERROR: <songID trackID> song_title" + * "ERROR: <songID trackID> song_title" * * If no arguments are provided, the example runs with default data from {@link MusicProfilesData}. */ http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/gsa/ApplyFunction.java ---------------------------------------------------------------------- diff --git a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/gsa/ApplyFunction.java b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/gsa/ApplyFunction.java index 5a8e97a..e9add7c 100755 --- a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/gsa/ApplyFunction.java +++ b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/gsa/ApplyFunction.java @@ -45,7 +45,7 @@ public abstract class ApplyFunction<K, VV, M> implements Serializable { /** * Retrieves the number of vertices in the graph. - * @return the number of vertices if the {@link IterationConfigurationion#setOptNumVertices(boolean)} + * @return the number of vertices if the {@link org.apache.flink.graph.IterationConfiguration#setOptNumVertices(boolean)} * option has been set; -1 otherwise. */ public long getNumberOfVertices() { @@ -80,15 +80,11 @@ public abstract class ApplyFunction<K, VV, M> implements Serializable { /** * This method is executed once per superstep before the vertex update function is invoked for each vertex. - * - * @throws Exception Exceptions in the pre-superstep phase cause the superstep to fail. */ public void preSuperstep() {} /** * This method is executed once per superstep after the vertex update function has been invoked for each vertex. - * - * @throws Exception Exceptions in the post-superstep phase cause the superstep to fail. */ public void postSuperstep() {} http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/gsa/GatherFunction.java ---------------------------------------------------------------------- diff --git a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/gsa/GatherFunction.java b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/gsa/GatherFunction.java index 563b20e..d914f2a 100755 --- a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/gsa/GatherFunction.java +++ b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/gsa/GatherFunction.java @@ -43,7 +43,7 @@ public abstract class GatherFunction<VV, EV, M> implements Serializable { /** * Retrieves the number of vertices in the graph. - * @return the number of vertices if the {@link IterationConfigurationion#setOptNumVertices(boolean)} + * @return the number of vertices if the {@link org.apache.flink.graph.IterationConfiguration#setOptNumVertices(boolean)} * option has been set; -1 otherwise. */ public long getNumberOfVertices() { @@ -69,15 +69,11 @@ public abstract class GatherFunction<VV, EV, M> implements Serializable { /** * This method is executed once per superstep before the vertex update function is invoked for each vertex. - * - * @throws Exception Exceptions in the pre-superstep phase cause the superstep to fail. */ public void preSuperstep() {} /** * This method is executed once per superstep after the vertex update function has been invoked for each vertex. - * - * @throws Exception Exceptions in the post-superstep phase cause the superstep to fail. */ public void postSuperstep() {} http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/gsa/Neighbor.java ---------------------------------------------------------------------- diff --git a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/gsa/Neighbor.java b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/gsa/Neighbor.java index 7fa1ed2..4c970fb 100755 --- a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/gsa/Neighbor.java +++ b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/gsa/Neighbor.java @@ -21,8 +21,8 @@ package org.apache.flink.graph.gsa; import org.apache.flink.api.java.tuple.Tuple2; /** - * This class represents a <sourceVertex, edge> pair - * This is a wrapper around Tuple2<VV, EV> for convenience in the GatherFunction + * This class represents a {@code <sourceVertex, edge>} pair + * This is a wrapper around {@code Tuple2<VV, EV>} for convenience in the GatherFunction * @param <VV> the vertex value type * @param <EV> the edge value type */ http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/gsa/SumFunction.java ---------------------------------------------------------------------- diff --git a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/gsa/SumFunction.java b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/gsa/SumFunction.java index f27e275..68e8d27 100755 --- a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/gsa/SumFunction.java +++ b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/gsa/SumFunction.java @@ -43,7 +43,7 @@ public abstract class SumFunction<VV, EV, M> implements Serializable { /** * Retrieves the number of vertices in the graph. - * @return the number of vertices if the {@link IterationConfigurationion#setOptNumVertices(boolean)} + * @return the number of vertices if the {@link org.apache.flink.graph.IterationConfiguration#setOptNumVertices(boolean)} * option has been set; -1 otherwise. */ public long getNumberOfVertices() { @@ -69,15 +69,11 @@ public abstract class SumFunction<VV, EV, M> implements Serializable { /** * This method is executed once per superstep before the vertex update function is invoked for each vertex. - * - * @throws Exception Exceptions in the pre-superstep phase cause the superstep to fail. */ public void preSuperstep() {} /** * This method is executed once per superstep after the vertex update function has been invoked for each vertex. - * - * @throws Exception Exceptions in the post-superstep phase cause the superstep to fail. */ public void postSuperstep() {} http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/library/ConnectedComponents.java ---------------------------------------------------------------------- diff --git a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/library/ConnectedComponents.java b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/library/ConnectedComponents.java index 745c103..a190cc5 100644 --- a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/library/ConnectedComponents.java +++ b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/library/ConnectedComponents.java @@ -41,7 +41,7 @@ import org.apache.flink.types.NullValue; * * The result is a DataSet of vertices, where the vertex value corresponds to the assigned component ID. * - * @see {@link org.apache.flink.graph.library.GSAConnectedComponents} + * @see org.apache.flink.graph.library.GSAConnectedComponents */ @SuppressWarnings("serial") public class ConnectedComponents<K, EV> implements GraphAlgorithm<K, Long, EV, DataSet<Vertex<K, Long>>> { http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/library/GSAConnectedComponents.java ---------------------------------------------------------------------- diff --git a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/library/GSAConnectedComponents.java b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/library/GSAConnectedComponents.java index 4269517..a44ba14 100755 --- a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/library/GSAConnectedComponents.java +++ b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/library/GSAConnectedComponents.java @@ -34,7 +34,7 @@ import org.apache.flink.types.NullValue; * This implementation assumes that the vertices of the input Graph are initialized with unique, Long component IDs. * The result is a DataSet of vertices, where the vertex value corresponds to the assigned component ID. * - * @see {@link org.apache.flink.graph.library.ConnectedComponents} + * @see org.apache.flink.graph.library.ConnectedComponents */ public class GSAConnectedComponents<K, EV> implements GraphAlgorithm<K, Long, EV, DataSet<Vertex<K, Long>>> { http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/library/GSAPageRank.java ---------------------------------------------------------------------- diff --git a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/library/GSAPageRank.java b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/library/GSAPageRank.java index 43a5e5c..99624ca 100644 --- a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/library/GSAPageRank.java +++ b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/library/GSAPageRank.java @@ -61,7 +61,7 @@ public class GSAPageRank<K> implements GraphAlgorithm<K, Double, Double, DataSet /** * Creates an instance of the GSA PageRank algorithm. * If the number of vertices of the input graph is known, - * use the {@link GSAPageRank#GSAPageRank(double, long)} constructor instead. + * use the {@link GSAPageRank#GSAPageRank(double, int)} constructor instead. * * The implementation assumes that each page has at least one incoming and one outgoing link. * http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/library/PageRank.java ---------------------------------------------------------------------- diff --git a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/library/PageRank.java b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/library/PageRank.java index dfbe14b..935058d 100644 --- a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/library/PageRank.java +++ b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/library/PageRank.java @@ -62,7 +62,7 @@ public class PageRank<K> implements GraphAlgorithm<K, Double, Double, DataSet<Ve /** * Creates an instance of the PageRank algorithm. * If the number of vertices of the input graph is known, - * use the {@link PageRank#PageRank(double, long)} constructor instead. + * use the {@link PageRank#PageRank(double, int)} constructor instead. * * The implementation assumes that each page has at least one incoming and one outgoing link. * http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/spargel/MessagingFunction.java ---------------------------------------------------------------------- diff --git a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/spargel/MessagingFunction.java b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/spargel/MessagingFunction.java index 271db86..f29fc9d 100644 --- a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/spargel/MessagingFunction.java +++ b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/spargel/MessagingFunction.java @@ -53,7 +53,7 @@ public abstract class MessagingFunction<K, VV, Message, EV> implements Serializa /** * Retrieves the number of vertices in the graph. - * @return the number of vertices if the {@link IterationConfiguration#setOptNumVertices(boolean)} + * @return the number of vertices if the {@link org.apache.flink.graph.IterationConfiguration#setOptNumVertices(boolean)} * option has been set; -1 otherwise. */ public long getNumberOfVertices() { @@ -314,8 +314,7 @@ public abstract class MessagingFunction<K, VV, Message, EV> implements Serializa /** * Retrieves the vertex in-degree (number of in-coming edges). - * @return The in-degree of this vertex if the {@link IterationConfiguration#setOptDegrees(boolean)} - * option has been set; -1 otherwise. + * @return The in-degree of this vertex */ public long getInDegree() { return inDegree; @@ -327,8 +326,7 @@ public abstract class MessagingFunction<K, VV, Message, EV> implements Serializa /** * Retrieve the vertex out-degree (number of out-going edges). - * @return The out-degree of this vertex if the {@link IterationConfiguration#setOptDegrees(boolean)} - * option has been set; -1 otherwise. + * @return The out-degree of this vertex */ public long getOutDegree() { return outDegree; http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/spargel/VertexCentricIteration.java ---------------------------------------------------------------------- diff --git a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/spargel/VertexCentricIteration.java b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/spargel/VertexCentricIteration.java index fdc39ff..be48f9c 100644 --- a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/spargel/VertexCentricIteration.java +++ b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/spargel/VertexCentricIteration.java @@ -46,7 +46,7 @@ import com.google.common.base.Preconditions; /** * This class represents iterative graph computations, programmed in a vertex-centric perspective. - * It is a special case of <i>Bulk Synchronous Parallel<i> computation. The paradigm has also been + * It is a special case of <i>Bulk Synchronous Parallel</i> computation. The paradigm has also been * implemented by Google's <i>Pregel</i> system and by <i>Apache Giraph</i>. * <p> * Vertex centric algorithms operate on graphs, which are defined through vertices and edges. The http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/spargel/VertexUpdateFunction.java ---------------------------------------------------------------------- diff --git a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/spargel/VertexUpdateFunction.java b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/spargel/VertexUpdateFunction.java index 248925b..857cef5 100644 --- a/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/spargel/VertexUpdateFunction.java +++ b/flink-libraries/flink-gelly/src/main/java/org/apache/flink/graph/spargel/VertexUpdateFunction.java @@ -30,12 +30,12 @@ import org.apache.flink.util.Collector; /** * This class must be extended by functions that compute the state of the vertex depending on the old state and the - * incoming messages. The central method is {@link #updateVertex(Comparable, Object, MessageIterator)}, which is + * incoming messages. The central method is {@link #updateVertex(Vertex, MessageIterator)}, which is * invoked once per vertex per superstep. * - * <K> The vertex key type. - * <VV> The vertex value type. - * <Message> The message type. + * {@code <K>} The vertex key type. + * {@code <VV>} The vertex value type. + * {@code <Message>} The message type. */ public abstract class VertexUpdateFunction<K, VV, Message> implements Serializable { @@ -50,7 +50,7 @@ public abstract class VertexUpdateFunction<K, VV, Message> implements Serializab /** * Retrieves the number of vertices in the graph. - * @return the number of vertices if the {@link IterationConfiguration#setOptNumVertices(boolean)} + * @return the number of vertices if the {@link org.apache.flink.graph.IterationConfiguration#setOptNumVertices(boolean)} * option has been set; -1 otherwise. */ public long getNumberOfVertices() { @@ -206,8 +206,7 @@ public abstract class VertexUpdateFunction<K, VV, Message> implements Serializab /** * Retrieves the vertex in-degree (number of in-coming edges). - * @return The in-degree of this vertex if the {@link IterationConfiguration#setOptDegrees(boolean)} - * option has been set; -1 otherwise. + * @return The in-degree of this vertex */ public long getInDegree() { return inDegree; @@ -219,8 +218,7 @@ public abstract class VertexUpdateFunction<K, VV, Message> implements Serializab /** * Retrieve the vertex out-degree (number of out-going edges). - * @return The out-degree of this vertex if the {@link IterationConfiguration#setOptDegrees(boolean)} - * option has been set; -1 otherwise. + * @return The out-degree of this vertex */ public long getOutDegree() { return outDegree; http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-optimizer/src/main/java/org/apache/flink/optimizer/dag/OptimizerNode.java ---------------------------------------------------------------------- diff --git a/flink-optimizer/src/main/java/org/apache/flink/optimizer/dag/OptimizerNode.java b/flink-optimizer/src/main/java/org/apache/flink/optimizer/dag/OptimizerNode.java index 490c304..a3bf995 100644 --- a/flink-optimizer/src/main/java/org/apache/flink/optimizer/dag/OptimizerNode.java +++ b/flink-optimizer/src/main/java/org/apache/flink/optimizer/dag/OptimizerNode.java @@ -50,7 +50,7 @@ import org.apache.flink.util.Visitor; * The OptimizerNode is the base class of all nodes in the optimizer DAG. The optimizer DAG is the * optimizer's representation of a program, created before the actual optimization (which creates different * candidate plans and computes their cost). - * <p>> + * <p> * Nodes in the DAG correspond (almost) one-to-one to the operators in a program. The optimizer DAG is constructed * to hold the additional information that the optimizer needs: * <ul> http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-optimizer/src/main/java/org/apache/flink/optimizer/operators/OperatorDescriptorSingle.java ---------------------------------------------------------------------- diff --git a/flink-optimizer/src/main/java/org/apache/flink/optimizer/operators/OperatorDescriptorSingle.java b/flink-optimizer/src/main/java/org/apache/flink/optimizer/operators/OperatorDescriptorSingle.java index c8be5d4..ffa317c 100644 --- a/flink-optimizer/src/main/java/org/apache/flink/optimizer/operators/OperatorDescriptorSingle.java +++ b/flink-optimizer/src/main/java/org/apache/flink/optimizer/operators/OperatorDescriptorSingle.java @@ -35,7 +35,7 @@ import org.apache.flink.optimizer.plan.SingleInputPlanNode; * Abstract base class for Operator descriptions which instantiates the node and sets the driver * strategy and the sorting and grouping keys. Returns possible local and global properties and * updates them after the operation has been performed. - * @see org.apache.flink.compiler.dag.SingleInputNode + * @see org.apache.flink.optimizer.dag.SingleInputNode */ public abstract class OperatorDescriptorSingle implements AbstractOperatorDescriptor { http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-optimizer/src/main/java/org/apache/flink/optimizer/plan/Channel.java ---------------------------------------------------------------------- diff --git a/flink-optimizer/src/main/java/org/apache/flink/optimizer/plan/Channel.java b/flink-optimizer/src/main/java/org/apache/flink/optimizer/plan/Channel.java index b139b62..a4448c6 100644 --- a/flink-optimizer/src/main/java/org/apache/flink/optimizer/plan/Channel.java +++ b/flink-optimizer/src/main/java/org/apache/flink/optimizer/plan/Channel.java @@ -186,7 +186,6 @@ public class Channel implements EstimateProvider, Cloneable, DumpableConnection< * Sets the data exchange mode (batch / pipelined) to use for the data * exchange of this channel. * - * @return The data exchange mode of this channel. */ public void setDataExchangeMode(DataExchangeMode dataExchangeMode) { this.dataExchangeMode = checkNotNull(dataExchangeMode); http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-optimizer/src/main/java/org/apache/flink/optimizer/postpass/SparseKeySchema.java ---------------------------------------------------------------------- diff --git a/flink-optimizer/src/main/java/org/apache/flink/optimizer/postpass/SparseKeySchema.java b/flink-optimizer/src/main/java/org/apache/flink/optimizer/postpass/SparseKeySchema.java index e14888e..1545d6f 100644 --- a/flink-optimizer/src/main/java/org/apache/flink/optimizer/postpass/SparseKeySchema.java +++ b/flink-optimizer/src/main/java/org/apache/flink/optimizer/postpass/SparseKeySchema.java @@ -26,7 +26,7 @@ import java.util.Map.Entry; import org.apache.flink.types.Key; /** - * Class encapsulating a schema map (int column position -> column type) and a reference counter. + * Class encapsulating a schema map (int column position -> column type) and a reference counter. */ public class SparseKeySchema extends AbstractSchema<Class<? extends Key<?>>> { http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-runtime/src/main/java/org/apache/flink/runtime/blob/BlobService.java ---------------------------------------------------------------------- diff --git a/flink-runtime/src/main/java/org/apache/flink/runtime/blob/BlobService.java b/flink-runtime/src/main/java/org/apache/flink/runtime/blob/BlobService.java index a2400b5..e6ea8d3 100644 --- a/flink-runtime/src/main/java/org/apache/flink/runtime/blob/BlobService.java +++ b/flink-runtime/src/main/java/org/apache/flink/runtime/blob/BlobService.java @@ -52,7 +52,6 @@ public interface BlobService { /** * Shutdown method which is called to terminate the blob service. - * @throws IOException */ void shutdown(); } http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/PendingCheckpoint.java ---------------------------------------------------------------------- diff --git a/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/PendingCheckpoint.java b/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/PendingCheckpoint.java index 81159f6..19c65d4 100644 --- a/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/PendingCheckpoint.java +++ b/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/PendingCheckpoint.java @@ -149,7 +149,6 @@ public class PendingCheckpoint { /** * Discards the pending checkpoint, releasing all held resources. - * @throws Exception */ public void discard(ClassLoader userClassLoader, boolean discardStateHandle) { synchronized (lock) { http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-runtime/src/main/java/org/apache/flink/runtime/execution/ExecutionState.java ---------------------------------------------------------------------- diff --git a/flink-runtime/src/main/java/org/apache/flink/runtime/execution/ExecutionState.java b/flink-runtime/src/main/java/org/apache/flink/runtime/execution/ExecutionState.java index 9f4a5a7..e3e3256 100644 --- a/flink-runtime/src/main/java/org/apache/flink/runtime/execution/ExecutionState.java +++ b/flink-runtime/src/main/java/org/apache/flink/runtime/execution/ExecutionState.java @@ -22,7 +22,7 @@ package org.apache.flink.runtime.execution; * An enumeration of all states that a task can be in during its execution. * Tasks usually start in the state {@code CREATED} and switch states according to * this diagram: - * <pre> + * <pre>{@code * * CREATED -> SCHEDULED -> DEPLOYING -> RUNNING -> FINISHED * | | | @@ -33,7 +33,7 @@ package org.apache.flink.runtime.execution; * +-------------------------+ * * ... -> FAILED - * </pre> + * }</pre> * * <p>It is possible to enter the {@code FAILED} state from any other state.</p> * http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-runtime/src/main/java/org/apache/flink/runtime/filecache/FileCache.java ---------------------------------------------------------------------- diff --git a/flink-runtime/src/main/java/org/apache/flink/runtime/filecache/FileCache.java b/flink-runtime/src/main/java/org/apache/flink/runtime/filecache/FileCache.java index fa5ad85..1b8b1d9 100644 --- a/flink-runtime/src/main/java/org/apache/flink/runtime/filecache/FileCache.java +++ b/flink-runtime/src/main/java/org/apache/flink/runtime/filecache/FileCache.java @@ -50,7 +50,7 @@ import org.slf4j.LoggerFactory; /** * The FileCache is used to create the local files for the registered cache files when a task is deployed. * The files will be removed when the task is unregistered after a 5 second delay. - * A given file x will be placed in "<system-tmp-dir>/tmp_<jobID>/". + * A given file x will be placed in "{@code <system-tmp-dir>/tmp_<jobID>/}". */ public class FileCache { http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-runtime/src/main/java/org/apache/flink/runtime/instance/SlotSharingGroupAssignment.java ---------------------------------------------------------------------- diff --git a/flink-runtime/src/main/java/org/apache/flink/runtime/instance/SlotSharingGroupAssignment.java b/flink-runtime/src/main/java/org/apache/flink/runtime/instance/SlotSharingGroupAssignment.java index 94249de..7d666fe 100644 --- a/flink-runtime/src/main/java/org/apache/flink/runtime/instance/SlotSharingGroupAssignment.java +++ b/flink-runtime/src/main/java/org/apache/flink/runtime/instance/SlotSharingGroupAssignment.java @@ -57,7 +57,7 @@ import org.slf4j.LoggerFactory; * * <p>Consider a job set up like this:</p> * - * <pre> + * <pre>{@code * +-------------- Slot Sharing Group --------------+ * | | * | +-- Co Location Group --+ | @@ -66,7 +66,7 @@ import org.slf4j.LoggerFactory; * | | | | * | +-----------------------+ | * +------------------------------------------------+ - * </pre> + * }</pre> * * <p>The slot hierarchy in the slot sharing group will look like the following</p> * http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-runtime/src/main/java/org/apache/flink/runtime/io/disk/iomanager/BlockChannelReader.java ---------------------------------------------------------------------- diff --git a/flink-runtime/src/main/java/org/apache/flink/runtime/io/disk/iomanager/BlockChannelReader.java b/flink-runtime/src/main/java/org/apache/flink/runtime/io/disk/iomanager/BlockChannelReader.java index 957052e..8e33cda 100644 --- a/flink-runtime/src/main/java/org/apache/flink/runtime/io/disk/iomanager/BlockChannelReader.java +++ b/flink-runtime/src/main/java/org/apache/flink/runtime/io/disk/iomanager/BlockChannelReader.java @@ -46,7 +46,7 @@ public interface BlockChannelReader<T> extends FileIOChannel { * such a segment is available, or until an error occurs in the reader, or the reader is closed. * <p> * WARNING: If this method is invoked without any segment ever returning (for example, because the - * {@link #readBlock(T)} method has not been invoked appropriately), the method may block + * {@link #readBlock} method has not been invoked appropriately), the method may block * forever. * * @return The next memory segment from the reader's return queue. http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-runtime/src/main/java/org/apache/flink/runtime/io/disk/iomanager/BlockChannelWriter.java ---------------------------------------------------------------------- diff --git a/flink-runtime/src/main/java/org/apache/flink/runtime/io/disk/iomanager/BlockChannelWriter.java b/flink-runtime/src/main/java/org/apache/flink/runtime/io/disk/iomanager/BlockChannelWriter.java index ccf065a..d6cfd7e 100644 --- a/flink-runtime/src/main/java/org/apache/flink/runtime/io/disk/iomanager/BlockChannelWriter.java +++ b/flink-runtime/src/main/java/org/apache/flink/runtime/io/disk/iomanager/BlockChannelWriter.java @@ -35,7 +35,7 @@ public interface BlockChannelWriter<T> extends BlockChannelWriterWithCallback<T> * writer is closed. * <p> * NOTE: If this method is invoked without any segment ever returning (for example, because the - * {@link #writeBlock(T)} method has not been invoked accordingly), the method may block + * {@link #writeBlock} method has not been invoked accordingly), the method may block * forever. * * @return The next memory segment from the writers's return queue. http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/partition/consumer/InputGate.java ---------------------------------------------------------------------- diff --git a/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/partition/consumer/InputGate.java b/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/partition/consumer/InputGate.java index f18c7e4..d49b27e 100644 --- a/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/partition/consumer/InputGate.java +++ b/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/partition/consumer/InputGate.java @@ -32,17 +32,17 @@ import java.io.IOException; * <p> As an example, consider a map-reduce program, where the map operator produces data and the * reduce operator consumes the produced data. * - * <pre> + * <pre>{@code * +-----+ +---------------------+ +--------+ * | Map | = produce => | Intermediate Result | <= consume = | Reduce | * +-----+ +---------------------+ +--------+ - * </pre> + * }</pre> * * <p> When deploying such a program in parallel, the intermediate result will be partitioned over its * producing parallel subtasks; each of these partitions is furthermore partitioned into one or more * subpartitions. * - * <pre> + * <pre>{@code * Intermediate result * +-----------------------------------------+ * | +----------------+ | +-----------------------+ @@ -57,7 +57,7 @@ import java.io.IOException; * +-------+ | +-------------+ +=> | Subpartition 2 | | <==+======== | Input Gate | Reduce 2 | * | +----------------+ | +-----------------------+ * +-----------------------------------------+ - * </pre> + * }</pre> * * <p> In the above example, two map subtasks produce the intermediate result in parallel, resulting * in two partitions (Partition 1 and 2). Each of these partitions is further partitioned into two http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/partition/consumer/SingleInputGate.java ---------------------------------------------------------------------- diff --git a/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/partition/consumer/SingleInputGate.java b/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/partition/consumer/SingleInputGate.java index 896fa9c..47f9228 100644 --- a/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/partition/consumer/SingleInputGate.java +++ b/flink-runtime/src/main/java/org/apache/flink/runtime/io/network/partition/consumer/SingleInputGate.java @@ -66,17 +66,17 @@ import static com.google.common.base.Preconditions.checkState; * <p> As an example, consider a map-reduce program, where the map operator produces data and the * reduce operator consumes the produced data. * - * <pre> + * <pre>{@code * +-----+ +---------------------+ +--------+ * | Map | = produce => | Intermediate Result | <= consume = | Reduce | * +-----+ +---------------------+ +--------+ - * </pre> + * }</pre> * * <p> When deploying such a program in parallel, the intermediate result will be partitioned over its * producing parallel subtasks; each of these partitions is furthermore partitioned into one or more * subpartitions. * - * <pre> + * <pre>{@code * Intermediate result * +-----------------------------------------+ * | +----------------+ | +-----------------------+ @@ -91,7 +91,7 @@ import static com.google.common.base.Preconditions.checkState; * +-------+ | +-------------+ +=> | Subpartition 2 | | <==+======== | Input Gate | Reduce 2 | * | +----------------+ | +-----------------------+ * +-----------------------------------------+ - * </pre> + * }</pre> * * <p> In the above example, two map subtasks produce the intermediate result in parallel, resulting * in two partitions (Partition 1 and 2). Each of these partitions is further partitioned into two http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-runtime/src/main/java/org/apache/flink/runtime/iterative/task/IterationIntermediateTask.java ---------------------------------------------------------------------- diff --git a/flink-runtime/src/main/java/org/apache/flink/runtime/iterative/task/IterationIntermediateTask.java b/flink-runtime/src/main/java/org/apache/flink/runtime/iterative/task/IterationIntermediateTask.java index 0d266c2..6836237 100644 --- a/flink-runtime/src/main/java/org/apache/flink/runtime/iterative/task/IterationIntermediateTask.java +++ b/flink-runtime/src/main/java/org/apache/flink/runtime/iterative/task/IterationIntermediateTask.java @@ -33,7 +33,7 @@ import org.slf4j.LoggerFactory; import java.io.IOException; /** - * An intermediate iteration task, which runs a {@link org.apache.flink.runtime.operators.PactDriver} inside. + * An intermediate iteration task, which runs a {@link org.apache.flink.runtime.operators.Driver} inside. * <p> * It will propagate {@link EndOfSuperstepEvent}s and {@link TerminationEvent}s to it's connected tasks. Furthermore * intermediate tasks can also update the iteration state, either the workset or the solution set. http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-runtime/src/main/java/org/apache/flink/runtime/jobmanager/scheduler/Scheduler.java ---------------------------------------------------------------------- diff --git a/flink-runtime/src/main/java/org/apache/flink/runtime/jobmanager/scheduler/Scheduler.java b/flink-runtime/src/main/java/org/apache/flink/runtime/jobmanager/scheduler/Scheduler.java index cb99e52..963fc4c 100644 --- a/flink-runtime/src/main/java/org/apache/flink/runtime/jobmanager/scheduler/Scheduler.java +++ b/flink-runtime/src/main/java/org/apache/flink/runtime/jobmanager/scheduler/Scheduler.java @@ -57,7 +57,7 @@ import scala.concurrent.ExecutionContext; * <p>The scheduler supports two scheduling modes:</p> * <ul> * <li>Immediate scheduling: A request for a task slot immediately returns a task slot, if one is - * available, or throws a {@link NoResourceAvailableException}</li>. + * available, or throws a {@link NoResourceAvailableException}.</li> * <li>Queued Scheduling: A request for a task slot is queued and returns a future that will be * fulfilled as soon as a slot becomes available.</li> * </ul> http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-runtime/src/main/java/org/apache/flink/runtime/operators/sort/AbstractMergeInnerJoinIterator.java ---------------------------------------------------------------------- diff --git a/flink-runtime/src/main/java/org/apache/flink/runtime/operators/sort/AbstractMergeInnerJoinIterator.java b/flink-runtime/src/main/java/org/apache/flink/runtime/operators/sort/AbstractMergeInnerJoinIterator.java index db56227..1b54f20 100644 --- a/flink-runtime/src/main/java/org/apache/flink/runtime/operators/sort/AbstractMergeInnerJoinIterator.java +++ b/flink-runtime/src/main/java/org/apache/flink/runtime/operators/sort/AbstractMergeInnerJoinIterator.java @@ -53,7 +53,7 @@ public abstract class AbstractMergeInnerJoinIterator<T1, T2, O> extends Abstract /** * Calls the <code>JoinFunction#join()</code> method for all two key-value pairs that share the same key and come * from different inputs. The output of the <code>join()</code> method is forwarded. - * <p/> + * <p> * This method first zig-zags between the two sorted inputs in order to find a common * key, and then calls the join stub with the cross product of the values. * http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-runtime/src/main/java/org/apache/flink/runtime/operators/util/BloomFilter.java ---------------------------------------------------------------------- diff --git a/flink-runtime/src/main/java/org/apache/flink/runtime/operators/util/BloomFilter.java b/flink-runtime/src/main/java/org/apache/flink/runtime/operators/util/BloomFilter.java index 947a56b..0ad3c6d 100644 --- a/flink-runtime/src/main/java/org/apache/flink/runtime/operators/util/BloomFilter.java +++ b/flink-runtime/src/main/java/org/apache/flink/runtime/operators/util/BloomFilter.java @@ -6,9 +6,9 @@ * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at - * <p/> + * <p> * http://www.apache.org/licenses/LICENSE-2.0 - * <p/> + * <p> * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. @@ -34,11 +34,11 @@ import static com.google.common.base.Preconditions.checkArgument; * During the creation of bloom filter expected number of entries must be specified. If the number * of insertions exceed the specified initial number of entries then false positive probability will * increase accordingly. - * <p/> + * <p> * Internally, this implementation of bloom filter uses MemorySegment to store BitSet, BloomFilter and * BitSet are designed to be able to switch between different MemorySegments, so that Flink can share * the same BloomFilter/BitSet object instance for different bloom filters. - * <p/> + * <p> * Part of this class refers to the implementation from Apache Hive project * https://github.com/apache/hive/blob/master/common/src/java/org/apache/hive/common/util/BloomFilter.java */ http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-runtime/src/main/java/org/apache/flink/runtime/webmonitor/WebMonitorUtils.java ---------------------------------------------------------------------- diff --git a/flink-runtime/src/main/java/org/apache/flink/runtime/webmonitor/WebMonitorUtils.java b/flink-runtime/src/main/java/org/apache/flink/runtime/webmonitor/WebMonitorUtils.java index 181d6d6..a7a3458 100644 --- a/flink-runtime/src/main/java/org/apache/flink/runtime/webmonitor/WebMonitorUtils.java +++ b/flink-runtime/src/main/java/org/apache/flink/runtime/webmonitor/WebMonitorUtils.java @@ -113,7 +113,7 @@ public final class WebMonitorUtils { /** * Starts the web runtime monitor. Because the actual implementation of the runtime monitor is * in another project, we load the runtime monitor dynamically. - * <p/> + * <p> * Because failure to start the web runtime monitor is not considered fatal, this method does * not throw any exceptions, but only logs them. * http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-runtime/src/main/java/org/apache/flink/runtime/yarn/AbstractFlinkYarnCluster.java ---------------------------------------------------------------------- diff --git a/flink-runtime/src/main/java/org/apache/flink/runtime/yarn/AbstractFlinkYarnCluster.java b/flink-runtime/src/main/java/org/apache/flink/runtime/yarn/AbstractFlinkYarnCluster.java index 3f78898..d7ef904 100644 --- a/flink-runtime/src/main/java/org/apache/flink/runtime/yarn/AbstractFlinkYarnCluster.java +++ b/flink-runtime/src/main/java/org/apache/flink/runtime/yarn/AbstractFlinkYarnCluster.java @@ -100,7 +100,9 @@ public abstract class AbstractFlinkYarnCluster { /** * Disconnect from the ApplicationMaster without stopping the session - * (therefore, use the {@see shutdown()} method. + * (therefore, use the {@link AbstractFlinkYarnCluster#shutdown(boolean)} method. + * + * @see AbstractFlinkYarnCluster#shutdown(boolean) */ public abstract void disconnect(); http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-runtime/src/main/java/org/apache/flink/runtime/zookeeper/StateStorageHelper.java ---------------------------------------------------------------------- diff --git a/flink-runtime/src/main/java/org/apache/flink/runtime/zookeeper/StateStorageHelper.java b/flink-runtime/src/main/java/org/apache/flink/runtime/zookeeper/StateStorageHelper.java index d18cace..36fb849 100644 --- a/flink-runtime/src/main/java/org/apache/flink/runtime/zookeeper/StateStorageHelper.java +++ b/flink-runtime/src/main/java/org/apache/flink/runtime/zookeeper/StateStorageHelper.java @@ -23,7 +23,7 @@ import org.apache.flink.runtime.state.StateHandle; import java.io.Serializable; /** - * State storage helper which is used by {@ZooKeeperStateHandleStore} to persiste state before + * State storage helper which is used by {@link ZooKeeperStateHandleStore} to persiste state before * the state handle is written to ZooKeeper. * * @param <T> http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-runtime/src/main/java/org/apache/flink/runtime/zookeeper/ZooKeeperStateHandleStore.java ---------------------------------------------------------------------- diff --git a/flink-runtime/src/main/java/org/apache/flink/runtime/zookeeper/ZooKeeperStateHandleStore.java b/flink-runtime/src/main/java/org/apache/flink/runtime/zookeeper/ZooKeeperStateHandleStore.java index 6073a39..1d89d64 100644 --- a/flink-runtime/src/main/java/org/apache/flink/runtime/zookeeper/ZooKeeperStateHandleStore.java +++ b/flink-runtime/src/main/java/org/apache/flink/runtime/zookeeper/ZooKeeperStateHandleStore.java @@ -50,13 +50,13 @@ import static com.google.common.base.Preconditions.checkNotNull; * <p>ZooKeeper holds the ground truth about state handles, i.e. the following holds: * * <pre> - * State handle in ZooKeeper => State handle exists + * State handle in ZooKeeper => State handle exists * </pre> * * But not: * * <pre> - * State handle exists => State handle in ZooKeeper + * State handle exists => State handle in ZooKeeper * </pre> * * There can be lingering state handles when failures happen during operation. They @@ -316,7 +316,7 @@ public class ZooKeeperStateHandleStore<T extends Serializable> { /** * Removes a state handle from ZooKeeper. * - * <p><stong>Important</stong>: this does not discard the state handle. If you want to + * <p><strong>Important</strong>: this does not discard the state handle. If you want to * discard the state handle call {@link #removeAndDiscardState(String)}. * * @param pathInZooKeeper Path of state handle to remove (expected to start with a '/') @@ -331,7 +331,7 @@ public class ZooKeeperStateHandleStore<T extends Serializable> { /** * Removes a state handle from ZooKeeper asynchronously. * - * <p><stong>Important</stong>: this does not discard the state handle. If you want to + * <p><strong>Important</strong>: this does not discard the state handle. If you want to * discard the state handle call {@link #removeAndDiscardState(String)}. * * @param pathInZooKeeper Path of state handle to remove (expected to start with a '/') http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-staging/flink-hadoop-compatibility/src/main/java/org/apache/flink/hadoopcompatibility/mapred/HadoopMapFunction.java ---------------------------------------------------------------------- diff --git a/flink-staging/flink-hadoop-compatibility/src/main/java/org/apache/flink/hadoopcompatibility/mapred/HadoopMapFunction.java b/flink-staging/flink-hadoop-compatibility/src/main/java/org/apache/flink/hadoopcompatibility/mapred/HadoopMapFunction.java index 4d81daf..83ab23d 100644 --- a/flink-staging/flink-hadoop-compatibility/src/main/java/org/apache/flink/hadoopcompatibility/mapred/HadoopMapFunction.java +++ b/flink-staging/flink-hadoop-compatibility/src/main/java/org/apache/flink/hadoopcompatibility/mapred/HadoopMapFunction.java @@ -111,7 +111,7 @@ public final class HadoopMapFunction<KEYIN, VALUEIN, KEYOUT, VALUEOUT> /** * Custom serialization methods. - * @see http://docs.oracle.com/javase/7/docs/api/java/io/Serializable.html + * @see <a href="http://docs.oracle.com/javase/7/docs/api/java/io/Serializable.html">http://docs.oracle.com/javase/7/docs/api/java/io/Serializable.html</a> */ private void writeObject(final ObjectOutputStream out) throws IOException { out.writeObject(mapper.getClass()); http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-staging/flink-hadoop-compatibility/src/main/java/org/apache/flink/hadoopcompatibility/mapred/HadoopReduceCombineFunction.java ---------------------------------------------------------------------- diff --git a/flink-staging/flink-hadoop-compatibility/src/main/java/org/apache/flink/hadoopcompatibility/mapred/HadoopReduceCombineFunction.java b/flink-staging/flink-hadoop-compatibility/src/main/java/org/apache/flink/hadoopcompatibility/mapred/HadoopReduceCombineFunction.java index f00859f..97b9768 100644 --- a/flink-staging/flink-hadoop-compatibility/src/main/java/org/apache/flink/hadoopcompatibility/mapred/HadoopReduceCombineFunction.java +++ b/flink-staging/flink-hadoop-compatibility/src/main/java/org/apache/flink/hadoopcompatibility/mapred/HadoopReduceCombineFunction.java @@ -138,7 +138,7 @@ public final class HadoopReduceCombineFunction<KEYIN, VALUEIN, KEYOUT, VALUEOUT> /** * Custom serialization methods. - * @see http://docs.oracle.com/javase/7/docs/api/java/io/Serializable.html + * @see <a href="http://docs.oracle.com/javase/7/docs/api/java/io/Serializable.html">http://docs.oracle.com/javase/7/docs/api/java/io/Serializable.html</a> */ private void writeObject(final ObjectOutputStream out) throws IOException { http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-staging/flink-hadoop-compatibility/src/main/java/org/apache/flink/hadoopcompatibility/mapred/HadoopReduceFunction.java ---------------------------------------------------------------------- diff --git a/flink-staging/flink-hadoop-compatibility/src/main/java/org/apache/flink/hadoopcompatibility/mapred/HadoopReduceFunction.java b/flink-staging/flink-hadoop-compatibility/src/main/java/org/apache/flink/hadoopcompatibility/mapred/HadoopReduceFunction.java index 6943421..1c47696 100644 --- a/flink-staging/flink-hadoop-compatibility/src/main/java/org/apache/flink/hadoopcompatibility/mapred/HadoopReduceFunction.java +++ b/flink-staging/flink-hadoop-compatibility/src/main/java/org/apache/flink/hadoopcompatibility/mapred/HadoopReduceFunction.java @@ -118,8 +118,8 @@ public final class HadoopReduceFunction<KEYIN, VALUEIN, KEYOUT, VALUEOUT> } /** - * Custom serialization methods. - * @see http://docs.oracle.com/javase/7/docs/api/java/io/Serializable.html + * Custom serialization methods + * @see <a href="http://docs.oracle.com/javase/7/docs/api/java/io/Serializable.html">http://docs.oracle.com/javase/7/docs/api/java/io/Serializable.html</a> */ private void writeObject(final ObjectOutputStream out) throws IOException { http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-staging/flink-scala-shell/src/main/java/org/apache/flink/api/java/JarHelper.java ---------------------------------------------------------------------- diff --git a/flink-staging/flink-scala-shell/src/main/java/org/apache/flink/api/java/JarHelper.java b/flink-staging/flink-scala-shell/src/main/java/org/apache/flink/api/java/JarHelper.java index e085c25..83fb342 100644 --- a/flink-staging/flink-scala-shell/src/main/java/org/apache/flink/api/java/JarHelper.java +++ b/flink-staging/flink-scala-shell/src/main/java/org/apache/flink/api/java/JarHelper.java @@ -37,7 +37,7 @@ import java.util.jar.JarInputStream; * * Copied from http://grepcode.com/file_/repo1.maven.org/maven2/org.apache.xmlbeans/xmlbeans/2.4.0/org/apache/xmlbeans/impl/common/JarHelper.java/?v=source * - * @author Patrick Calahan <p...@bea.com> + * @author Patrick Calahan <a href="mailto:p...@bea.com">p...@bea.com</a> */ public class JarHelper { http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-staging/flink-table/src/main/java/org/apache/flink/api/java/table/package-info.java ---------------------------------------------------------------------- diff --git a/flink-staging/flink-table/src/main/java/org/apache/flink/api/java/table/package-info.java b/flink-staging/flink-table/src/main/java/org/apache/flink/api/java/table/package-info.java index 6f9f0a3..97113bb 100644 --- a/flink-staging/flink-table/src/main/java/org/apache/flink/api/java/table/package-info.java +++ b/flink-staging/flink-table/src/main/java/org/apache/flink/api/java/table/package-info.java @@ -31,7 +31,7 @@ * <p> * Example: * - * <code> + * <pre>{@code * ExecutionEnvironment env = ExecutionEnvironment.createCollectionsEnvironment(); * * DataSet<WC> input = env.fromElements( @@ -50,11 +50,11 @@ * * result.print(); * env.execute(); - * </code> + * }</pre> * * <p> * As seen above, a {@link org.apache.flink.api.table.Table} can be converted back to the - * underlying API representation using {@link org.apache.flink.api.java.table.TableEnvironment.toSet} - * or {@link org.apache.flink.api.java.table.TableEnvironment.toStream}. + * underlying API representation using {@link org.apache.flink.api.java.table.TableEnvironment#toDataSet(org.apache.flink.api.table.Table, java.lang.Class)} + * or {@link org.apache.flink.api.java.table.TableEnvironment#toDataStream(org.apache.flink.api.table.Table, java.lang.Class)}}. */ package org.apache.flink.api.java.table; http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-streaming-connectors/flink-connector-kafka/src/main/java/org/apache/flink/streaming/connectors/kafka/FlinkKafkaConsumer.java ---------------------------------------------------------------------- diff --git a/flink-streaming-connectors/flink-connector-kafka/src/main/java/org/apache/flink/streaming/connectors/kafka/FlinkKafkaConsumer.java b/flink-streaming-connectors/flink-connector-kafka/src/main/java/org/apache/flink/streaming/connectors/kafka/FlinkKafkaConsumer.java index 8066b3c..c7ae1cc 100644 --- a/flink-streaming-connectors/flink-connector-kafka/src/main/java/org/apache/flink/streaming/connectors/kafka/FlinkKafkaConsumer.java +++ b/flink-streaming-connectors/flink-connector-kafka/src/main/java/org/apache/flink/streaming/connectors/kafka/FlinkKafkaConsumer.java @@ -79,7 +79,7 @@ import static com.google.common.base.Preconditions.checkNotNull; * <ul> * <li>{@link FetcherType#NEW_HIGH_LEVEL}: A fetcher based on the new Kafka consumer API. * This fetcher is generally more robust, but works only with later versions of - * Kafka (> 0.8.2).</li> + * Kafka (> 0.8.2).</li> * * <li>{@link FetcherType#LEGACY_LOW_LEVEL}: A fetcher based on the old low-level consumer API. * This fetcher is works also with older versions of Kafka (0.8.1). The fetcher interprets http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-streaming-connectors/flink-connector-kafka/src/main/java/org/apache/flink/streaming/connectors/kafka/partitioner/FixedPartitioner.java ---------------------------------------------------------------------- diff --git a/flink-streaming-connectors/flink-connector-kafka/src/main/java/org/apache/flink/streaming/connectors/kafka/partitioner/FixedPartitioner.java b/flink-streaming-connectors/flink-connector-kafka/src/main/java/org/apache/flink/streaming/connectors/kafka/partitioner/FixedPartitioner.java index 346a7d5..7ab7290 100644 --- a/flink-streaming-connectors/flink-connector-kafka/src/main/java/org/apache/flink/streaming/connectors/kafka/partitioner/FixedPartitioner.java +++ b/flink-streaming-connectors/flink-connector-kafka/src/main/java/org/apache/flink/streaming/connectors/kafka/partitioner/FixedPartitioner.java @@ -28,24 +28,24 @@ import java.io.Serializable; * # More Flink partitions than kafka partitions * <pre> * Flink Sinks: Kafka Partitions - * 1 ----------------> 1 + * 1 ----------------> 1 * 2 --------------/ * 3 -------------/ * 4 ------------/ * </pre> - * --> Some (or all) kafka partitions contain the output of more than one flink partition + * Some (or all) kafka partitions contain the output of more than one flink partition * *# Fewer Flink partitions than Kafka * <pre> * Flink Sinks: Kafka Partitions - * 1 ----------------> 1 - * 2 ----------------> 2 - * 3 - * 4 - * 5 + * 1 ----------------> 1 + * 2 ----------------> 2 + * 3 + * 4 + * 5 * </pre> * - * --> Not all Kafka partitions contain data + * Not all Kafka partitions contain data * To avoid such an unbalanced partitioning, use a round-robin kafka partitioner. (note that this will * cause a lot of network connections between all the Flink instances and all the Kafka brokers * http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/iteration/IterateExample.java ---------------------------------------------------------------------- diff --git a/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/iteration/IterateExample.java b/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/iteration/IterateExample.java index 2cf66b9..52ec896 100644 --- a/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/iteration/IterateExample.java +++ b/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/iteration/IterateExample.java @@ -32,10 +32,11 @@ import java.util.List; import java.util.Random; /** - * Example illustrating iterations in Flink streaming. <p/> <p> The program sums up random numbers and counts additions + * Example illustrating iterations in Flink streaming. + * <p> The program sums up random numbers and counts additions * it performs to reach a specific threshold in an iterative streaming fashion. </p> - * <p/> - * <p/> + * + * <p> * This example shows how to use: <ul> <li>streaming iterations, <li>buffer timeout to enhance latency, <li>directed * outputs. </ul> */ http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/ml/IncrementalLearningSkeleton.java ---------------------------------------------------------------------- diff --git a/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/ml/IncrementalLearningSkeleton.java b/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/ml/IncrementalLearningSkeleton.java index ce227e4..32cf430 100644 --- a/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/ml/IncrementalLearningSkeleton.java +++ b/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/ml/IncrementalLearningSkeleton.java @@ -40,7 +40,7 @@ import java.util.concurrent.TimeUnit; * incremental Alternating Least Squares model while also providing the * predictions. * - * <p/> + * <p> * This example shows how to use: * <ul> * <li>Connected streams @@ -188,7 +188,6 @@ public class IncrementalLearningSkeleton { /** * Creates newData using the model produced in batch-processing and the * up-to-date partial model. - * <p/> * <p> * By defaults emits the Integer 0 for every newData and the Integer 1 * for every model update. http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/socket/SocketTextStreamWordCount.java ---------------------------------------------------------------------- diff --git a/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/socket/SocketTextStreamWordCount.java b/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/socket/SocketTextStreamWordCount.java index 17add2c..cecabdd 100644 --- a/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/socket/SocketTextStreamWordCount.java +++ b/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/socket/SocketTextStreamWordCount.java @@ -26,19 +26,16 @@ import org.apache.flink.streaming.examples.wordcount.WordCount.Tokenizer; * This example shows an implementation of WordCount with data from a text * socket. To run the example make sure that the service providing the text data * is already up and running. - * <p/> - * <p/> + * <p> * To start an example socket text stream on your local machine run netcat from * a command line: <code>nc -lk 9999</code>, where the parameter specifies the * port number. - * <p/> - * <p/> - * <p/> + * </p> + * <p> * Usage: * <code>SocketTextStreamWordCount <hostname> <port> <result path></code> - * <br> - * <p/> - * <p/> + * </p> + * <p> * This example shows how to: * <ul> * <li>use StreamExecutionEnvironment.socketTextStream http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/twitter/TwitterStream.java ---------------------------------------------------------------------- diff --git a/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/twitter/TwitterStream.java b/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/twitter/TwitterStream.java index c2477b5..06872f0 100644 --- a/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/twitter/TwitterStream.java +++ b/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/twitter/TwitterStream.java @@ -31,16 +31,15 @@ import java.util.StringTokenizer; /** * Implements the "TwitterStream" program that computes a most used word * occurrence over JSON files in a streaming fashion. - * <p/> - * <p/> + * <p> * The input is a JSON text file with lines separated by newline characters. - * <p/> - * <p/> + * </p> + * <p> * Usage: <code>TwitterStream <text path></code><br> * If no parameters are provided, the program is run with default data from * {@link TwitterStreamData}. - * <p/> - * <p/> + * </p> + * <p> * This example shows how to: * <ul> * <li>acquire external data, @@ -88,13 +87,11 @@ public class TwitterStream { /** * Makes sentences from English tweets. - * <p/> * <p> * Implements a string tokenizer that splits sentences into words as a * user-defined FlatMapFunction. The function takes a line (String) and - * splits it into multiple pairs in the form of "(word,1)" (Tuple2<String, - * Integer>). - * </p> + * splits it into multiple pairs in the form of "(word,1)" ({@code Tuple2<String, + * Integer>}). */ public static class SelectEnglishAndTokenizeFlatMap extends JSONParseFlatMap<String, Tuple2<String, Integer>> { private static final long serialVersionUID = 1L; http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/wordcount/PojoExample.java ---------------------------------------------------------------------- diff --git a/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/wordcount/PojoExample.java b/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/wordcount/PojoExample.java index 591ef51..9b0b63c 100644 --- a/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/wordcount/PojoExample.java +++ b/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/wordcount/PojoExample.java @@ -126,8 +126,8 @@ public class PojoExample { /** * Implements the string tokenizer that splits sentences into words as a * user-defined FlatMapFunction. The function takes a line (String) and - * splits it into multiple pairs in the form of "(word,1)" (Tuple2<String, - * Integer>). + * splits it into multiple pairs in the form of "(word,1)" ({@code Tuple2<String, + * Integer>}). */ public static final class Tokenizer implements FlatMapFunction<String, Word> { private static final long serialVersionUID = 1L; http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/wordcount/WordCount.java ---------------------------------------------------------------------- diff --git a/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/wordcount/WordCount.java b/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/wordcount/WordCount.java index a594c94..a76671e 100644 --- a/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/wordcount/WordCount.java +++ b/flink-streaming-examples/src/main/java/org/apache/flink/streaming/examples/wordcount/WordCount.java @@ -87,8 +87,8 @@ public class WordCount { /** * Implements the string tokenizer that splits sentences into words as a * user-defined FlatMapFunction. The function takes a line (String) and - * splits it into multiple pairs in the form of "(word,1)" (Tuple2<String, - * Integer>). + * splits it into multiple pairs in the form of "(word,1)" ({@code Tuple2<String, + * Integer>}). */ public static final class Tokenizer implements FlatMapFunction<String, Tuple2<String, Integer>> { private static final long serialVersionUID = 1L; http://git-wip-us.apache.org/repos/asf/flink/blob/e1f30b04/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/datastream/DataStream.java ---------------------------------------------------------------------- diff --git a/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/datastream/DataStream.java b/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/datastream/DataStream.java index 176a07f..309eb05 100644 --- a/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/datastream/DataStream.java +++ b/flink-streaming-java/src/main/java/org/apache/flink/streaming/api/datastream/DataStream.java @@ -598,7 +598,7 @@ public class DataStream<T> { } /** - * Initiates a Project transformation on a {@link Tuple} {@link DataStream}.<br/> + * Initiates a Project transformation on a {@link Tuple} {@link DataStream}.<br> * <b>Note: Only Tuple DataStreams can be projected.</b> * * <p>