Repository: spark
Updated Branches:
  refs/heads/master 07c12c09a -> aefe79890


[MINOR][BUILD] Fix javadoc8 break

## What changes were proposed in this pull request?

Several javadoc8 breaks have been introduced. This PR proposes fix those 
instances so that we can build Scala/Java API docs.

```
[error] 
.../spark/sql/core/target/java/org/apache/spark/sql/streaming/GroupState.java:6:
 error: reference not found
[error]  * <code>flatMapGroupsWithState</code> operations on {link 
KeyValueGroupedDataset}.
[error]                                                             ^
[error] 
.../spark/sql/core/target/java/org/apache/spark/sql/streaming/GroupState.java:10:
 error: reference not found
[error]  * Both, <code>mapGroupsWithState</code> and 
<code>flatMapGroupsWithState</code> in {link KeyValueGroupedDataset}
[error]                                                                         
                   ^
[error] 
.../spark/sql/core/target/java/org/apache/spark/sql/streaming/GroupState.java:51:
 error: reference not found
[error]  *    {link GroupStateTimeout.ProcessingTimeTimeout}) or event time 
(i.e.
[error]              ^
[error] 
.../spark/sql/core/target/java/org/apache/spark/sql/streaming/GroupState.java:52:
 error: reference not found
[error]  *    {link GroupStateTimeout.EventTimeTimeout}).
[error]              ^
[error] 
.../spark/sql/core/target/java/org/apache/spark/sql/streaming/GroupState.java:158:
 error: reference not found
[error]  *           Spark SQL types (see {link Encoder} for more details).
[error]                                          ^
[error] 
.../spark/mllib/target/java/org/apache/spark/ml/fpm/FPGrowthParams.java:26: 
error: bad use of '>'
[error]    * Number of partitions (>=1) used by parallel FP-growth. By default 
the param is not set, and
[error]                            ^
[error] 
.../spark/sql/core/src/main/java/org/apache/spark/api/java/function/FlatMapGroupsWithStateFunction.java:30:
 error: reference not found
[error]  * {link 
org.apache.spark.sql.KeyValueGroupedDataset#flatMapGroupsWithState(
[error]           ^
[error] 
.../spark/sql/core/target/java/org/apache/spark/sql/KeyValueGroupedDataset.java:211:
 error: reference not found
[error]    * See {link GroupState} for more details.
[error]                 ^
[error] 
.../spark/sql/core/target/java/org/apache/spark/sql/KeyValueGroupedDataset.java:232:
 error: reference not found
[error]    * See {link GroupState} for more details.
[error]                 ^
[error] 
.../spark/sql/core/target/java/org/apache/spark/sql/KeyValueGroupedDataset.java:254:
 error: reference not found
[error]    * See {link GroupState} for more details.
[error]                 ^
[error] 
.../spark/sql/core/target/java/org/apache/spark/sql/KeyValueGroupedDataset.java:277:
 error: reference not found
[error]    * See {link GroupState} for more details.
[error]                 ^
[error] .../spark/core/target/java/org/apache/spark/TaskContextImpl.java:10: 
error: reference not found
[error]  * {link TaskMetrics} &amp; {link MetricsSystem} objects are not thread 
safe.
[error]           ^
[error] .../spark/core/target/java/org/apache/spark/TaskContextImpl.java:10: 
error: reference not found
[error]  * {link TaskMetrics} &amp; {link MetricsSystem} objects are not thread 
safe.
[error]                                     ^
[info] 13 errors
```

```
jekyll 3.3.1 | Error:  Unidoc generation failed
```

## How was this patch tested?

Manually via `jekyll build`

Author: hyukjinkwon <gurwls...@gmail.com>

Closes #17389 from HyukjinKwon/minor-javadoc8-fix.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/aefe7989
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/aefe7989
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/aefe7989

Branch: refs/heads/master
Commit: aefe79890541bc0829f184e03eb3961739ca8ef2
Parents: 07c12c0
Author: hyukjinkwon <gurwls...@gmail.com>
Authored: Thu Mar 23 08:41:30 2017 +0000
Committer: Sean Owen <so...@cloudera.com>
Committed: Thu Mar 23 08:41:30 2017 +0000

----------------------------------------------------------------------
 .../org/apache/spark/TaskContextImpl.scala      |  2 +-
 .../org/apache/spark/ml/fpm/FPGrowth.scala      |  4 ++--
 .../FlatMapGroupsWithStateFunction.java         |  2 +-
 .../spark/sql/KeyValueGroupedDataset.scala      |  8 +++----
 .../apache/spark/sql/streaming/GroupState.scala | 22 ++++++++++----------
 5 files changed, 19 insertions(+), 19 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/spark/blob/aefe7989/core/src/main/scala/org/apache/spark/TaskContextImpl.scala
----------------------------------------------------------------------
diff --git a/core/src/main/scala/org/apache/spark/TaskContextImpl.scala 
b/core/src/main/scala/org/apache/spark/TaskContextImpl.scala
index ea8dcdf..f346cf8 100644
--- a/core/src/main/scala/org/apache/spark/TaskContextImpl.scala
+++ b/core/src/main/scala/org/apache/spark/TaskContextImpl.scala
@@ -38,7 +38,7 @@ import org.apache.spark.util._
  * callbacks are protected by locking on the context instance. For instance, 
this ensures
  * that you cannot add a completion listener in one thread while we are 
completing (and calling
  * the completion listeners) in another thread. Other state is immutable, 
however the exposed
- * [[TaskMetrics]] & [[MetricsSystem]] objects are not thread safe.
+ * `TaskMetrics` & `MetricsSystem` objects are not thread safe.
  */
 private[spark] class TaskContextImpl(
     val stageId: Int,

http://git-wip-us.apache.org/repos/asf/spark/blob/aefe7989/mllib/src/main/scala/org/apache/spark/ml/fpm/FPGrowth.scala
----------------------------------------------------------------------
diff --git a/mllib/src/main/scala/org/apache/spark/ml/fpm/FPGrowth.scala 
b/mllib/src/main/scala/org/apache/spark/ml/fpm/FPGrowth.scala
index e2bc270..65cc806 100644
--- a/mllib/src/main/scala/org/apache/spark/ml/fpm/FPGrowth.scala
+++ b/mllib/src/main/scala/org/apache/spark/ml/fpm/FPGrowth.scala
@@ -69,8 +69,8 @@ private[fpm] trait FPGrowthParams extends Params with 
HasPredictionCol {
   def getMinSupport: Double = $(minSupport)
 
   /**
-   * Number of partitions (>=1) used by parallel FP-growth. By default the 
param is not set, and
-   * partition number of the input dataset is used.
+   * Number of partitions (at least 1) used by parallel FP-growth. By default 
the param is not
+   * set, and partition number of the input dataset is used.
    * @group expertParam
    */
   @Since("2.2.0")

http://git-wip-us.apache.org/repos/asf/spark/blob/aefe7989/sql/core/src/main/java/org/apache/spark/api/java/function/FlatMapGroupsWithStateFunction.java
----------------------------------------------------------------------
diff --git 
a/sql/core/src/main/java/org/apache/spark/api/java/function/FlatMapGroupsWithStateFunction.java
 
b/sql/core/src/main/java/org/apache/spark/api/java/function/FlatMapGroupsWithStateFunction.java
index 026b37c..802949c 100644
--- 
a/sql/core/src/main/java/org/apache/spark/api/java/function/FlatMapGroupsWithStateFunction.java
+++ 
b/sql/core/src/main/java/org/apache/spark/api/java/function/FlatMapGroupsWithStateFunction.java
@@ -27,7 +27,7 @@ import org.apache.spark.sql.streaming.GroupState;
 /**
  * ::Experimental::
  * Base interface for a map function used in
- * {@link org.apache.spark.sql.KeyValueGroupedDataset#flatMapGroupsWithState(
+ * {@code org.apache.spark.sql.KeyValueGroupedDataset.flatMapGroupsWithState(
  * FlatMapGroupsWithStateFunction, org.apache.spark.sql.streaming.OutputMode,
  * org.apache.spark.sql.Encoder, org.apache.spark.sql.Encoder)}
  * @since 2.1.1

http://git-wip-us.apache.org/repos/asf/spark/blob/aefe7989/sql/core/src/main/scala/org/apache/spark/sql/KeyValueGroupedDataset.scala
----------------------------------------------------------------------
diff --git 
a/sql/core/src/main/scala/org/apache/spark/sql/KeyValueGroupedDataset.scala 
b/sql/core/src/main/scala/org/apache/spark/sql/KeyValueGroupedDataset.scala
index 87c5621..022c2f5 100644
--- a/sql/core/src/main/scala/org/apache/spark/sql/KeyValueGroupedDataset.scala
+++ b/sql/core/src/main/scala/org/apache/spark/sql/KeyValueGroupedDataset.scala
@@ -298,7 +298,7 @@ class KeyValueGroupedDataset[K, V] private[sql](
    * For a static batch Dataset, the function will be invoked once per group. 
For a streaming
    * Dataset, the function will be invoked for each group repeatedly in every 
trigger, and
    * updates to each group's state will be saved across invocations.
-   * See [[GroupState]] for more details.
+   * See `GroupState` for more details.
    *
    * @tparam S The type of the user-defined state. Must be encodable to Spark 
SQL types.
    * @tparam U The type of the output objects. Must be encodable to Spark SQL 
types.
@@ -328,7 +328,7 @@ class KeyValueGroupedDataset[K, V] private[sql](
    * For a static batch Dataset, the function will be invoked once per group. 
For a streaming
    * Dataset, the function will be invoked for each group repeatedly in every 
trigger, and
    * updates to each group's state will be saved across invocations.
-   * See [[GroupState]] for more details.
+   * See `GroupState` for more details.
    *
    * @tparam S The type of the user-defined state. Must be encodable to Spark 
SQL types.
    * @tparam U The type of the output objects. Must be encodable to Spark SQL 
types.
@@ -360,7 +360,7 @@ class KeyValueGroupedDataset[K, V] private[sql](
    * For a static batch Dataset, the function will be invoked once per group. 
For a streaming
    * Dataset, the function will be invoked for each group repeatedly in every 
trigger, and
    * updates to each group's state will be saved across invocations.
-   * See [[GroupState]] for more details.
+   * See `GroupState` for more details.
    *
    * @tparam S The type of the user-defined state. Must be encodable to Spark 
SQL types.
    * @tparam U The type of the output objects. Must be encodable to Spark SQL 
types.
@@ -400,7 +400,7 @@ class KeyValueGroupedDataset[K, V] private[sql](
    * For a static batch Dataset, the function will be invoked once per group. 
For a streaming
    * Dataset, the function will be invoked for each group repeatedly in every 
trigger, and
    * updates to each group's state will be saved across invocations.
-   * See [[GroupState]] for more details.
+   * See `GroupState` for more details.
    *
    * @tparam S The type of the user-defined state. Must be encodable to Spark 
SQL types.
    * @tparam U The type of the output objects. Must be encodable to Spark SQL 
types.

http://git-wip-us.apache.org/repos/asf/spark/blob/aefe7989/sql/core/src/main/scala/org/apache/spark/sql/streaming/GroupState.scala
----------------------------------------------------------------------
diff --git 
a/sql/core/src/main/scala/org/apache/spark/sql/streaming/GroupState.scala 
b/sql/core/src/main/scala/org/apache/spark/sql/streaming/GroupState.scala
index 60a4d0d..15df906 100644
--- a/sql/core/src/main/scala/org/apache/spark/sql/streaming/GroupState.scala
+++ b/sql/core/src/main/scala/org/apache/spark/sql/streaming/GroupState.scala
@@ -18,18 +18,18 @@
 package org.apache.spark.sql.streaming
 
 import org.apache.spark.annotation.{Experimental, InterfaceStability}
-import org.apache.spark.sql.{Encoder, KeyValueGroupedDataset}
+import org.apache.spark.sql.KeyValueGroupedDataset
 import org.apache.spark.sql.catalyst.plans.logical.LogicalGroupState
 
 /**
  * :: Experimental ::
  *
  * Wrapper class for interacting with per-group state data in 
`mapGroupsWithState` and
- * `flatMapGroupsWithState` operations on [[KeyValueGroupedDataset]].
+ * `flatMapGroupsWithState` operations on `KeyValueGroupedDataset`.
  *
  * Detail description on `[map/flatMap]GroupsWithState` operation
  * --------------------------------------------------------------
- * Both, `mapGroupsWithState` and `flatMapGroupsWithState` in 
[[KeyValueGroupedDataset]]
+ * Both, `mapGroupsWithState` and `flatMapGroupsWithState` in 
`KeyValueGroupedDataset`
  * will invoke the user-given function on each group (defined by the grouping 
function in
  * `Dataset.groupByKey()`) while maintaining user-defined per-group state 
between invocations.
  * For a static batch Dataset, the function will be invoked once per group. 
For a streaming
@@ -70,8 +70,8 @@ import 
org.apache.spark.sql.catalyst.plans.logical.LogicalGroupState
  *    `[map|flatMap]GroupsWithState`, but the exact timeout duration/timestamp 
is configurable per
  *    group by calling `setTimeout...()` in `GroupState`.
  *  - Timeouts can be either based on processing time (i.e.
- *    [[GroupStateTimeout.ProcessingTimeTimeout]]) or event time (i.e.
- *    [[GroupStateTimeout.EventTimeTimeout]]).
+ *    `GroupStateTimeout.ProcessingTimeTimeout`) or event time (i.e.
+ *    `GroupStateTimeout.EventTimeTimeout`).
  *  - With `ProcessingTimeTimeout`, the timeout duration can be set by calling
  *    `GroupState.setTimeoutDuration`. The timeout will occur when the clock 
has advanced by the set
  *    duration. Guarantees provided by this timeout with a duration of D ms 
are as follows:
@@ -177,7 +177,7 @@ import 
org.apache.spark.sql.catalyst.plans.logical.LogicalGroupState
  * }}}
  *
  * @tparam S User-defined type of the state to be stored for each group. Must 
be encodable into
- *           Spark SQL types (see [[Encoder]] for more details).
+ *           Spark SQL types (see `Encoder` for more details).
  * @since 2.2.0
  */
 @Experimental
@@ -224,7 +224,7 @@ trait GroupState[S] extends LogicalGroupState[S] {
   /**
    * Set the timeout duration for this key as a string. For example, "1 hour", 
"2 days", etc.
    *
-   * @note, ProcessingTimeTimeout must be enabled in 
`[map/flatmap]GroupsWithStates`.
+   * @note ProcessingTimeTimeout must be enabled in 
`[map/flatmap]GroupsWithStates`.
    */
   @throws[IllegalArgumentException]("if 'duration' is not a valid duration")
   @throws[IllegalStateException]("when state is either not initialized, or 
already removed")
@@ -240,7 +240,7 @@ trait GroupState[S] extends LogicalGroupState[S] {
    * Set the timeout timestamp for this key as milliseconds in epoch time.
    * This timestamp cannot be older than the current watermark.
    *
-   * @note, EventTimeTimeout must be enabled in 
`[map/flatmap]GroupsWithStates`.
+   * @note EventTimeTimeout must be enabled in `[map/flatmap]GroupsWithStates`.
    */
   def setTimeoutTimestamp(timestampMs: Long): Unit
 
@@ -254,7 +254,7 @@ trait GroupState[S] extends LogicalGroupState[S] {
    * The final timestamp (including the additional duration) cannot be older 
than the
    * current watermark.
    *
-   * @note, EventTimeTimeout must be enabled in 
`[map/flatmap]GroupsWithStates`.
+   * @note EventTimeTimeout must be enabled in `[map/flatmap]GroupsWithStates`.
    */
   def setTimeoutTimestamp(timestampMs: Long, additionalDuration: String): Unit
 
@@ -265,7 +265,7 @@ trait GroupState[S] extends LogicalGroupState[S] {
    * Set the timeout timestamp for this key as a java.sql.Date.
    * This timestamp cannot be older than the current watermark.
    *
-   * @note, EventTimeTimeout must be enabled in 
`[map/flatmap]GroupsWithStates`.
+   * @note EventTimeTimeout must be enabled in `[map/flatmap]GroupsWithStates`.
    */
   def setTimeoutTimestamp(timestamp: java.sql.Date): Unit
 
@@ -279,7 +279,7 @@ trait GroupState[S] extends LogicalGroupState[S] {
    * The final timestamp (including the additional duration) cannot be older 
than the
    * current watermark.
    *
-   * @note, EventTimeTimeout must be enabled in 
`[map/flatmap]GroupsWithStates`.
+   * @note EventTimeTimeout must be enabled in `[map/flatmap]GroupsWithStates`.
    */
   def setTimeoutTimestamp(timestamp: java.sql.Date, additionalDuration: 
String): Unit
 }


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org

Reply via email to