Repository: spark
Updated Branches:
  refs/heads/branch-2.2 10c68366e -> 63098dc31


[DOCS] Fix unreachable links in the document

## What changes were proposed in this pull request?

Recently, I found two unreachable links in the document and fixed them.
Because of small changes related to the document, I don't file this issue in 
JIRA but please suggest I should do it if you think it's needed.

## How was this patch tested?

Tested manually.

Author: Kousuke Saruta <saru...@oss.nttdata.co.jp>

Closes #19195 from sarutak/fix-unreachable-link.

(cherry picked from commit 957558235b7537c706c6ab4779655aa57838ebac)
Signed-off-by: Sean Owen <so...@cloudera.com>


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/63098dc3
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/63098dc3
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/63098dc3

Branch: refs/heads/branch-2.2
Commit: 63098dc3170bf4289091d97b7beb63dd0e2356c5
Parents: 10c6836
Author: Kousuke Saruta <saru...@oss.nttdata.co.jp>
Authored: Tue Sep 12 15:07:04 2017 +0100
Committer: Sean Owen <so...@cloudera.com>
Committed: Tue Sep 12 15:07:21 2017 +0100

----------------------------------------------------------------------
 docs/building-spark.md        | 2 +-
 docs/rdd-programming-guide.md | 2 +-
 2 files changed, 2 insertions(+), 2 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/spark/blob/63098dc3/docs/building-spark.md
----------------------------------------------------------------------
diff --git a/docs/building-spark.md b/docs/building-spark.md
index 777635a..14164f1 100644
--- a/docs/building-spark.md
+++ b/docs/building-spark.md
@@ -119,7 +119,7 @@ should run continuous compilation (i.e. wait for changes). 
However, this has not
 extensively. A couple of gotchas to note:
 
 * it only scans the paths `src/main` and `src/test` (see
-[docs](http://scala-tools.org/mvnsites/maven-scala-plugin/usage_cc.html)), so 
it will only work
+[docs](http://davidb.github.io/scala-maven-plugin/example_cc.html)), so it 
will only work
 from within certain submodules that have that structure.
 
 * you'll typically need to run `mvn install` from the project root for 
compilation within

http://git-wip-us.apache.org/repos/asf/spark/blob/63098dc3/docs/rdd-programming-guide.md
----------------------------------------------------------------------
diff --git a/docs/rdd-programming-guide.md b/docs/rdd-programming-guide.md
index 8e6c36b..e3a31b8 100644
--- a/docs/rdd-programming-guide.md
+++ b/docs/rdd-programming-guide.md
@@ -604,7 +604,7 @@ before the `reduce`, which would cause `lineLengths` to be 
saved in memory after
 Spark's API relies heavily on passing functions in the driver program to run 
on the cluster.
 There are two recommended ways to do this:
 
-* [Anonymous function 
syntax](http://docs.scala-lang.org/tutorials/tour/anonymous-function-syntax.html),
+* [Anonymous function 
syntax](http://docs.scala-lang.org/tour/basics.html#functions),
   which can be used for short pieces of code.
 * Static methods in a global singleton object. For example, you can define 
`object MyFunctions` and then
   pass `MyFunctions.func1`, as follows:


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org

Reply via email to