[1/2] [SPARK-1390] Refactoring of matrices backed by RDDs
Repository: spark Updated Branches: refs/heads/master fa0524fd0 - 9689b663a http://git-wip-us.apache.org/repos/asf/spark/blob/9689b663/mllib/src/test/scala/org/apache/spark/mllib/linalg/SVDSuite.scala -- diff --git a/mllib/src/test/scala/org/apache/spark/mllib/linalg/SVDSuite.scala b/mllib/src/test/scala/org/apache/spark/mllib/linalg/SVDSuite.scala deleted file mode 100644 index 20e2b0f..000 --- a/mllib/src/test/scala/org/apache/spark/mllib/linalg/SVDSuite.scala +++ /dev/null @@ -1,194 +0,0 @@ -/* - * Licensed to the Apache Software Foundation (ASF) under one or more - * contributor license agreements. See the NOTICE file distributed with - * this work for additional information regarding copyright ownership. - * The ASF licenses this file to You under the Apache License, Version 2.0 - * (the License); you may not use this file except in compliance with - * the License. You may obtain a copy of the License at - * - *http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an AS IS BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -package org.apache.spark.mllib.linalg - -import scala.util.Random - -import org.scalatest.BeforeAndAfterAll -import org.scalatest.FunSuite - -import org.jblas.{DoubleMatrix, Singular, MatrixFunctions} - -import org.apache.spark.SparkContext -import org.apache.spark.SparkContext._ -import org.apache.spark.rdd.RDD - -import org.apache.spark.mllib.util._ - -import org.jblas._ - -class SVDSuite extends FunSuite with BeforeAndAfterAll { - @transient private var sc: SparkContext = _ - - override def beforeAll() { -sc = new SparkContext(local, test) - } - - override def afterAll() { -sc.stop() -System.clearProperty(spark.driver.port) - } - - val EPSILON = 1e-4 - - // Return jblas matrix from sparse matrix RDD - def getDenseMatrix(matrix: SparseMatrix) : DoubleMatrix = { -val data = matrix.data -val m = matrix.m -val n = matrix.n -val ret = DoubleMatrix.zeros(m, n) -matrix.data.collect().map(x = ret.put(x.i, x.j, x.mval)) -ret - } - - def assertMatrixApproximatelyEquals(a: DoubleMatrix, b: DoubleMatrix) { -assert(a.rows == b.rows a.columns == b.columns, - dimension mismatch: $a.rows vs $b.rows and $a.columns vs $b.columns) -for (i - 0 until a.columns) { - val aCol = a.getColumn(i) - val bCol = b.getColumn(i) - val diff = Math.min(aCol.sub(bCol).norm1, aCol.add(bCol).norm1) - assert(diff EPSILON, matrix mismatch: + diff) -} - } - - test(full rank matrix svd) { -val m = 10 -val n = 3 -val datarr = Array.tabulate(m,n){ (a, b) = - MatrixEntry(a, b, (a + 2).toDouble * (b + 1) / (1 + a + b)) }.flatten -val data = sc.makeRDD(datarr, 3) - -val a = SparseMatrix(data, m, n) - -val decomposed = new SVD().setK(n).compute(a) -val u = decomposed.U -val v = decomposed.V -val s = decomposed.S - -val denseA = getDenseMatrix(a) -val svd = Singular.sparseSVD(denseA) - -val retu = getDenseMatrix(u) -val rets = getDenseMatrix(s) -val retv = getDenseMatrix(v) - - -// check individual decomposition -assertMatrixApproximatelyEquals(retu, svd(0)) -assertMatrixApproximatelyEquals(rets, DoubleMatrix.diag(svd(1))) -assertMatrixApproximatelyEquals(retv, svd(2)) - -// check multiplication guarantee -assertMatrixApproximatelyEquals(retu.mmul(rets).mmul(retv.transpose), denseA) - } - - test(dense full rank matrix svd) { -val m = 10 -val n = 3 -val datarr = Array.tabulate(m,n){ (a, b) = - MatrixEntry(a, b, (a + 2).toDouble * (b + 1) / (1 + a + b)) }.flatten -val data = sc.makeRDD(datarr, 3) - -val a = LAUtils.sparseToTallSkinnyDense(SparseMatrix(data, m, n)) - -val decomposed = new SVD().setK(n).setComputeU(true).compute(a) -val u = LAUtils.denseToSparse(decomposed.U) -val v = decomposed.V -val s = decomposed.S - -val denseA = getDenseMatrix(LAUtils.denseToSparse(a)) -val svd = Singular.sparseSVD(denseA) - -val retu = getDenseMatrix(u) -val rets = DoubleMatrix.diag(new DoubleMatrix(s)) -val retv = new DoubleMatrix(v) - - -// check individual decomposition -assertMatrixApproximatelyEquals(retu, svd(0)) -assertMatrixApproximatelyEquals(rets, DoubleMatrix.diag(svd(1))) -assertMatrixApproximatelyEquals(retv, svd(2)) - -// check multiplication guarantee -assertMatrixApproximatelyEquals(retu.mmul(rets).mmul(retv.transpose), denseA) - } - - test(rank one matrix svd) { -val m = 10 -val n = 3 -val data = sc.makeRDD(Array.tabulate(m, n){ (a,b) = - MatrixEntry(a, b, 1.0) }.flatten ) -val k = 1 - -
svn commit: r5003 - /dev/spark/spark-0.9.1/
Author: tdas Date: Wed Apr 9 19:48:15 2014 New Revision: 5003 Log: Moved Spark 0.9.1 from dev tree to release tree, within incubator/spark as a temporary place until www.apache.org/dist/spark is created. Removed: dev/spark/spark-0.9.1/
svn commit: r1586127 [2/2] - in /spark: ./ news/_posts/ releases/_posts/ site/ site/docs/ site/mllib/ site/news/ site/releases/ site/screencasts/ site/streaming/
Modified: spark/site/releases/spark-release-0-5-0.html URL: http://svn.apache.org/viewvc/spark/site/releases/spark-release-0-5-0.html?rev=1586127r1=1586126r2=1586127view=diff == --- spark/site/releases/spark-release-0-5-0.html (original) +++ spark/site/releases/spark-release-0-5-0.html Wed Apr 9 20:17:50 2014 @@ -124,6 +124,9 @@ h5Latest News/h5 ul class=list-unstyled + lia href=/news/spark-0-9-1-released.htmlSpark 0.9.1 released/a + span class=small(Apr 09, 2014)/span/li + lia href=/news/submit-talks-to-spark-summit-2014.htmlSubmissions and registration open for Spark Summit 2014/a span class=small(Mar 20, 2014)/span/li @@ -133,9 +136,6 @@ lia href=/news/spark-0-9-0-released.htmlSpark 0.9.0 released/a span class=small(Feb 02, 2014)/span/li - lia href=/news/spark-0-8-1-released.htmlSpark 0.8.1 released/a - span class=small(Dec 19, 2013)/span/li - /ul p class=small style=text-align: right;a href=/news/index.htmlArchive/a/p /div Modified: spark/site/releases/spark-release-0-5-1.html URL: http://svn.apache.org/viewvc/spark/site/releases/spark-release-0-5-1.html?rev=1586127r1=1586126r2=1586127view=diff == --- spark/site/releases/spark-release-0-5-1.html (original) +++ spark/site/releases/spark-release-0-5-1.html Wed Apr 9 20:17:50 2014 @@ -124,6 +124,9 @@ h5Latest News/h5 ul class=list-unstyled + lia href=/news/spark-0-9-1-released.htmlSpark 0.9.1 released/a + span class=small(Apr 09, 2014)/span/li + lia href=/news/submit-talks-to-spark-summit-2014.htmlSubmissions and registration open for Spark Summit 2014/a span class=small(Mar 20, 2014)/span/li @@ -133,9 +136,6 @@ lia href=/news/spark-0-9-0-released.htmlSpark 0.9.0 released/a span class=small(Feb 02, 2014)/span/li - lia href=/news/spark-0-8-1-released.htmlSpark 0.8.1 released/a - span class=small(Dec 19, 2013)/span/li - /ul p class=small style=text-align: right;a href=/news/index.htmlArchive/a/p /div Modified: spark/site/releases/spark-release-0-5-2.html URL: http://svn.apache.org/viewvc/spark/site/releases/spark-release-0-5-2.html?rev=1586127r1=1586126r2=1586127view=diff == --- spark/site/releases/spark-release-0-5-2.html (original) +++ spark/site/releases/spark-release-0-5-2.html Wed Apr 9 20:17:50 2014 @@ -124,6 +124,9 @@ h5Latest News/h5 ul class=list-unstyled + lia href=/news/spark-0-9-1-released.htmlSpark 0.9.1 released/a + span class=small(Apr 09, 2014)/span/li + lia href=/news/submit-talks-to-spark-summit-2014.htmlSubmissions and registration open for Spark Summit 2014/a span class=small(Mar 20, 2014)/span/li @@ -133,9 +136,6 @@ lia href=/news/spark-0-9-0-released.htmlSpark 0.9.0 released/a span class=small(Feb 02, 2014)/span/li - lia href=/news/spark-0-8-1-released.htmlSpark 0.8.1 released/a - span class=small(Dec 19, 2013)/span/li - /ul p class=small style=text-align: right;a href=/news/index.htmlArchive/a/p /div Modified: spark/site/releases/spark-release-0-6-0.html URL: http://svn.apache.org/viewvc/spark/site/releases/spark-release-0-6-0.html?rev=1586127r1=1586126r2=1586127view=diff == --- spark/site/releases/spark-release-0-6-0.html (original) +++ spark/site/releases/spark-release-0-6-0.html Wed Apr 9 20:17:50 2014 @@ -124,6 +124,9 @@ h5Latest News/h5 ul class=list-unstyled + lia href=/news/spark-0-9-1-released.htmlSpark 0.9.1 released/a + span class=small(Apr 09, 2014)/span/li + lia href=/news/submit-talks-to-spark-summit-2014.htmlSubmissions and registration open for Spark Summit 2014/a span class=small(Mar 20, 2014)/span/li @@ -133,9 +136,6 @@ lia href=/news/spark-0-9-0-released.htmlSpark 0.9.0 released/a span class=small(Feb 02, 2014)/span/li - lia href=/news/spark-0-8-1-released.htmlSpark 0.8.1 released/a - span class=small(Dec 19, 2013)/span/li - /ul p class=small style=text-align: right;a href=/news/index.htmlArchive/a/p /div Modified: spark/site/releases/spark-release-0-6-1.html URL: http://svn.apache.org/viewvc/spark/site/releases/spark-release-0-6-1.html?rev=1586127r1=1586126r2=1586127view=diff == ---
git commit: SPARK-1407 drain event queue before stopping event logger
Repository: spark Updated Branches: refs/heads/master bde9cc11f - eb5f2b642 SPARK-1407 drain event queue before stopping event logger Author: Kan Zhang kzh...@apache.org Closes #366 from kanzhang/SPARK-1407 and squashes the following commits: cd0629f [Kan Zhang] code refactoring and adding test b073ee6 [Kan Zhang] SPARK-1407 drain event queue before stopping event logger Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/eb5f2b64 Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/eb5f2b64 Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/eb5f2b64 Branch: refs/heads/master Commit: eb5f2b64230faa69a53815cb61bcc87aeb233d20 Parents: bde9cc1 Author: Kan Zhang kzh...@apache.org Authored: Wed Apr 9 15:24:33 2014 -0700 Committer: Patrick Wendell pwend...@gmail.com Committed: Wed Apr 9 15:25:29 2014 -0700 -- .../scala/org/apache/spark/SparkContext.scala | 4 +- .../spark/scheduler/LiveListenerBus.scala | 32 -- .../spark/scheduler/SparkListenerSuite.scala| 45 .../org/apache/spark/examples/SparkHdfsLR.scala | 2 +- 4 files changed, 67 insertions(+), 16 deletions(-) -- http://git-wip-us.apache.org/repos/asf/spark/blob/eb5f2b64/core/src/main/scala/org/apache/spark/SparkContext.scala -- diff --git a/core/src/main/scala/org/apache/spark/SparkContext.scala b/core/src/main/scala/org/apache/spark/SparkContext.scala index f775051..7630523 100644 --- a/core/src/main/scala/org/apache/spark/SparkContext.scala +++ b/core/src/main/scala/org/apache/spark/SparkContext.scala @@ -931,7 +931,6 @@ class SparkContext(config: SparkConf) extends Logging { /** Shut down the SparkContext. */ def stop() { ui.stop() -eventLogger.foreach(_.stop()) // Do this only if not stopped already - best case effort. // prevent NPE if stopped more than once. val dagSchedulerCopy = dagScheduler @@ -940,13 +939,14 @@ class SparkContext(config: SparkConf) extends Logging { metadataCleaner.cancel() cleaner.foreach(_.stop()) dagSchedulerCopy.stop() - listenerBus.stop() taskScheduler = null // TODO: Cache.stop()? env.stop() SparkEnv.set(null) ShuffleMapTask.clearCache() ResultTask.clearCache() + listenerBus.stop() + eventLogger.foreach(_.stop()) logInfo(Successfully stopped SparkContext) } else { logInfo(SparkContext already stopped) http://git-wip-us.apache.org/repos/asf/spark/blob/eb5f2b64/core/src/main/scala/org/apache/spark/scheduler/LiveListenerBus.scala -- diff --git a/core/src/main/scala/org/apache/spark/scheduler/LiveListenerBus.scala b/core/src/main/scala/org/apache/spark/scheduler/LiveListenerBus.scala index 353a486..76f3e32 100644 --- a/core/src/main/scala/org/apache/spark/scheduler/LiveListenerBus.scala +++ b/core/src/main/scala/org/apache/spark/scheduler/LiveListenerBus.scala @@ -36,6 +36,22 @@ private[spark] class LiveListenerBus extends SparkListenerBus with Logging { private val eventQueue = new LinkedBlockingQueue[SparkListenerEvent](EVENT_QUEUE_CAPACITY) private var queueFullErrorMessageLogged = false private var started = false + private val listenerThread = new Thread(SparkListenerBus) { +setDaemon(true) +override def run() { + while (true) { +val event = eventQueue.take +if (event == SparkListenerShutdown) { + // Get out of the while loop and shutdown the daemon thread + return +} +postToAll(event) + } +} + } + + // Exposed for testing + @volatile private[spark] var stopCalled = false /** * Start sending events to attached listeners. @@ -48,20 +64,8 @@ private[spark] class LiveListenerBus extends SparkListenerBus with Logging { if (started) { throw new IllegalStateException(Listener bus already started!) } +listenerThread.start() started = true -new Thread(SparkListenerBus) { - setDaemon(true) - override def run() { -while (true) { - val event = eventQueue.take - if (event == SparkListenerShutdown) { -// Get out of the while loop and shutdown the daemon thread -return - } - postToAll(event) -} - } -}.start() } def post(event: SparkListenerEvent) { @@ -93,9 +97,11 @@ private[spark] class LiveListenerBus extends SparkListenerBus with Logging { } def stop() { +stopCalled = true if (!started) { throw new IllegalStateException(Attempted to stop a listener bus that has not yet started!) } post(SparkListenerShutdown) +listenerThread.join() } }
svn commit: r1586167 [2/2] - in /spark/site/docs/0.9.1: ./ api/pyspark/
Modified: spark/site/docs/0.9.1/tuning.html URL: http://svn.apache.org/viewvc/spark/site/docs/0.9.1/tuning.html?rev=1586167r1=1586166r2=1586167view=diff == --- spark/site/docs/0.9.1/tuning.html (original) +++ spark/site/docs/0.9.1/tuning.html Wed Apr 9 22:56:47 2014 @@ -25,6 +25,19 @@ link rel=stylesheet href=css/pygments-default.css +!-- Google analytics script -- +script type=text/javascript + var _gaq = _gaq || []; + _gaq.push(['_setAccount', 'UA-32518208-1']); + _gaq.push(['_trackPageview']); + + (function() { +var ga = document.createElement('script'); ga.type = 'text/javascript'; ga.async = true; +ga.src = ('https:' == document.location.protocol ? 'https://ssl' : 'http://www') + '.google-analytics.com/ga.js'; +var s = document.getElementsByTagName('script')[0]; s.parentNode.insertBefore(ga, s); + })(); +/script + /head body
svn commit: r1586169 - in /spark: releases/_posts/2014-04-09-spark-release-0-9-1.md site/docs/0.9.1/0.9.1-docs.tgz site/releases/spark-release-0-9-1.html
Author: tdas Date: Wed Apr 9 23:19:13 2014 New Revision: 1586169 URL: http://svn.apache.org/r1586169 Log: Updated 0.9.1 release notes Removed: spark/site/docs/0.9.1/0.9.1-docs.tgz Modified: spark/releases/_posts/2014-04-09-spark-release-0-9-1.md spark/site/releases/spark-release-0-9-1.html Modified: spark/releases/_posts/2014-04-09-spark-release-0-9-1.md URL: http://svn.apache.org/viewvc/spark/releases/_posts/2014-04-09-spark-release-0-9-1.md?rev=1586169r1=1586168r2=1586169view=diff == --- spark/releases/_posts/2014-04-09-spark-release-0-9-1.md (original) +++ spark/releases/_posts/2014-04-09-spark-release-0-9-1.md Wed Apr 9 23:19:13 2014 @@ -59,7 +59,7 @@ Several bug fixes were made to YARN depl * Fixed bug in Python de-pickling [[SPARK-1135](https://issues.apache.org/jira/browse/SPARK-1135)] * Fixed bug in serialization of strings longer than 64K [[SPARK-1043](https://issues.apache.org/jira/browse/SPARK-1043)] * Fixed bug that made jobs hang when base file is not available [[SPARK-1025](https://issues.apache.org/jira/browse/SPARK-1025)] -* Added Missing RDD operations to PySpark - top, zip, foldByKey and setName +* Added Missing RDD operations to PySpark - top, zip, foldByKey, repartition, coallesce, getStorageLevel, setName and toDebugString ### Improvements to documentation * Streaming: Added documentation on running streaming application from spark-shell Modified: spark/site/releases/spark-release-0-9-1.html URL: http://svn.apache.org/viewvc/spark/site/releases/spark-release-0-9-1.html?rev=1586169r1=1586168r2=1586169view=diff == --- spark/site/releases/spark-release-0-9-1.html (original) +++ spark/site/releases/spark-release-0-9-1.html Wed Apr 9 23:19:13 2014 @@ -217,7 +217,7 @@ liFixed bug in Python de-pickling [a href=https://issues.apache.org/jira/browse/SPARK-1135;SPARK-1135/a]/li liFixed bug in serialization of strings longer than 64K [a href=https://issues.apache.org/jira/browse/SPARK-1043;SPARK-1043/a] /li liFixed bug that made jobs hang when base file is not available [a href=https://issues.apache.org/jira/browse/SPARK-1025;SPARK-1025/a] /li - liAdded Missing RDD operations to PySpark - top, zip, foldByKey and setName /li + liAdded Missing RDD operations to PySpark - top, zip, foldByKey, repartition, coallesce, getStorageLevel, setName and toDebugString/li /ul h3 id=improvements-to-documentationImprovements to documentation/h3
git commit: [SPARK-1357 (fix)] remove empty line after :: DeveloperApi/Experimental ::
Repository: spark Updated Branches: refs/heads/master eb5f2b642 - 0adc932ad [SPARK-1357 (fix)] remove empty line after :: DeveloperApi/Experimental :: Remove empty line after :: DeveloperApi/Experimental :: in comments to make the original doc show up in the preview of the generated html docs. Thanks @andrewor14 ! Author: Xiangrui Meng m...@databricks.com Closes #373 from mengxr/api and squashes the following commits: 9c35bdc [Xiangrui Meng] remove the empty line after :: DeveloperApi/Experimental :: Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/0adc932a Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/0adc932a Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/0adc932a Branch: refs/heads/master Commit: 0adc932add413a1754107b21d5ecfb38c0c3a4eb Parents: eb5f2b6 Author: Xiangrui Meng m...@databricks.com Authored: Wed Apr 9 17:08:17 2014 -0700 Committer: Patrick Wendell pwend...@gmail.com Committed: Wed Apr 9 17:08:17 2014 -0700 -- .../spark/mllib/api/python/PythonMLLibAPI.scala | 1 - .../spark/mllib/classification/NaiveBayes.scala | 1 - .../apache/spark/mllib/clustering/KMeans.scala | 4 -- .../linalg/distributed/CoordinateMatrix.scala | 1 - .../linalg/distributed/IndexedRowMatrix.scala | 2 - .../mllib/linalg/distributed/RowMatrix.scala| 1 - .../spark/mllib/optimization/Gradient.scala | 4 -- .../mllib/optimization/GradientDescent.scala| 2 - .../spark/mllib/optimization/Optimizer.scala| 1 - .../spark/mllib/optimization/Updater.scala | 4 -- .../apache/spark/mllib/recommendation/ALS.scala | 1 - .../MatrixFactorizationModel.scala | 1 - .../regression/GeneralizedLinearAlgorithm.scala | 1 - .../apache/spark/mllib/tree/DecisionTree.scala | 1 - .../spark/mllib/tree/configuration/Algo.scala | 1 - .../mllib/tree/configuration/FeatureType.scala | 1 - .../tree/configuration/QuantileStrategy.scala | 1 - .../mllib/tree/configuration/Strategy.scala | 1 - .../spark/mllib/tree/impurity/Entropy.scala | 2 - .../apache/spark/mllib/tree/impurity/Gini.scala | 2 - .../spark/mllib/tree/impurity/Impurity.scala| 3 -- .../spark/mllib/tree/impurity/Variance.scala| 2 - .../mllib/tree/model/DecisionTreeModel.scala| 1 - .../mllib/tree/model/InformationGainStats.scala | 1 - .../apache/spark/mllib/tree/model/Node.scala| 1 - .../apache/spark/mllib/tree/model/Split.scala | 1 - .../spark/mllib/util/DataValidators.scala | 1 - .../spark/mllib/util/KMeansDataGenerator.scala | 1 - .../spark/mllib/util/LinearDataGenerator.scala | 1 - .../util/LogisticRegressionDataGenerator.scala | 1 - .../spark/mllib/util/MFDataGenerator.scala | 43 ++-- .../org/apache/spark/mllib/util/MLUtils.scala | 2 - .../spark/mllib/util/SVMDataGenerator.scala | 1 - 33 files changed, 21 insertions(+), 71 deletions(-) -- http://git-wip-us.apache.org/repos/asf/spark/blob/0adc932a/mllib/src/main/scala/org/apache/spark/mllib/api/python/PythonMLLibAPI.scala -- diff --git a/mllib/src/main/scala/org/apache/spark/mllib/api/python/PythonMLLibAPI.scala b/mllib/src/main/scala/org/apache/spark/mllib/api/python/PythonMLLibAPI.scala index ae27c57..a6c049e 100644 --- a/mllib/src/main/scala/org/apache/spark/mllib/api/python/PythonMLLibAPI.scala +++ b/mllib/src/main/scala/org/apache/spark/mllib/api/python/PythonMLLibAPI.scala @@ -30,7 +30,6 @@ import org.apache.spark.rdd.RDD /** * :: DeveloperApi :: - * * The Java stubs necessary for the Python mllib bindings. */ @DeveloperApi http://git-wip-us.apache.org/repos/asf/spark/blob/0adc932a/mllib/src/main/scala/org/apache/spark/mllib/classification/NaiveBayes.scala -- diff --git a/mllib/src/main/scala/org/apache/spark/mllib/classification/NaiveBayes.scala b/mllib/src/main/scala/org/apache/spark/mllib/classification/NaiveBayes.scala index 5a45f12..1865885 100644 --- a/mllib/src/main/scala/org/apache/spark/mllib/classification/NaiveBayes.scala +++ b/mllib/src/main/scala/org/apache/spark/mllib/classification/NaiveBayes.scala @@ -29,7 +29,6 @@ import org.apache.spark.rdd.RDD /** * :: Experimental :: - * * Model for Naive Bayes Classifiers. * * @param labels list of labels http://git-wip-us.apache.org/repos/asf/spark/blob/0adc932a/mllib/src/main/scala/org/apache/spark/mllib/clustering/KMeans.scala -- diff --git a/mllib/src/main/scala/org/apache/spark/mllib/clustering/KMeans.scala b/mllib/src/main/scala/org/apache/spark/mllib/clustering/KMeans.scala index 8f565eb..90cf852 100644 ---
svn commit: r1586186 - /spark/site/examples.html
Author: andrew Date: Thu Apr 10 00:52:59 2014 New Revision: 1586186 URL: http://svn.apache.org/r1586186 Log: adding changed versions of compiled files as well Modified: spark/site/examples.html Modified: spark/site/examples.html URL: http://svn.apache.org/viewvc/spark/site/examples.html?rev=1586186r1=1586185r2=1586186view=diff == --- spark/site/examples.html (original) +++ spark/site/examples.html Thu Apr 10 00:52:59 2014 @@ -409,8 +409,8 @@ previous ones, and emactions/em, whi pMany additional examples are distributed with Spark:/p ul - liBasic Spark: a href=https://github.com/apache/incubator-spark/tree/master/examples/src/main/scala/org/apache/spark/examples;Scala examples/a, a href=https://github.com/apache/incubator-spark/tree/master/examples/src/main/java/org/apache/spark/examples;Java examples/a, a href=https://github.com/apache/incubator-spark/tree/master/python/examples;Python examples/a/li - liSpark Streaming: a href=https://github.com/apache/incubator-spark/tree/master/examples/src/main/scala/org/apache/spark/streaming/examples;Scala examples/a, a href=https://github.com/apache/incubator-spark/tree/master/examples/src/main/java/org/apache/spark/streaming/examples;Java examples/a/li + liBasic Spark: a href=https://github.com/apache/spark/tree/master/examples/src/main/scala/org/apache/spark/examples;Scala examples/a, a href=https://github.com/apache/spark/tree/master/examples/src/main/java/org/apache/spark/examples;Java examples/a, a href=https://github.com/apache/spark/tree/master/python/examples;Python examples/a/li + liSpark Streaming: a href=https://github.com/apache/spark/tree/master/examples/src/main/scala/org/apache/spark/streaming/examples;Scala examples/a, a href=https://github.com/apache/spark/tree/master/examples/src/main/java/org/apache/spark/streaming/examples;Java examples/a/li /ul
svn commit: r1586197 - in /spark: releases/_posts/2014-04-09-spark-release-0-9-1.md site/releases/spark-release-0-9-1.html
Author: tdas Date: Thu Apr 10 02:17:44 2014 New Revision: 1586197 URL: http://svn.apache.org/r1586197 Log: Fixed typo Modified: spark/releases/_posts/2014-04-09-spark-release-0-9-1.md spark/site/releases/spark-release-0-9-1.html Modified: spark/releases/_posts/2014-04-09-spark-release-0-9-1.md URL: http://svn.apache.org/viewvc/spark/releases/_posts/2014-04-09-spark-release-0-9-1.md?rev=1586197r1=1586196r2=1586197view=diff == --- spark/releases/_posts/2014-04-09-spark-release-0-9-1.md (original) +++ spark/releases/_posts/2014-04-09-spark-release-0-9-1.md Thu Apr 10 02:17:44 2014 @@ -59,7 +59,7 @@ Several bug fixes were made to YARN depl * Fixed bug in Python de-pickling [[SPARK-1135](https://issues.apache.org/jira/browse/SPARK-1135)] * Fixed bug in serialization of strings longer than 64K [[SPARK-1043](https://issues.apache.org/jira/browse/SPARK-1043)] * Fixed bug that made jobs hang when base file is not available [[SPARK-1025](https://issues.apache.org/jira/browse/SPARK-1025)] -* Added Missing RDD operations to PySpark - top, zip, foldByKey, repartition, coallesce, getStorageLevel, setName and toDebugString +* Added Missing RDD operations to PySpark - top, zip, foldByKey, repartition, coalesce, getStorageLevel, setName and toDebugString ### Improvements to documentation * Streaming: Added documentation on running streaming application from spark-shell Modified: spark/site/releases/spark-release-0-9-1.html URL: http://svn.apache.org/viewvc/spark/site/releases/spark-release-0-9-1.html?rev=1586197r1=1586196r2=1586197view=diff == --- spark/site/releases/spark-release-0-9-1.html (original) +++ spark/site/releases/spark-release-0-9-1.html Thu Apr 10 02:17:44 2014 @@ -217,7 +217,7 @@ liFixed bug in Python de-pickling [a href=https://issues.apache.org/jira/browse/SPARK-1135;SPARK-1135/a]/li liFixed bug in serialization of strings longer than 64K [a href=https://issues.apache.org/jira/browse/SPARK-1043;SPARK-1043/a] /li liFixed bug that made jobs hang when base file is not available [a href=https://issues.apache.org/jira/browse/SPARK-1025;SPARK-1025/a] /li - liAdded Missing RDD operations to PySpark - top, zip, foldByKey, repartition, coallesce, getStorageLevel, setName and toDebugString/li + liAdded Missing RDD operations to PySpark - top, zip, foldByKey, repartition, coalesce, getStorageLevel, setName and toDebugString/li /ul h3 id=improvements-to-documentationImprovements to documentation/h3