[1/2] [SPARK-1390] Refactoring of matrices backed by RDDs

2014-04-09 Thread pwendell
Repository: spark
Updated Branches:
  refs/heads/master fa0524fd0 - 9689b663a


http://git-wip-us.apache.org/repos/asf/spark/blob/9689b663/mllib/src/test/scala/org/apache/spark/mllib/linalg/SVDSuite.scala
--
diff --git a/mllib/src/test/scala/org/apache/spark/mllib/linalg/SVDSuite.scala 
b/mllib/src/test/scala/org/apache/spark/mllib/linalg/SVDSuite.scala
deleted file mode 100644
index 20e2b0f..000
--- a/mllib/src/test/scala/org/apache/spark/mllib/linalg/SVDSuite.scala
+++ /dev/null
@@ -1,194 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one or more
- * contributor license agreements.  See the NOTICE file distributed with
- * this work for additional information regarding copyright ownership.
- * The ASF licenses this file to You under the Apache License, Version 2.0
- * (the License); you may not use this file except in compliance with
- * the License.  You may obtain a copy of the License at
- *
- *http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an AS IS BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-package org.apache.spark.mllib.linalg
-
-import scala.util.Random
-
-import org.scalatest.BeforeAndAfterAll
-import org.scalatest.FunSuite
-
-import org.jblas.{DoubleMatrix, Singular, MatrixFunctions}
-
-import org.apache.spark.SparkContext
-import org.apache.spark.SparkContext._
-import org.apache.spark.rdd.RDD
-
-import org.apache.spark.mllib.util._
-
-import org.jblas._
-
-class SVDSuite extends FunSuite with BeforeAndAfterAll {
-  @transient private var sc: SparkContext = _
-
-  override def beforeAll() {
-sc = new SparkContext(local, test)
-  }
-
-  override def afterAll() {
-sc.stop()
-System.clearProperty(spark.driver.port)
-  }
-
-  val EPSILON = 1e-4
-
-  // Return jblas matrix from sparse matrix RDD
-  def getDenseMatrix(matrix: SparseMatrix) : DoubleMatrix = {
-val data = matrix.data
-val m = matrix.m
-val n = matrix.n
-val ret = DoubleMatrix.zeros(m, n)
-matrix.data.collect().map(x = ret.put(x.i, x.j, x.mval))
-ret
-  }
-
-  def assertMatrixApproximatelyEquals(a: DoubleMatrix, b: DoubleMatrix) {
-assert(a.rows == b.rows  a.columns == b.columns,
-  dimension mismatch: $a.rows vs $b.rows and $a.columns vs $b.columns)
-for (i - 0 until a.columns) {
-  val aCol = a.getColumn(i)
-  val bCol = b.getColumn(i)
-  val diff = Math.min(aCol.sub(bCol).norm1, aCol.add(bCol).norm1)
-  assert(diff  EPSILON, matrix mismatch:  + diff)
-}
-  }
-
-  test(full rank matrix svd) {
-val m = 10
-val n = 3
-val datarr = Array.tabulate(m,n){ (a, b) =
-  MatrixEntry(a, b, (a + 2).toDouble * (b + 1) / (1 + a + b)) }.flatten
-val data = sc.makeRDD(datarr, 3)
-
-val a = SparseMatrix(data, m, n)
-
-val decomposed = new SVD().setK(n).compute(a)
-val u = decomposed.U
-val v = decomposed.V
-val s = decomposed.S
-
-val denseA = getDenseMatrix(a)
-val svd = Singular.sparseSVD(denseA)
-
-val retu = getDenseMatrix(u)
-val rets = getDenseMatrix(s)
-val retv = getDenseMatrix(v)
- 
- 
-// check individual decomposition  
-assertMatrixApproximatelyEquals(retu, svd(0))
-assertMatrixApproximatelyEquals(rets, DoubleMatrix.diag(svd(1)))
-assertMatrixApproximatelyEquals(retv, svd(2))
-
-// check multiplication guarantee
-assertMatrixApproximatelyEquals(retu.mmul(rets).mmul(retv.transpose), 
denseA)  
-  }
-
- test(dense full rank matrix svd) {
-val m = 10
-val n = 3
-val datarr = Array.tabulate(m,n){ (a, b) =
-  MatrixEntry(a, b, (a + 2).toDouble * (b + 1) / (1 + a + b)) }.flatten
-val data = sc.makeRDD(datarr, 3)
-
-val a = LAUtils.sparseToTallSkinnyDense(SparseMatrix(data, m, n))
-
-val decomposed = new SVD().setK(n).setComputeU(true).compute(a)
-val u = LAUtils.denseToSparse(decomposed.U)
-val v = decomposed.V
-val s = decomposed.S
-
-val denseA = getDenseMatrix(LAUtils.denseToSparse(a))
-val svd = Singular.sparseSVD(denseA)
-
-val retu = getDenseMatrix(u)
-val rets = DoubleMatrix.diag(new DoubleMatrix(s))
-val retv = new DoubleMatrix(v)
-
-
-// check individual decomposition  
-assertMatrixApproximatelyEquals(retu, svd(0))
-assertMatrixApproximatelyEquals(rets, DoubleMatrix.diag(svd(1)))
-assertMatrixApproximatelyEquals(retv, svd(2))
-
-// check multiplication guarantee
-assertMatrixApproximatelyEquals(retu.mmul(rets).mmul(retv.transpose), 
denseA)
-  }
-
- test(rank one matrix svd) {
-val m = 10
-val n = 3   
-val data = sc.makeRDD(Array.tabulate(m, n){ (a,b) =
-  MatrixEntry(a, b, 1.0) }.flatten )
-val k = 1
-
-

svn commit: r5003 - /dev/spark/spark-0.9.1/

2014-04-09 Thread tdas
Author: tdas
Date: Wed Apr  9 19:48:15 2014
New Revision: 5003

Log:
Moved Spark 0.9.1 from dev tree to release tree, within incubator/spark as a 
temporary place
until www.apache.org/dist/spark is created.

Removed:
dev/spark/spark-0.9.1/



svn commit: r1586127 [2/2] - in /spark: ./ news/_posts/ releases/_posts/ site/ site/docs/ site/mllib/ site/news/ site/releases/ site/screencasts/ site/streaming/

2014-04-09 Thread tdas
Modified: spark/site/releases/spark-release-0-5-0.html
URL: 
http://svn.apache.org/viewvc/spark/site/releases/spark-release-0-5-0.html?rev=1586127r1=1586126r2=1586127view=diff
==
--- spark/site/releases/spark-release-0-5-0.html (original)
+++ spark/site/releases/spark-release-0-5-0.html Wed Apr  9 20:17:50 2014
@@ -124,6 +124,9 @@
   h5Latest News/h5
   ul class=list-unstyled
 
+  lia href=/news/spark-0-9-1-released.htmlSpark 0.9.1 
released/a
+  span class=small(Apr 09, 2014)/span/li
+
   lia 
href=/news/submit-talks-to-spark-summit-2014.htmlSubmissions and 
registration open for Spark Summit 2014/a
   span class=small(Mar 20, 2014)/span/li
 
@@ -133,9 +136,6 @@
   lia href=/news/spark-0-9-0-released.htmlSpark 0.9.0 
released/a
   span class=small(Feb 02, 2014)/span/li
 
-  lia href=/news/spark-0-8-1-released.htmlSpark 0.8.1 
released/a
-  span class=small(Dec 19, 2013)/span/li
-
   /ul
   p class=small style=text-align: right;a 
href=/news/index.htmlArchive/a/p
 /div

Modified: spark/site/releases/spark-release-0-5-1.html
URL: 
http://svn.apache.org/viewvc/spark/site/releases/spark-release-0-5-1.html?rev=1586127r1=1586126r2=1586127view=diff
==
--- spark/site/releases/spark-release-0-5-1.html (original)
+++ spark/site/releases/spark-release-0-5-1.html Wed Apr  9 20:17:50 2014
@@ -124,6 +124,9 @@
   h5Latest News/h5
   ul class=list-unstyled
 
+  lia href=/news/spark-0-9-1-released.htmlSpark 0.9.1 
released/a
+  span class=small(Apr 09, 2014)/span/li
+
   lia 
href=/news/submit-talks-to-spark-summit-2014.htmlSubmissions and 
registration open for Spark Summit 2014/a
   span class=small(Mar 20, 2014)/span/li
 
@@ -133,9 +136,6 @@
   lia href=/news/spark-0-9-0-released.htmlSpark 0.9.0 
released/a
   span class=small(Feb 02, 2014)/span/li
 
-  lia href=/news/spark-0-8-1-released.htmlSpark 0.8.1 
released/a
-  span class=small(Dec 19, 2013)/span/li
-
   /ul
   p class=small style=text-align: right;a 
href=/news/index.htmlArchive/a/p
 /div

Modified: spark/site/releases/spark-release-0-5-2.html
URL: 
http://svn.apache.org/viewvc/spark/site/releases/spark-release-0-5-2.html?rev=1586127r1=1586126r2=1586127view=diff
==
--- spark/site/releases/spark-release-0-5-2.html (original)
+++ spark/site/releases/spark-release-0-5-2.html Wed Apr  9 20:17:50 2014
@@ -124,6 +124,9 @@
   h5Latest News/h5
   ul class=list-unstyled
 
+  lia href=/news/spark-0-9-1-released.htmlSpark 0.9.1 
released/a
+  span class=small(Apr 09, 2014)/span/li
+
   lia 
href=/news/submit-talks-to-spark-summit-2014.htmlSubmissions and 
registration open for Spark Summit 2014/a
   span class=small(Mar 20, 2014)/span/li
 
@@ -133,9 +136,6 @@
   lia href=/news/spark-0-9-0-released.htmlSpark 0.9.0 
released/a
   span class=small(Feb 02, 2014)/span/li
 
-  lia href=/news/spark-0-8-1-released.htmlSpark 0.8.1 
released/a
-  span class=small(Dec 19, 2013)/span/li
-
   /ul
   p class=small style=text-align: right;a 
href=/news/index.htmlArchive/a/p
 /div

Modified: spark/site/releases/spark-release-0-6-0.html
URL: 
http://svn.apache.org/viewvc/spark/site/releases/spark-release-0-6-0.html?rev=1586127r1=1586126r2=1586127view=diff
==
--- spark/site/releases/spark-release-0-6-0.html (original)
+++ spark/site/releases/spark-release-0-6-0.html Wed Apr  9 20:17:50 2014
@@ -124,6 +124,9 @@
   h5Latest News/h5
   ul class=list-unstyled
 
+  lia href=/news/spark-0-9-1-released.htmlSpark 0.9.1 
released/a
+  span class=small(Apr 09, 2014)/span/li
+
   lia 
href=/news/submit-talks-to-spark-summit-2014.htmlSubmissions and 
registration open for Spark Summit 2014/a
   span class=small(Mar 20, 2014)/span/li
 
@@ -133,9 +136,6 @@
   lia href=/news/spark-0-9-0-released.htmlSpark 0.9.0 
released/a
   span class=small(Feb 02, 2014)/span/li
 
-  lia href=/news/spark-0-8-1-released.htmlSpark 0.8.1 
released/a
-  span class=small(Dec 19, 2013)/span/li
-
   /ul
   p class=small style=text-align: right;a 
href=/news/index.htmlArchive/a/p
 /div

Modified: spark/site/releases/spark-release-0-6-1.html
URL: 
http://svn.apache.org/viewvc/spark/site/releases/spark-release-0-6-1.html?rev=1586127r1=1586126r2=1586127view=diff
==
--- 

git commit: SPARK-1407 drain event queue before stopping event logger

2014-04-09 Thread pwendell
Repository: spark
Updated Branches:
  refs/heads/master bde9cc11f - eb5f2b642


SPARK-1407 drain event queue before stopping event logger

Author: Kan Zhang kzh...@apache.org

Closes #366 from kanzhang/SPARK-1407 and squashes the following commits:

cd0629f [Kan Zhang] code refactoring and adding test
b073ee6 [Kan Zhang] SPARK-1407 drain event queue before stopping event logger


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/eb5f2b64
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/eb5f2b64
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/eb5f2b64

Branch: refs/heads/master
Commit: eb5f2b64230faa69a53815cb61bcc87aeb233d20
Parents: bde9cc1
Author: Kan Zhang kzh...@apache.org
Authored: Wed Apr 9 15:24:33 2014 -0700
Committer: Patrick Wendell pwend...@gmail.com
Committed: Wed Apr 9 15:25:29 2014 -0700

--
 .../scala/org/apache/spark/SparkContext.scala   |  4 +-
 .../spark/scheduler/LiveListenerBus.scala   | 32 --
 .../spark/scheduler/SparkListenerSuite.scala| 45 
 .../org/apache/spark/examples/SparkHdfsLR.scala |  2 +-
 4 files changed, 67 insertions(+), 16 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/eb5f2b64/core/src/main/scala/org/apache/spark/SparkContext.scala
--
diff --git a/core/src/main/scala/org/apache/spark/SparkContext.scala 
b/core/src/main/scala/org/apache/spark/SparkContext.scala
index f775051..7630523 100644
--- a/core/src/main/scala/org/apache/spark/SparkContext.scala
+++ b/core/src/main/scala/org/apache/spark/SparkContext.scala
@@ -931,7 +931,6 @@ class SparkContext(config: SparkConf) extends Logging {
   /** Shut down the SparkContext. */
   def stop() {
 ui.stop()
-eventLogger.foreach(_.stop())
 // Do this only if not stopped already - best case effort.
 // prevent NPE if stopped more than once.
 val dagSchedulerCopy = dagScheduler
@@ -940,13 +939,14 @@ class SparkContext(config: SparkConf) extends Logging {
   metadataCleaner.cancel()
   cleaner.foreach(_.stop())
   dagSchedulerCopy.stop()
-  listenerBus.stop()
   taskScheduler = null
   // TODO: Cache.stop()?
   env.stop()
   SparkEnv.set(null)
   ShuffleMapTask.clearCache()
   ResultTask.clearCache()
+  listenerBus.stop()
+  eventLogger.foreach(_.stop())
   logInfo(Successfully stopped SparkContext)
 } else {
   logInfo(SparkContext already stopped)

http://git-wip-us.apache.org/repos/asf/spark/blob/eb5f2b64/core/src/main/scala/org/apache/spark/scheduler/LiveListenerBus.scala
--
diff --git 
a/core/src/main/scala/org/apache/spark/scheduler/LiveListenerBus.scala 
b/core/src/main/scala/org/apache/spark/scheduler/LiveListenerBus.scala
index 353a486..76f3e32 100644
--- a/core/src/main/scala/org/apache/spark/scheduler/LiveListenerBus.scala
+++ b/core/src/main/scala/org/apache/spark/scheduler/LiveListenerBus.scala
@@ -36,6 +36,22 @@ private[spark] class LiveListenerBus extends 
SparkListenerBus with Logging {
   private val eventQueue = new 
LinkedBlockingQueue[SparkListenerEvent](EVENT_QUEUE_CAPACITY)
   private var queueFullErrorMessageLogged = false
   private var started = false
+  private val listenerThread = new Thread(SparkListenerBus) {
+setDaemon(true)
+override def run() {
+  while (true) {
+val event = eventQueue.take
+if (event == SparkListenerShutdown) {
+  // Get out of the while loop and shutdown the daemon thread
+  return
+}
+postToAll(event)
+  }
+}
+  }
+
+  // Exposed for testing
+  @volatile private[spark] var stopCalled = false
 
   /**
* Start sending events to attached listeners.
@@ -48,20 +64,8 @@ private[spark] class LiveListenerBus extends 
SparkListenerBus with Logging {
 if (started) {
   throw new IllegalStateException(Listener bus already started!)
 }
+listenerThread.start()
 started = true
-new Thread(SparkListenerBus) {
-  setDaemon(true)
-  override def run() {
-while (true) {
-  val event = eventQueue.take
-  if (event == SparkListenerShutdown) {
-// Get out of the while loop and shutdown the daemon thread
-return
-  }
-  postToAll(event)
-}
-  }
-}.start()
   }
 
   def post(event: SparkListenerEvent) {
@@ -93,9 +97,11 @@ private[spark] class LiveListenerBus extends 
SparkListenerBus with Logging {
   }
 
   def stop() {
+stopCalled = true
 if (!started) {
   throw new IllegalStateException(Attempted to stop a listener bus that 
has not yet started!)
 }
 post(SparkListenerShutdown)
+listenerThread.join()
   }
 }

svn commit: r1586167 [2/2] - in /spark/site/docs/0.9.1: ./ api/pyspark/

2014-04-09 Thread tdas
Modified: spark/site/docs/0.9.1/tuning.html
URL: 
http://svn.apache.org/viewvc/spark/site/docs/0.9.1/tuning.html?rev=1586167r1=1586166r2=1586167view=diff
==
--- spark/site/docs/0.9.1/tuning.html (original)
+++ spark/site/docs/0.9.1/tuning.html Wed Apr  9 22:56:47 2014
@@ -25,6 +25,19 @@
 link rel=stylesheet href=css/pygments-default.css
 
 
+!-- Google analytics script --
+script type=text/javascript
+  var _gaq = _gaq || [];
+  _gaq.push(['_setAccount', 'UA-32518208-1']);
+  _gaq.push(['_trackPageview']);
+
+  (function() {
+var ga = document.createElement('script'); ga.type = 
'text/javascript'; ga.async = true;
+ga.src = ('https:' == document.location.protocol ? 'https://ssl' : 
'http://www') + '.google-analytics.com/ga.js';
+var s = document.getElementsByTagName('script')[0]; 
s.parentNode.insertBefore(ga, s);
+  })();
+/script
+
 
 /head
 body




svn commit: r1586169 - in /spark: releases/_posts/2014-04-09-spark-release-0-9-1.md site/docs/0.9.1/0.9.1-docs.tgz site/releases/spark-release-0-9-1.html

2014-04-09 Thread tdas
Author: tdas
Date: Wed Apr  9 23:19:13 2014
New Revision: 1586169

URL: http://svn.apache.org/r1586169
Log:
Updated 0.9.1 release notes

Removed:
spark/site/docs/0.9.1/0.9.1-docs.tgz
Modified:
spark/releases/_posts/2014-04-09-spark-release-0-9-1.md
spark/site/releases/spark-release-0-9-1.html

Modified: spark/releases/_posts/2014-04-09-spark-release-0-9-1.md
URL: 
http://svn.apache.org/viewvc/spark/releases/_posts/2014-04-09-spark-release-0-9-1.md?rev=1586169r1=1586168r2=1586169view=diff
==
--- spark/releases/_posts/2014-04-09-spark-release-0-9-1.md (original)
+++ spark/releases/_posts/2014-04-09-spark-release-0-9-1.md Wed Apr  9 23:19:13 
2014
@@ -59,7 +59,7 @@ Several bug fixes were made to YARN depl
 * Fixed bug in Python de-pickling 
[[SPARK-1135](https://issues.apache.org/jira/browse/SPARK-1135)]
 * Fixed bug in serialization of strings longer than 64K 
[[SPARK-1043](https://issues.apache.org/jira/browse/SPARK-1043)] 
 * Fixed bug that made jobs hang when base file is not available 
[[SPARK-1025](https://issues.apache.org/jira/browse/SPARK-1025)] 
-* Added Missing RDD operations to PySpark - top, zip, foldByKey and setName 
+* Added Missing RDD operations to PySpark - top, zip, foldByKey, repartition, 
coallesce, getStorageLevel, setName and toDebugString
 
 ### Improvements to documentation
 * Streaming: Added documentation on running streaming application from 
spark-shell

Modified: spark/site/releases/spark-release-0-9-1.html
URL: 
http://svn.apache.org/viewvc/spark/site/releases/spark-release-0-9-1.html?rev=1586169r1=1586168r2=1586169view=diff
==
--- spark/site/releases/spark-release-0-9-1.html (original)
+++ spark/site/releases/spark-release-0-9-1.html Wed Apr  9 23:19:13 2014
@@ -217,7 +217,7 @@
   liFixed bug in Python de-pickling [a 
href=https://issues.apache.org/jira/browse/SPARK-1135;SPARK-1135/a]/li
   liFixed bug in serialization of strings longer than 64K [a 
href=https://issues.apache.org/jira/browse/SPARK-1043;SPARK-1043/a] /li
   liFixed bug that made jobs hang when base file is not available [a 
href=https://issues.apache.org/jira/browse/SPARK-1025;SPARK-1025/a] /li
-  liAdded Missing RDD operations to PySpark - top, zip, foldByKey and 
setName /li
+  liAdded Missing RDD operations to PySpark - top, zip, foldByKey, 
repartition, coallesce, getStorageLevel, setName and toDebugString/li
 /ul
 
 h3 id=improvements-to-documentationImprovements to documentation/h3




git commit: [SPARK-1357 (fix)] remove empty line after :: DeveloperApi/Experimental ::

2014-04-09 Thread pwendell
Repository: spark
Updated Branches:
  refs/heads/master eb5f2b642 - 0adc932ad


[SPARK-1357 (fix)] remove empty line after :: DeveloperApi/Experimental ::

Remove empty line after :: DeveloperApi/Experimental :: in comments to make the 
original doc show up in the preview of the generated html docs. Thanks 
@andrewor14 !

Author: Xiangrui Meng m...@databricks.com

Closes #373 from mengxr/api and squashes the following commits:

9c35bdc [Xiangrui Meng] remove the empty line after :: 
DeveloperApi/Experimental ::


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/0adc932a
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/0adc932a
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/0adc932a

Branch: refs/heads/master
Commit: 0adc932add413a1754107b21d5ecfb38c0c3a4eb
Parents: eb5f2b6
Author: Xiangrui Meng m...@databricks.com
Authored: Wed Apr 9 17:08:17 2014 -0700
Committer: Patrick Wendell pwend...@gmail.com
Committed: Wed Apr 9 17:08:17 2014 -0700

--
 .../spark/mllib/api/python/PythonMLLibAPI.scala |  1 -
 .../spark/mllib/classification/NaiveBayes.scala |  1 -
 .../apache/spark/mllib/clustering/KMeans.scala  |  4 --
 .../linalg/distributed/CoordinateMatrix.scala   |  1 -
 .../linalg/distributed/IndexedRowMatrix.scala   |  2 -
 .../mllib/linalg/distributed/RowMatrix.scala|  1 -
 .../spark/mllib/optimization/Gradient.scala |  4 --
 .../mllib/optimization/GradientDescent.scala|  2 -
 .../spark/mllib/optimization/Optimizer.scala|  1 -
 .../spark/mllib/optimization/Updater.scala  |  4 --
 .../apache/spark/mllib/recommendation/ALS.scala |  1 -
 .../MatrixFactorizationModel.scala  |  1 -
 .../regression/GeneralizedLinearAlgorithm.scala |  1 -
 .../apache/spark/mllib/tree/DecisionTree.scala  |  1 -
 .../spark/mllib/tree/configuration/Algo.scala   |  1 -
 .../mllib/tree/configuration/FeatureType.scala  |  1 -
 .../tree/configuration/QuantileStrategy.scala   |  1 -
 .../mllib/tree/configuration/Strategy.scala |  1 -
 .../spark/mllib/tree/impurity/Entropy.scala |  2 -
 .../apache/spark/mllib/tree/impurity/Gini.scala |  2 -
 .../spark/mllib/tree/impurity/Impurity.scala|  3 --
 .../spark/mllib/tree/impurity/Variance.scala|  2 -
 .../mllib/tree/model/DecisionTreeModel.scala|  1 -
 .../mllib/tree/model/InformationGainStats.scala |  1 -
 .../apache/spark/mllib/tree/model/Node.scala|  1 -
 .../apache/spark/mllib/tree/model/Split.scala   |  1 -
 .../spark/mllib/util/DataValidators.scala   |  1 -
 .../spark/mllib/util/KMeansDataGenerator.scala  |  1 -
 .../spark/mllib/util/LinearDataGenerator.scala  |  1 -
 .../util/LogisticRegressionDataGenerator.scala  |  1 -
 .../spark/mllib/util/MFDataGenerator.scala  | 43 ++--
 .../org/apache/spark/mllib/util/MLUtils.scala   |  2 -
 .../spark/mllib/util/SVMDataGenerator.scala |  1 -
 33 files changed, 21 insertions(+), 71 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/0adc932a/mllib/src/main/scala/org/apache/spark/mllib/api/python/PythonMLLibAPI.scala
--
diff --git 
a/mllib/src/main/scala/org/apache/spark/mllib/api/python/PythonMLLibAPI.scala 
b/mllib/src/main/scala/org/apache/spark/mllib/api/python/PythonMLLibAPI.scala
index ae27c57..a6c049e 100644
--- 
a/mllib/src/main/scala/org/apache/spark/mllib/api/python/PythonMLLibAPI.scala
+++ 
b/mllib/src/main/scala/org/apache/spark/mllib/api/python/PythonMLLibAPI.scala
@@ -30,7 +30,6 @@ import org.apache.spark.rdd.RDD
 
 /**
  * :: DeveloperApi ::
- *
  * The Java stubs necessary for the Python mllib bindings.
  */
 @DeveloperApi

http://git-wip-us.apache.org/repos/asf/spark/blob/0adc932a/mllib/src/main/scala/org/apache/spark/mllib/classification/NaiveBayes.scala
--
diff --git 
a/mllib/src/main/scala/org/apache/spark/mllib/classification/NaiveBayes.scala 
b/mllib/src/main/scala/org/apache/spark/mllib/classification/NaiveBayes.scala
index 5a45f12..1865885 100644
--- 
a/mllib/src/main/scala/org/apache/spark/mllib/classification/NaiveBayes.scala
+++ 
b/mllib/src/main/scala/org/apache/spark/mllib/classification/NaiveBayes.scala
@@ -29,7 +29,6 @@ import org.apache.spark.rdd.RDD
 
 /**
  * :: Experimental ::
- *
  * Model for Naive Bayes Classifiers.
  *
  * @param labels list of labels

http://git-wip-us.apache.org/repos/asf/spark/blob/0adc932a/mllib/src/main/scala/org/apache/spark/mllib/clustering/KMeans.scala
--
diff --git 
a/mllib/src/main/scala/org/apache/spark/mllib/clustering/KMeans.scala 
b/mllib/src/main/scala/org/apache/spark/mllib/clustering/KMeans.scala
index 8f565eb..90cf852 100644
--- 

svn commit: r1586186 - /spark/site/examples.html

2014-04-09 Thread andrew
Author: andrew
Date: Thu Apr 10 00:52:59 2014
New Revision: 1586186

URL: http://svn.apache.org/r1586186
Log:
adding changed versions of compiled files as well

Modified:
spark/site/examples.html

Modified: spark/site/examples.html
URL: 
http://svn.apache.org/viewvc/spark/site/examples.html?rev=1586186r1=1586185r2=1586186view=diff
==
--- spark/site/examples.html (original)
+++ spark/site/examples.html Thu Apr 10 00:52:59 2014
@@ -409,8 +409,8 @@ previous ones, and emactions/em, whi
 pMany additional examples are distributed with Spark:/p
 
 ul
-  liBasic Spark: a 
href=https://github.com/apache/incubator-spark/tree/master/examples/src/main/scala/org/apache/spark/examples;Scala
 examples/a, a 
href=https://github.com/apache/incubator-spark/tree/master/examples/src/main/java/org/apache/spark/examples;Java
 examples/a, a 
href=https://github.com/apache/incubator-spark/tree/master/python/examples;Python
 examples/a/li
-  liSpark Streaming: a 
href=https://github.com/apache/incubator-spark/tree/master/examples/src/main/scala/org/apache/spark/streaming/examples;Scala
 examples/a, a 
href=https://github.com/apache/incubator-spark/tree/master/examples/src/main/java/org/apache/spark/streaming/examples;Java
 examples/a/li
+  liBasic Spark: a 
href=https://github.com/apache/spark/tree/master/examples/src/main/scala/org/apache/spark/examples;Scala
 examples/a, a 
href=https://github.com/apache/spark/tree/master/examples/src/main/java/org/apache/spark/examples;Java
 examples/a, a 
href=https://github.com/apache/spark/tree/master/python/examples;Python 
examples/a/li
+  liSpark Streaming: a 
href=https://github.com/apache/spark/tree/master/examples/src/main/scala/org/apache/spark/streaming/examples;Scala
 examples/a, a 
href=https://github.com/apache/spark/tree/master/examples/src/main/java/org/apache/spark/streaming/examples;Java
 examples/a/li
 /ul
 
 




svn commit: r1586197 - in /spark: releases/_posts/2014-04-09-spark-release-0-9-1.md site/releases/spark-release-0-9-1.html

2014-04-09 Thread tdas
Author: tdas
Date: Thu Apr 10 02:17:44 2014
New Revision: 1586197

URL: http://svn.apache.org/r1586197
Log:
Fixed typo

Modified:
spark/releases/_posts/2014-04-09-spark-release-0-9-1.md
spark/site/releases/spark-release-0-9-1.html

Modified: spark/releases/_posts/2014-04-09-spark-release-0-9-1.md
URL: 
http://svn.apache.org/viewvc/spark/releases/_posts/2014-04-09-spark-release-0-9-1.md?rev=1586197r1=1586196r2=1586197view=diff
==
--- spark/releases/_posts/2014-04-09-spark-release-0-9-1.md (original)
+++ spark/releases/_posts/2014-04-09-spark-release-0-9-1.md Thu Apr 10 02:17:44 
2014
@@ -59,7 +59,7 @@ Several bug fixes were made to YARN depl
 * Fixed bug in Python de-pickling 
[[SPARK-1135](https://issues.apache.org/jira/browse/SPARK-1135)]
 * Fixed bug in serialization of strings longer than 64K 
[[SPARK-1043](https://issues.apache.org/jira/browse/SPARK-1043)] 
 * Fixed bug that made jobs hang when base file is not available 
[[SPARK-1025](https://issues.apache.org/jira/browse/SPARK-1025)] 
-* Added Missing RDD operations to PySpark - top, zip, foldByKey, repartition, 
coallesce, getStorageLevel, setName and toDebugString
+* Added Missing RDD operations to PySpark - top, zip, foldByKey, repartition, 
coalesce, getStorageLevel, setName and toDebugString
 
 ### Improvements to documentation
 * Streaming: Added documentation on running streaming application from 
spark-shell

Modified: spark/site/releases/spark-release-0-9-1.html
URL: 
http://svn.apache.org/viewvc/spark/site/releases/spark-release-0-9-1.html?rev=1586197r1=1586196r2=1586197view=diff
==
--- spark/site/releases/spark-release-0-9-1.html (original)
+++ spark/site/releases/spark-release-0-9-1.html Thu Apr 10 02:17:44 2014
@@ -217,7 +217,7 @@
   liFixed bug in Python de-pickling [a 
href=https://issues.apache.org/jira/browse/SPARK-1135;SPARK-1135/a]/li
   liFixed bug in serialization of strings longer than 64K [a 
href=https://issues.apache.org/jira/browse/SPARK-1043;SPARK-1043/a] /li
   liFixed bug that made jobs hang when base file is not available [a 
href=https://issues.apache.org/jira/browse/SPARK-1025;SPARK-1025/a] /li
-  liAdded Missing RDD operations to PySpark - top, zip, foldByKey, 
repartition, coallesce, getStorageLevel, setName and toDebugString/li
+  liAdded Missing RDD operations to PySpark - top, zip, foldByKey, 
repartition, coalesce, getStorageLevel, setName and toDebugString/li
 /ul
 
 h3 id=improvements-to-documentationImprovements to documentation/h3