spark git commit: BUILD: Minor tweaks to internal build scripts

2015-03-03 Thread pwendell
Repository: spark
Updated Branches:
  refs/heads/branch-1.3 1aa846165 - ae60eb998


BUILD: Minor tweaks to internal build scripts

This adds two features:
1. The ability to publish with a different maven version than
   that specified in the release source.
2. Forking of different Zinc instances during the parallel dist
   creation (to help with some stability issues).


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/ae60eb99
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/ae60eb99
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/ae60eb99

Branch: refs/heads/branch-1.3
Commit: ae60eb9984de56c36c7f63220f9281fbaac10931
Parents: 1aa8461
Author: Patrick Wendell patr...@databricks.com
Authored: Tue Mar 3 00:38:12 2015 -0800
Committer: Patrick Wendell patr...@databricks.com
Committed: Tue Mar 3 01:54:06 2015 -0800

--
 dev/create-release/create-release.sh | 24 +++-
 1 file changed, 19 insertions(+), 5 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/ae60eb99/dev/create-release/create-release.sh
--
diff --git a/dev/create-release/create-release.sh 
b/dev/create-release/create-release.sh
index da15ce3..6f87fcd 100755
--- a/dev/create-release/create-release.sh
+++ b/dev/create-release/create-release.sh
@@ -34,6 +34,9 @@ ASF_PASSWORD=${ASF_PASSWORD:-XXX}
 GPG_PASSPHRASE=${GPG_PASSPHRASE:-XXX}
 GIT_BRANCH=${GIT_BRANCH:-branch-1.0}
 RELEASE_VERSION=${RELEASE_VERSION:-1.2.0}
+# Allows publishing under a different version identifier than
+# was present in the actual release sources (e.g. rc-X)
+PUBLISH_VERSION=${PUBLISH_VERSION:-$RELEASE_VERSION} 
 NEXT_VERSION=${NEXT_VERSION:-1.2.1}
 RC_NAME=${RC_NAME:-rc2}
 
@@ -97,30 +100,35 @@ if [[ ! $@ =~ --skip-publish ]]; then
   pushd spark
   git checkout --force $GIT_TAG 
   
+  # Substitute in case published version is different than released
+  old=^\( \{2,4\}\)version${RELEASE_VERSION}\/version$
+  new=\1version${PUBLISH_VERSION}\/version
+  find . -name pom.xml | grep -v dev | xargs -I {} sed -i \
+-e s/${old}/${new}/ {}
+
   # Using Nexus API documented here:
   # 
https://support.sonatype.com/entries/39720203-Uploading-to-a-Staging-Repository-via-REST-API
   echo Creating Nexus staging repository
-  repo_request=promoteRequestdatadescriptionApache Spark 
$GIT_TAG/description/data/promoteRequest
+  repo_request=promoteRequestdatadescriptionApache Spark $GIT_TAG 
(published as $PUBLISH_VERSION)/description/data/promoteRequest
   out=$(curl -X POST -d $repo_request -u $ASF_USERNAME:$ASF_PASSWORD \
 -H Content-Type:application/xml -v \
 $NEXUS_ROOT/profiles/$NEXUS_PROFILE/start)
   staged_repo_id=$(echo $out | sed -e 
s/.*\(orgapachespark-[0-9]\{4\}\).*/\1/)
   echo Created Nexus staging repository: $staged_repo_id
 
-  rm -rf $SPARK_REPO
-
-  mvn -DskipTests -Dhadoop.version=2.2.0 -Dyarn.version=2.2.0 \
+  build/mvn -DskipTests -Dhadoop.version=2.2.0 -Dyarn.version=2.2.0 \
 -Pyarn -Phive -Phadoop-2.2 -Pspark-ganglia-lgpl -Pkinesis-asl \
 clean install
 
   ./dev/change-version-to-2.11.sh
   
-  mvn -DskipTests -Dhadoop.version=2.2.0 -Dyarn.version=2.2.0 \
+  build/mvn -DskipTests -Dhadoop.version=2.2.0 -Dyarn.version=2.2.0 \
 -Dscala-2.11 -Pyarn -Phive -Phadoop-2.2 -Pspark-ganglia-lgpl -Pkinesis-asl 
\
 clean install
 
   ./dev/change-version-to-2.10.sh
 
+  rm -rf $SPARK_REPO
   pushd $SPARK_REPO
 
   # Remove any extra files generated during install
@@ -197,6 +205,12 @@ if [[ ! $@ =~ --skip-package ]]; then
   ./dev/change-version-to-2.11.sh
 fi
 
+# Create new Zinc instances for each binary release to avoid interference
+# that causes OOM's and random compiler crashes.
+zinc_port=${zinc_port:-3030}
+zinc_port=$[$zinc_port + 1]
+export ZINC_PORT=$zinc_port
+
 ./make-distribution.sh --name $NAME --tgz $FLAGS 21 | tee 
../binary-release-$NAME.log
 cd ..
 cp spark-$RELEASE_VERSION-bin-$NAME/spark-$RELEASE_VERSION-bin-$NAME.tgz .


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[1/3] spark git commit: Adding CHANGES.txt for Spark 1.3

2015-03-03 Thread pwendell
Repository: spark
Updated Branches:
  refs/heads/branch-1.3 ae60eb998 - ce7158cf7


http://git-wip-us.apache.org/repos/asf/spark/blob/ce7158cf/dev/create-release/generate-changelist.py
--
diff --git a/dev/create-release/generate-changelist.py 
b/dev/create-release/generate-changelist.py
index 2e1a35a..2a2e4b7 100755
--- a/dev/create-release/generate-changelist.py
+++ b/dev/create-release/generate-changelist.py
@@ -31,8 +31,8 @@ import time
 import traceback
 
 SPARK_HOME = os.environ[SPARK_HOME]
-NEW_RELEASE_VERSION = 1.0.0
-PREV_RELEASE_GIT_TAG = v0.9.1
+NEW_RELEASE_VERSION = 1.3.0
+PREV_RELEASE_GIT_TAG = v1.2.1
 
 CHANGELIST = CHANGES.txt
 OLD_CHANGELIST = %s.old % (CHANGELIST)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[2/3] spark git commit: Adding CHANGES.txt for Spark 1.3

2015-03-03 Thread pwendell
http://git-wip-us.apache.org/repos/asf/spark/blob/ce7158cf/CHANGES.txt
--
diff --git a/CHANGES.txt b/CHANGES.txt
new file mode 100644
index 000..46d612a
--- /dev/null
+++ b/CHANGES.txt
@@ -0,0 +1,6520 @@
+Spark Change Log
+
+
+Release 1.3.0
+
+  BUILD: Minor tweaks to internal build scripts
+  Patrick Wendell patr...@databricks.com
+  2015-03-03 00:38:12 -0800
+  Commit: ae60eb9
+
+  HOTFIX: Bump HBase version in MapR profiles.
+  Patrick Wendell patr...@databricks.com
+  2015-03-03 01:38:07 -0800
+  Commit: 1aa8461
+
+  [SPARK-5537][MLlib][Docs] Add user guide for multinomial logistic regression
+  DB Tsai dbt...@alpinenow.com
+  2015-03-02 22:37:12 -0800
+  Commit: 841d2a2, github.com/apache/spark/pull/4866
+
+  [SPARK-6120] [mllib] Warnings about memory in tree, ensemble model save
+  Joseph K. Bradley jos...@databricks.com
+  2015-03-02 22:33:51 -0800
+  Commit: 81648a7, github.com/apache/spark/pull/4864
+
+  [SPARK-6097][MLLIB] Support tree model save/load in PySpark/MLlib
+  Xiangrui Meng m...@databricks.com
+  2015-03-02 22:27:01 -0800
+  Commit: 62c53be, github.com/apache/spark/pull/4854
+
+  [SPARK-5310][SQL] Fixes to Docs and Datasources API
+  Reynold Xin r...@databricks.com, Michael Armbrust mich...@databricks.com
+  2015-03-02 22:14:08 -0800
+  Commit: 4e6e008, github.com/apache/spark/pull/4868
+
+  [SPARK-5950][SQL]Insert array into a metastore table saved as parquet should 
work when using datasource api
+  Yin Huai yh...@databricks.com
+  2015-03-02 19:31:55 -0800
+  Commit: 1b490e9, github.com/apache/spark/pull/4826
+
+  [SPARK-6127][Streaming][Docs] Add Kafka to Python api docs
+  Tathagata Das tathagata.das1...@gmail.com
+  2015-03-02 18:40:46 -0800
+  Commit: ffd0591, github.com/apache/spark/pull/4860
+
+  [SPARK-5537] Add user guide for multinomial logistic regression
+  Xiangrui Meng m...@databricks.com, DB Tsai dbt...@alpinenow.com
+  2015-03-02 18:10:50 -0800
+  Commit: 11389f0, github.com/apache/spark/pull/4801
+
+  [SPARK-6121][SQL][MLLIB] simpleString for UDT
+  Xiangrui Meng m...@databricks.com
+  2015-03-02 17:14:34 -0800
+  Commit: 1b8ab57, github.com/apache/spark/pull/4858
+
+  [SPARK-6048] SparkConf should not translate deprecated configs on set
+  Andrew Or and...@databricks.com
+  2015-03-02 16:36:42 -0800
+  Commit: ea69cf2, github.com/apache/spark/pull/4799
+
+  [SPARK-6066] Make event log format easier to parse
+  Andrew Or and...@databricks.com
+  2015-03-02 16:34:32 -0800
+  Commit: 8100b79, github.com/apache/spark/pull/4821
+
+  [SPARK-6082] [SQL] Provides better error message for malformed rows when 
caching tables
+  Cheng Lian l...@databricks.com
+  2015-03-02 16:18:00 -0800
+  Commit: 866f281, github.com/apache/spark/pull/4842
+
+  [SPARK-6114][SQL] Avoid metastore conversions before plan is resolved
+  Michael Armbrust mich...@databricks.com
+  2015-03-02 16:10:54 -0800
+  Commit: 3899c7c, github.com/apache/spark/pull/4855
+
+  [SPARK-6050] [yarn] Relax matching of vcore count in received containers.
+  Marcelo Vanzin van...@cloudera.com
+  2015-03-02 16:41:43 -0600
+  Commit: 650d1e7, github.com/apache/spark/pull/4818
+
+  [SPARK-6040][SQL] Fix the percent bug in tablesample
+  q00251598 qiyad...@huawei.com
+  2015-03-02 13:16:29 -0800
+  Commit: a83b9bb, github.com/apache/spark/pull/4789
+
+  [Minor] Fix doc typo for describing primitiveTerm effectiveness condition
+  Liang-Chi Hsieh vii...@gmail.com
+  2015-03-02 13:11:17 -0800
+  Commit: f92876a, github.com/apache/spark/pull/4762
+
+  SPARK-5390 [DOCS] Encourage users to post on Stack Overflow in Community Docs
+  Sean Owen so...@cloudera.com
+  2015-03-02 21:10:08 +
+  Commit: 58e7198, github.com/apache/spark/pull/4843
+
+  [DOCS] Refactored Dataframe join comment to use correct parameter ordering
+  Paul Power paul.po...@peerside.com
+  2015-03-02 13:08:47 -0800
+  Commit: 54ac243, github.com/apache/spark/pull/4847
+
+  [SPARK-6080] [PySpark] correct LogisticRegressionWithLBFGS regType parameter 
for pyspark
+  Yanbo Liang yblia...@gmail.com
+  2015-03-02 10:17:24 -0800
+  Commit: 4ffaf85, github.com/apache/spark/pull/4831
+
+  [SPARK-5741][SQL] Support the path contains comma in HiveContext
+  q00251598 qiyad...@huawei.com
+  2015-03-02 10:13:11 -0800
+  Commit: f476108, github.com/apache/spark/pull/4532
+
+  [SPARK-6111] Fixed usage string in documentation.
+  Kenneth Myers myer...@us.ibm.com
+  2015-03-02 17:25:24 +
+  Commit: b2b7f01, github.com/apache/spark/pull/4852
+
+  [SPARK-6052][SQL]In JSON schema inference, we should always set containsNull 
of an ArrayType to true
+  Yin Huai yh...@databricks.com
+  2015-03-02 23:18:07 +0800
+  Commit: a3fef2c, github.com/apache/spark/pull/4806
+
+  [SPARK-6073][SQL] Need to refresh metastore cache after append data in 
CreateMetastoreDataSourceAsSelect
+  Yin Huai yh...@databricks.com
+  2015-03-02 22:42:18 +0800
+  Commit: c59871c, 

[3/3] spark git commit: Adding CHANGES.txt for Spark 1.3

2015-03-03 Thread pwendell
Adding CHANGES.txt for Spark 1.3


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/ce7158cf
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/ce7158cf
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/ce7158cf

Branch: refs/heads/branch-1.3
Commit: ce7158cf70c1003c1011d9a755813b31feae91e4
Parents: ae60eb9
Author: Patrick Wendell patr...@databricks.com
Authored: Tue Mar 3 02:19:19 2015 -0800
Committer: Patrick Wendell patr...@databricks.com
Committed: Tue Mar 3 02:19:19 2015 -0800

--
 CHANGES.txt   | 6520 
 dev/create-release/generate-changelist.py |4 +-
 2 files changed, 6522 insertions(+), 2 deletions(-)
--



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[1/2] spark git commit: Preparing development version 1.3.1-SNAPSHOT

2015-03-03 Thread pwendell
Repository: spark
Updated Branches:
  refs/heads/branch-1.3 b012ed189 - 05d5a29eb


Preparing development version 1.3.1-SNAPSHOT


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/05d5a29e
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/05d5a29e
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/05d5a29e

Branch: refs/heads/branch-1.3
Commit: 05d5a29eb3193aeb57d177bafe39eb75edce72a1
Parents: 3af2687
Author: Patrick Wendell patr...@databricks.com
Authored: Tue Mar 3 10:23:07 2015 +
Committer: Patrick Wendell patr...@databricks.com
Committed: Tue Mar 3 10:23:07 2015 +

--
 assembly/pom.xml  | 2 +-
 bagel/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 examples/pom.xml  | 2 +-
 external/flume-sink/pom.xml   | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka-assembly/pom.xml   | 2 +-
 external/kafka/pom.xml| 2 +-
 external/mqtt/pom.xml | 2 +-
 external/twitter/pom.xml  | 2 +-
 external/zeromq/pom.xml   | 2 +-
 extras/java8-tests/pom.xml| 2 +-
 extras/kinesis-asl/pom.xml| 2 +-
 extras/spark-ganglia-lgpl/pom.xml | 2 +-
 graphx/pom.xml| 2 +-
 mllib/pom.xml | 2 +-
 network/common/pom.xml| 2 +-
 network/shuffle/pom.xml   | 2 +-
 network/yarn/pom.xml  | 2 +-
 pom.xml   | 2 +-
 repl/pom.xml  | 2 +-
 sql/catalyst/pom.xml  | 2 +-
 sql/core/pom.xml  | 2 +-
 sql/hive-thriftserver/pom.xml | 2 +-
 sql/hive/pom.xml  | 2 +-
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 yarn/pom.xml  | 2 +-
 28 files changed, 28 insertions(+), 28 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/05d5a29e/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index 0764d2d..87b3e6f 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.3.0/version
+version1.3.1-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/05d5a29e/bagel/pom.xml
--
diff --git a/bagel/pom.xml b/bagel/pom.xml
index 4f73cf7..4a13c58 100644
--- a/bagel/pom.xml
+++ b/bagel/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.3.0/version
+version1.3.1-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/05d5a29e/core/pom.xml
--
diff --git a/core/pom.xml b/core/pom.xml
index 5612149..aca0f58 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.3.0/version
+version1.3.1-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/05d5a29e/examples/pom.xml
--
diff --git a/examples/pom.xml b/examples/pom.xml
index f7d6030..c424592 100644
--- a/examples/pom.xml
+++ b/examples/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.3.0/version
+version1.3.1-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/05d5a29e/external/flume-sink/pom.xml
--
diff --git a/external/flume-sink/pom.xml b/external/flume-sink/pom.xml
index 45aa775..ef960a8 100644
--- a/external/flume-sink/pom.xml
+++ b/external/flume-sink/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.3.0/version
+version1.3.1-SNAPSHOT/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/05d5a29e/external/flume/pom.xml
--
diff --git a/external/flume/pom.xml b/external/flume/pom.xml
index 455304f..f01d6e8 100644
--- a/external/flume/pom.xml
+++ b/external/flume/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.3.0/version
+version1.3.1-SNAPSHOT/version
 

svn commit: r1663693 - /spark/site/downloads.html

2015-03-03 Thread srowen
Author: srowen
Date: Tue Mar  3 13:04:20 2015
New Revision: 1663693

URL: http://svn.apache.org/r1663693
Log:
SPARK-5055: fix Chose - Choose typo

Modified:
spark/site/downloads.html

Modified: spark/site/downloads.html
URL: 
http://svn.apache.org/viewvc/spark/site/downloads.html?rev=1663693r1=1663692r2=1663693view=diff
==
--- spark/site/downloads.html (original)
+++ spark/site/downloads.html Tue Mar  3 13:04:20 2015
@@ -182,15 +182,15 @@ $(document).ready(function() {
 
 ol
   li
-pChose a Spark release:
+pChoose a Spark release:
   select id=sparkVersionSelect 
onchange=javascript:onVersionSelect();/selectbr //p
   /li
   li
-pChose a package type:
+pChoose a package type:
   select id=sparkPackageSelect 
onchange=javascript:onPackageSelect();/selectbr //p
   /li
   li
-pChose a download type:
+pChoose a download type:
   select id=sparkDownloadSelect 
onchange=javascript:onDownloadSelect()/selectbr //p
   /li
   li



-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



spark git commit: Revert [SPARK-5423][Core] Cleanup resources in DiskMapIterator.finalize to ensure deleting the temp file

2015-03-03 Thread andrewor14
Repository: spark
Updated Branches:
  refs/heads/master e359794ce - 9af001749


Revert [SPARK-5423][Core] Cleanup resources in DiskMapIterator.finalize to 
ensure deleting the temp file

This reverts commit 90095bf3ce9304d09a32ceffaa99069079071b59.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/9af00174
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/9af00174
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/9af00174

Branch: refs/heads/master
Commit: 9af001749a37a86ccbf78063ec514a21801645fa
Parents: e359794
Author: Andrew Or and...@databricks.com
Authored: Tue Mar 3 13:03:52 2015 -0800
Committer: Andrew Or and...@databricks.com
Committed: Tue Mar 3 13:03:52 2015 -0800

--
 .../util/collection/ExternalAppendOnlyMap.scala | 52 
 1 file changed, 9 insertions(+), 43 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/9af00174/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
--
diff --git 
a/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
 
b/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
index fc7e86e..8a0f5a6 100644
--- 
a/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
+++ 
b/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
@@ -387,15 +387,6 @@ class ExternalAppendOnlyMap[K, V, C](
 private var batchIndex = 0  // Which batch we're in
 private var fileStream: FileInputStream = null
 
-@volatile private var closed = false
-
-// A volatile variable to remember which DeserializationStream is using. 
Need to set it when we
-// open a DeserializationStream. But we should use `deserializeStream` 
rather than
-// `deserializeStreamToBeClosed` to read the content because touching a 
volatile variable will
-// reduce the performance. It must be volatile so that we can see its 
correct value in the
-// `finalize` method, which could run in any thread.
-@volatile private var deserializeStreamToBeClosed: DeserializationStream = 
null
-
 // An intermediate stream that reads from exactly one batch
 // This guards against pre-fetching and other arbitrary behavior of higher 
level streams
 private var deserializeStream = nextBatchStream()
@@ -410,7 +401,6 @@ class ExternalAppendOnlyMap[K, V, C](
   // we're still in a valid batch.
   if (batchIndex  batchOffsets.length - 1) {
 if (deserializeStream != null) {
-  deserializeStreamToBeClosed = null
   deserializeStream.close()
   fileStream.close()
   deserializeStream = null
@@ -429,11 +419,7 @@ class ExternalAppendOnlyMap[K, V, C](
 
 val bufferedStream = new 
BufferedInputStream(ByteStreams.limit(fileStream, end - start))
 val compressedStream = blockManager.wrapForCompression(blockId, 
bufferedStream)
-// Before returning the stream, assign it to 
`deserializeStreamToBeClosed` so that we can
-// close it in `finalize` and also avoid to touch the volatile 
`deserializeStreamToBeClosed`
-// during reading the (K, C) pairs.
-deserializeStreamToBeClosed = ser.deserializeStream(compressedStream)
-deserializeStreamToBeClosed
+ser.deserializeStream(compressedStream)
   } else {
 // No more batches left
 cleanup()
@@ -482,34 +468,14 @@ class ExternalAppendOnlyMap[K, V, C](
   item
 }
 
-// TODO: Now only use `finalize` to ensure `close` gets called to clean up 
the resources. In the
-// future, we need some mechanism to ensure this gets called once the 
resources are not used.
-private def cleanup(): Unit = {
-  if (!closed) {
-closed = true
-batchIndex = batchOffsets.length  // Prevent reading any other batch
-fileStream = null
-try {
-  val ds = deserializeStreamToBeClosed
-  deserializeStreamToBeClosed = null
-  deserializeStream = null
-  if (ds != null) {
-ds.close()
-  }
-} finally {
-  if (file.exists()) {
-file.delete()
-  }
-}
-  }
-}
-
-override def finalize(): Unit = {
-  try {
-cleanup()
-  } finally {
-super.finalize()
-  }
+// TODO: Ensure this gets called even if the iterator isn't drained.
+private def cleanup() {
+  batchIndex = batchOffsets.length  // Prevent reading any other batch
+  val ds = deserializeStream
+  deserializeStream = null
+  fileStream = null
+  ds.close()
+  file.delete()
 }
   }
 


-
To 

spark git commit: Revert [SPARK-5423][Core] Cleanup resources in DiskMapIterator.finalize to ensure deleting the temp file

2015-03-03 Thread andrewor14
Repository: spark
Updated Branches:
  refs/heads/branch-1.1 91d0effb3 - d70754df0


Revert [SPARK-5423][Core] Cleanup resources in DiskMapIterator.finalize to 
ensure deleting the temp file

This reverts commit 36f3c499fd1ad53a68a084d6a16a2c68099e7049.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/d70754df
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/d70754df
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/d70754df

Branch: refs/heads/branch-1.1
Commit: d70754df04031a8bd99e2aa104303e98e358128e
Parents: 91d0eff
Author: Andrew Or and...@databricks.com
Authored: Tue Mar 3 13:05:13 2015 -0800
Committer: Andrew Or and...@databricks.com
Committed: Tue Mar 3 13:05:13 2015 -0800

--
 .../util/collection/ExternalAppendOnlyMap.scala | 52 
 1 file changed, 9 insertions(+), 43 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/d70754df/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
--
diff --git 
a/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
 
b/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
index abb1f11..5619b30 100644
--- 
a/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
+++ 
b/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
@@ -434,15 +434,6 @@ class ExternalAppendOnlyMap[K, V, C](
 private var batchIndex = 0  // Which batch we're in
 private var fileStream: FileInputStream = null
 
-@volatile private var closed = false
-
-// A volatile variable to remember which DeserializationStream is using. 
Need to set it when we
-// open a DeserializationStream. But we should use `deserializeStream` 
rather than
-// `deserializeStreamToBeClosed` to read the content because touching a 
volatile variable will
-// reduce the performance. It must be volatile so that we can see its 
correct value in the
-// `finalize` method, which could run in any thread.
-@volatile private var deserializeStreamToBeClosed: DeserializationStream = 
null
-
 // An intermediate stream that reads from exactly one batch
 // This guards against pre-fetching and other arbitrary behavior of higher 
level streams
 private var deserializeStream = nextBatchStream()
@@ -457,7 +448,6 @@ class ExternalAppendOnlyMap[K, V, C](
   // we're still in a valid batch.
   if (batchIndex  batchOffsets.length - 1) {
 if (deserializeStream != null) {
-  deserializeStreamToBeClosed = null
   deserializeStream.close()
   fileStream.close()
   deserializeStream = null
@@ -476,11 +466,7 @@ class ExternalAppendOnlyMap[K, V, C](
 
 val bufferedStream = new 
BufferedInputStream(ByteStreams.limit(fileStream, end - start))
 val compressedStream = blockManager.wrapForCompression(blockId, 
bufferedStream)
-// Before returning the stream, assign it to 
`deserializeStreamToBeClosed` so that we can
-// close it in `finalize` and also avoid to touch the volatile 
`deserializeStreamToBeClosed`
-// during reading the (K, C) pairs.
-deserializeStreamToBeClosed = ser.deserializeStream(compressedStream)
-deserializeStreamToBeClosed
+ser.deserializeStream(compressedStream)
   } else {
 // No more batches left
 cleanup()
@@ -529,34 +515,14 @@ class ExternalAppendOnlyMap[K, V, C](
   item
 }
 
-// TODO: Now only use `finalize` to ensure `close` gets called to clean up 
the resources. In the
-// future, we need some mechanism to ensure this gets called once the 
resources are not used.
-private def cleanup(): Unit = {
-  if (!closed) {
-closed = true
-batchIndex = batchOffsets.length  // Prevent reading any other batch
-fileStream = null
-try {
-  val ds = deserializeStreamToBeClosed
-  deserializeStreamToBeClosed = null
-  deserializeStream = null
-  if (ds != null) {
-ds.close()
-  }
-} finally {
-  if (file.exists()) {
-file.delete()
-  }
-}
-  }
-}
-
-override def finalize(): Unit = {
-  try {
-cleanup()
-  } finally {
-super.finalize()
-  }
+// TODO: Ensure this gets called even if the iterator isn't drained.
+private def cleanup() {
+  batchIndex = batchOffsets.length  // Prevent reading any other batch
+  val ds = deserializeStream
+  deserializeStream = null
+  fileStream = null
+  ds.close()
+  file.delete()
 }
   }
 


-

spark git commit: SPARK-1911 [DOCS] Warn users if their assembly jars are not built with Java 6

2015-03-03 Thread andrewor14
Repository: spark
Updated Branches:
  refs/heads/master 9af001749 - e750a6bfd


SPARK-1911 [DOCS] Warn users if their assembly jars are not built with Java 6

Add warning about building with Java 7+ and running the JAR on early Java 6.

CC andrewor14

Author: Sean Owen so...@cloudera.com

Closes #4874 from srowen/SPARK-1911 and squashes the following commits:

79fa2f6 [Sean Owen] Add warning about building with Java 7+ and running the JAR 
on early Java 6.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/e750a6bf
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/e750a6bf
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/e750a6bf

Branch: refs/heads/master
Commit: e750a6bfddf1d7bf7d3e99a424ec2b83a18b40d9
Parents: 9af0017
Author: Sean Owen so...@cloudera.com
Authored: Tue Mar 3 13:40:11 2015 -0800
Committer: Andrew Or and...@databricks.com
Committed: Tue Mar 3 13:40:11 2015 -0800

--
 docs/building-spark.md | 4 
 1 file changed, 4 insertions(+)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/e750a6bf/docs/building-spark.md
--
diff --git a/docs/building-spark.md b/docs/building-spark.md
index 4c3988e..57d0ca8 100644
--- a/docs/building-spark.md
+++ b/docs/building-spark.md
@@ -9,6 +9,10 @@ redirect_from: building-with-maven.html
 
 Building Spark using Maven requires Maven 3.0.4 or newer and Java 6+.
 
+**Note:** Building Spark with Java 7 or later can create JAR files that may 
not be
+readable with early versions of Java 6, due to the large number of files in 
the JAR
+archive. Build with Java 6 if this is an issue for your deployment.
+
 # Building with `build/mvn`
 
 Spark now comes packaged with a self-contained Maven installation to ease 
building and deployment of Spark from source located under the `build/` 
directory. This script will automatically download and setup all necessary 
build requirements ([Maven](https://maven.apache.org/), 
[Scala](http://www.scala-lang.org/), and 
[Zinc](https://github.com/typesafehub/zinc)) locally within the `build/` 
directory itself. It honors any `mvn` binary if present already, however, will 
pull down its own copy of Scala and Zinc regardless to ensure proper version 
requirements are met. `build/mvn` execution acts as a pass through to the `mvn` 
call allowing easy transition from previous build methods. As an example, one 
can build a version of Spark as follows:


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



spark git commit: [SPARK-6132] ContextCleaner race condition across SparkContexts

2015-03-03 Thread andrewor14
Repository: spark
Updated Branches:
  refs/heads/master e750a6bfd - fe63e8229


[SPARK-6132] ContextCleaner race condition across SparkContexts

The problem is that `ContextCleaner` may clean variables that belong to a 
different `SparkContext`. This can happen if the `SparkContext` to which the 
cleaner belongs stops, and a new one is started immediately afterwards in the 
same JVM. In this case, if the cleaner is in the middle of cleaning a 
broadcast, for instance, it will do so through `SparkEnv.get.blockManager`, 
which could be one that belongs to a different `SparkContext`.

JoshRosen and I suspect that this is the cause of many flaky tests, most 
notably the `JavaAPISuite`. We were able to reproduce the failure locally 
(though it is not deterministic and very hard to reproduce).

Author: Andrew Or and...@databricks.com

Closes #4869 from andrewor14/cleaner-masquerade and squashes the following 
commits:

29168c0 [Andrew Or] Synchronize ContextCleaner stop


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/fe63e822
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/fe63e822
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/fe63e822

Branch: refs/heads/master
Commit: fe63e822918a01e1c1d741052b932e9944745fb6
Parents: e750a6b
Author: Andrew Or and...@databricks.com
Authored: Tue Mar 3 13:44:05 2015 -0800
Committer: Andrew Or and...@databricks.com
Committed: Tue Mar 3 13:44:05 2015 -0800

--
 .../scala/org/apache/spark/ContextCleaner.scala | 39 +---
 1 file changed, 26 insertions(+), 13 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/fe63e822/core/src/main/scala/org/apache/spark/ContextCleaner.scala
--
diff --git a/core/src/main/scala/org/apache/spark/ContextCleaner.scala 
b/core/src/main/scala/org/apache/spark/ContextCleaner.scala
index 4a9d007..4dab886 100644
--- a/core/src/main/scala/org/apache/spark/ContextCleaner.scala
+++ b/core/src/main/scala/org/apache/spark/ContextCleaner.scala
@@ -105,9 +105,19 @@ private[spark] class ContextCleaner(sc: SparkContext) 
extends Logging {
 cleaningThread.start()
   }
 
-  /** Stop the cleaner. */
+  /**
+   * Stop the cleaning thread and wait until the thread has finished running 
its current task.
+   */
   def stop() {
 stopped = true
+// Interrupt the cleaning thread, but wait until the current task has 
finished before
+// doing so. This guards against the race condition where a cleaning 
thread may
+// potentially clean similarly named variables created by a different 
SparkContext,
+// resulting in otherwise inexplicable block-not-found exceptions 
(SPARK-6132).
+synchronized {
+  cleaningThread.interrupt()
+}
+cleaningThread.join()
   }
 
   /** Register a RDD for cleanup when it is garbage collected. */
@@ -140,18 +150,21 @@ private[spark] class ContextCleaner(sc: SparkContext) 
extends Logging {
   try {
 val reference = 
Option(referenceQueue.remove(ContextCleaner.REF_QUEUE_POLL_TIMEOUT))
   .map(_.asInstanceOf[CleanupTaskWeakReference])
-reference.map(_.task).foreach { task =
-  logDebug(Got cleaning task  + task)
-  referenceBuffer -= reference.get
-  task match {
-case CleanRDD(rddId) =
-  doCleanupRDD(rddId, blocking = blockOnCleanupTasks)
-case CleanShuffle(shuffleId) =
-  doCleanupShuffle(shuffleId, blocking = 
blockOnShuffleCleanupTasks)
-case CleanBroadcast(broadcastId) =
-  doCleanupBroadcast(broadcastId, blocking = blockOnCleanupTasks)
-case CleanAccum(accId) =
-  doCleanupAccum(accId, blocking = blockOnCleanupTasks)
+// Synchronize here to avoid being interrupted on stop()
+synchronized {
+  reference.map(_.task).foreach { task =
+logDebug(Got cleaning task  + task)
+referenceBuffer -= reference.get
+task match {
+  case CleanRDD(rddId) =
+doCleanupRDD(rddId, blocking = blockOnCleanupTasks)
+  case CleanShuffle(shuffleId) =
+doCleanupShuffle(shuffleId, blocking = 
blockOnShuffleCleanupTasks)
+  case CleanBroadcast(broadcastId) =
+doCleanupBroadcast(broadcastId, blocking = blockOnCleanupTasks)
+  case CleanAccum(accId) =
+doCleanupAccum(accId, blocking = blockOnCleanupTasks)
+}
   }
 }
   } catch {


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



spark git commit: Revert [SPARK-5423][Core] Cleanup resources in DiskMapIterator.finalize to ensure deleting the temp file

2015-03-03 Thread andrewor14
Repository: spark
Updated Branches:
  refs/heads/branch-1.3 05d5a29eb - ee4929d1d


Revert [SPARK-5423][Core] Cleanup resources in DiskMapIterator.finalize to 
ensure deleting the temp file

This reverts commit 25fae8e7e6c93b7817771342d370b73b40dcf92e.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/ee4929d1
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/ee4929d1
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/ee4929d1

Branch: refs/heads/branch-1.3
Commit: ee4929d1d38d83382ccdc22bf99f61f24f989c8b
Parents: 05d5a29
Author: Andrew Or and...@databricks.com
Authored: Tue Mar 3 13:04:15 2015 -0800
Committer: Andrew Or and...@databricks.com
Committed: Tue Mar 3 13:04:15 2015 -0800

--
 .../util/collection/ExternalAppendOnlyMap.scala | 52 
 1 file changed, 9 insertions(+), 43 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/ee4929d1/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
--
diff --git 
a/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
 
b/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
index fc7e86e..8a0f5a6 100644
--- 
a/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
+++ 
b/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
@@ -387,15 +387,6 @@ class ExternalAppendOnlyMap[K, V, C](
 private var batchIndex = 0  // Which batch we're in
 private var fileStream: FileInputStream = null
 
-@volatile private var closed = false
-
-// A volatile variable to remember which DeserializationStream is using. 
Need to set it when we
-// open a DeserializationStream. But we should use `deserializeStream` 
rather than
-// `deserializeStreamToBeClosed` to read the content because touching a 
volatile variable will
-// reduce the performance. It must be volatile so that we can see its 
correct value in the
-// `finalize` method, which could run in any thread.
-@volatile private var deserializeStreamToBeClosed: DeserializationStream = 
null
-
 // An intermediate stream that reads from exactly one batch
 // This guards against pre-fetching and other arbitrary behavior of higher 
level streams
 private var deserializeStream = nextBatchStream()
@@ -410,7 +401,6 @@ class ExternalAppendOnlyMap[K, V, C](
   // we're still in a valid batch.
   if (batchIndex  batchOffsets.length - 1) {
 if (deserializeStream != null) {
-  deserializeStreamToBeClosed = null
   deserializeStream.close()
   fileStream.close()
   deserializeStream = null
@@ -429,11 +419,7 @@ class ExternalAppendOnlyMap[K, V, C](
 
 val bufferedStream = new 
BufferedInputStream(ByteStreams.limit(fileStream, end - start))
 val compressedStream = blockManager.wrapForCompression(blockId, 
bufferedStream)
-// Before returning the stream, assign it to 
`deserializeStreamToBeClosed` so that we can
-// close it in `finalize` and also avoid to touch the volatile 
`deserializeStreamToBeClosed`
-// during reading the (K, C) pairs.
-deserializeStreamToBeClosed = ser.deserializeStream(compressedStream)
-deserializeStreamToBeClosed
+ser.deserializeStream(compressedStream)
   } else {
 // No more batches left
 cleanup()
@@ -482,34 +468,14 @@ class ExternalAppendOnlyMap[K, V, C](
   item
 }
 
-// TODO: Now only use `finalize` to ensure `close` gets called to clean up 
the resources. In the
-// future, we need some mechanism to ensure this gets called once the 
resources are not used.
-private def cleanup(): Unit = {
-  if (!closed) {
-closed = true
-batchIndex = batchOffsets.length  // Prevent reading any other batch
-fileStream = null
-try {
-  val ds = deserializeStreamToBeClosed
-  deserializeStreamToBeClosed = null
-  deserializeStream = null
-  if (ds != null) {
-ds.close()
-  }
-} finally {
-  if (file.exists()) {
-file.delete()
-  }
-}
-  }
-}
-
-override def finalize(): Unit = {
-  try {
-cleanup()
-  } finally {
-super.finalize()
-  }
+// TODO: Ensure this gets called even if the iterator isn't drained.
+private def cleanup() {
+  batchIndex = batchOffsets.length  // Prevent reading any other batch
+  val ds = deserializeStream
+  deserializeStream = null
+  fileStream = null
+  ds.close()
+  file.delete()
 }
   }
 


-

spark git commit: Revert [SPARK-5423][Core] Cleanup resources in DiskMapIterator.finalize to ensure deleting the temp file

2015-03-03 Thread andrewor14
Repository: spark
Updated Branches:
  refs/heads/branch-1.2 5226dc74d - eb30fc180


Revert [SPARK-5423][Core] Cleanup resources in DiskMapIterator.finalize to 
ensure deleting the temp file

This reverts commit 61bde0049fac324b5004eadfa22b02cd76cf2187.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/eb30fc18
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/eb30fc18
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/eb30fc18

Branch: refs/heads/branch-1.2
Commit: eb30fc180a8908ea294d6d0ad04e90e8a45813f4
Parents: 5226dc7
Author: Andrew Or and...@databricks.com
Authored: Tue Mar 3 13:04:50 2015 -0800
Committer: Andrew Or and...@databricks.com
Committed: Tue Mar 3 13:04:50 2015 -0800

--
 .../util/collection/ExternalAppendOnlyMap.scala | 52 
 1 file changed, 9 insertions(+), 43 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/eb30fc18/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
--
diff --git 
a/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
 
b/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
index fc7e86e..8a0f5a6 100644
--- 
a/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
+++ 
b/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala
@@ -387,15 +387,6 @@ class ExternalAppendOnlyMap[K, V, C](
 private var batchIndex = 0  // Which batch we're in
 private var fileStream: FileInputStream = null
 
-@volatile private var closed = false
-
-// A volatile variable to remember which DeserializationStream is using. 
Need to set it when we
-// open a DeserializationStream. But we should use `deserializeStream` 
rather than
-// `deserializeStreamToBeClosed` to read the content because touching a 
volatile variable will
-// reduce the performance. It must be volatile so that we can see its 
correct value in the
-// `finalize` method, which could run in any thread.
-@volatile private var deserializeStreamToBeClosed: DeserializationStream = 
null
-
 // An intermediate stream that reads from exactly one batch
 // This guards against pre-fetching and other arbitrary behavior of higher 
level streams
 private var deserializeStream = nextBatchStream()
@@ -410,7 +401,6 @@ class ExternalAppendOnlyMap[K, V, C](
   // we're still in a valid batch.
   if (batchIndex  batchOffsets.length - 1) {
 if (deserializeStream != null) {
-  deserializeStreamToBeClosed = null
   deserializeStream.close()
   fileStream.close()
   deserializeStream = null
@@ -429,11 +419,7 @@ class ExternalAppendOnlyMap[K, V, C](
 
 val bufferedStream = new 
BufferedInputStream(ByteStreams.limit(fileStream, end - start))
 val compressedStream = blockManager.wrapForCompression(blockId, 
bufferedStream)
-// Before returning the stream, assign it to 
`deserializeStreamToBeClosed` so that we can
-// close it in `finalize` and also avoid to touch the volatile 
`deserializeStreamToBeClosed`
-// during reading the (K, C) pairs.
-deserializeStreamToBeClosed = ser.deserializeStream(compressedStream)
-deserializeStreamToBeClosed
+ser.deserializeStream(compressedStream)
   } else {
 // No more batches left
 cleanup()
@@ -482,34 +468,14 @@ class ExternalAppendOnlyMap[K, V, C](
   item
 }
 
-// TODO: Now only use `finalize` to ensure `close` gets called to clean up 
the resources. In the
-// future, we need some mechanism to ensure this gets called once the 
resources are not used.
-private def cleanup(): Unit = {
-  if (!closed) {
-closed = true
-batchIndex = batchOffsets.length  // Prevent reading any other batch
-fileStream = null
-try {
-  val ds = deserializeStreamToBeClosed
-  deserializeStreamToBeClosed = null
-  deserializeStream = null
-  if (ds != null) {
-ds.close()
-  }
-} finally {
-  if (file.exists()) {
-file.delete()
-  }
-}
-  }
-}
-
-override def finalize(): Unit = {
-  try {
-cleanup()
-  } finally {
-super.finalize()
-  }
+// TODO: Ensure this gets called even if the iterator isn't drained.
+private def cleanup() {
+  batchIndex = batchOffsets.length  // Prevent reading any other batch
+  val ds = deserializeStream
+  deserializeStream = null
+  fileStream = null
+  ds.close()
+  file.delete()
 }
   }
 


-

spark git commit: SPARK-1911 [DOCS] Warn users if their assembly jars are not built with Java 6

2015-03-03 Thread andrewor14
Repository: spark
Updated Branches:
  refs/heads/branch-1.3 ee4929d1d - 8446ad0eb


SPARK-1911 [DOCS] Warn users if their assembly jars are not built with Java 6

Add warning about building with Java 7+ and running the JAR on early Java 6.

CC andrewor14

Author: Sean Owen so...@cloudera.com

Closes #4874 from srowen/SPARK-1911 and squashes the following commits:

79fa2f6 [Sean Owen] Add warning about building with Java 7+ and running the JAR 
on early Java 6.

(cherry picked from commit e750a6bfddf1d7bf7d3e99a424ec2b83a18b40d9)
Signed-off-by: Andrew Or and...@databricks.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/8446ad0e
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/8446ad0e
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/8446ad0e

Branch: refs/heads/branch-1.3
Commit: 8446ad0ebd2abb10ef405dc2ce53d2724604ce83
Parents: ee4929d
Author: Sean Owen so...@cloudera.com
Authored: Tue Mar 3 13:40:11 2015 -0800
Committer: Andrew Or and...@databricks.com
Committed: Tue Mar 3 13:40:16 2015 -0800

--
 docs/building-spark.md | 4 
 1 file changed, 4 insertions(+)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/8446ad0e/docs/building-spark.md
--
diff --git a/docs/building-spark.md b/docs/building-spark.md
index b666d14..d12ee8b 100644
--- a/docs/building-spark.md
+++ b/docs/building-spark.md
@@ -9,6 +9,10 @@ redirect_from: building-with-maven.html
 
 Building Spark using Maven requires Maven 3.0.4 or newer and Java 6+.
 
+**Note:** Building Spark with Java 7 or later can create JAR files that may 
not be
+readable with early versions of Java 6, due to the large number of files in 
the JAR
+archive. Build with Java 6 if this is an issue for your deployment.
+
 # Building with `build/mvn`
 
 Spark now comes packaged with a self-contained Maven installation to ease 
building and deployment of Spark from source located under the `build/` 
directory. This script will automatically download and setup all necessary 
build requirements ([Maven](https://maven.apache.org/), 
[Scala](http://www.scala-lang.org/), and 
[Zinc](https://github.com/typesafehub/zinc)) locally within the `build/` 
directory itself. It honors any `mvn` binary if present already, however, will 
pull down its own copy of Scala and Zinc regardless to ensure proper version 
requirements are met. `build/mvn` execution acts as a pass through to the `mvn` 
call allowing easy transition from previous build methods. As an example, one 
can build a version of Spark as follows:


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



spark git commit: [SPARK-6141][MLlib] Upgrade Breeze from 0.10 to 0.11 to fix convergence bug

2015-03-03 Thread meng
Repository: spark
Updated Branches:
  refs/heads/branch-1.3 9a0b75cdd - 9f249779f


[SPARK-6141][MLlib] Upgrade Breeze from 0.10 to 0.11 to fix convergence bug

LBFGS and OWLQN in Breeze 0.10 has convergence check bug.
This is fixed in 0.11, see the description in Breeze project for detail:

https://github.com/scalanlp/breeze/pull/373#issuecomment-76879760

Author: Xiangrui Meng m...@databricks.com
Author: DB Tsai dbt...@alpinenow.com
Author: DB Tsai dbt...@dbtsai.com

Closes #4879 from dbtsai/breeze and squashes the following commits:

d848f65 [DB Tsai] Merge pull request #1 from mengxr/AlpineNow-breeze
c2ca6ac [Xiangrui Meng] upgrade to breeze-0.11.1
35c2f26 [Xiangrui Meng] fix LRSuite
397a208 [DB Tsai] upgrade breeze

(cherry picked from commit 76e20a0a03cf2c02db35e00271924efb070eaaa5)
Signed-off-by: Xiangrui Meng m...@databricks.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/9f249779
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/9f249779
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/9f249779

Branch: refs/heads/branch-1.3
Commit: 9f249779ffe65131d1b9b95154a8ccd3a89fe022
Parents: 9a0b75c
Author: Xiangrui Meng m...@databricks.com
Authored: Tue Mar 3 23:52:02 2015 -0800
Committer: Xiangrui Meng m...@databricks.com
Committed: Tue Mar 3 23:52:10 2015 -0800

--
 mllib/pom.xml| 2 +-
 .../spark/mllib/classification/LogisticRegressionSuite.scala | 4 
 2 files changed, 5 insertions(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/9f249779/mllib/pom.xml
--
diff --git a/mllib/pom.xml b/mllib/pom.xml
index 32e9847..48c9d06 100644
--- a/mllib/pom.xml
+++ b/mllib/pom.xml
@@ -63,7 +63,7 @@
 dependency
   groupIdorg.scalanlp/groupId
   artifactIdbreeze_${scala.binary.version}/artifactId
-  version0.10/version
+  version0.11.1/version
   exclusions
 !-- This is included as a compile-scoped dependency by jtransforms, 
which is
  a dependency of breeze. --

http://git-wip-us.apache.org/repos/asf/spark/blob/9f249779/mllib/src/test/scala/org/apache/spark/mllib/classification/LogisticRegressionSuite.scala
--
diff --git 
a/mllib/src/test/scala/org/apache/spark/mllib/classification/LogisticRegressionSuite.scala
 
b/mllib/src/test/scala/org/apache/spark/mllib/classification/LogisticRegressionSuite.scala
index d2b40f2..aaa81da 100644
--- 
a/mllib/src/test/scala/org/apache/spark/mllib/classification/LogisticRegressionSuite.scala
+++ 
b/mllib/src/test/scala/org/apache/spark/mllib/classification/LogisticRegressionSuite.scala
@@ -372,8 +372,12 @@ class LogisticRegressionSuite extends FunSuite with 
MLlibTestSparkContext with M
 testRDD2.cache()
 testRDD3.cache()
 
+val numIteration = 10
+
 val lrA = new LogisticRegressionWithLBFGS().setIntercept(true)
+lrA.optimizer.setNumIterations(numIteration)
 val lrB = new 
LogisticRegressionWithLBFGS().setIntercept(true).setFeatureScaling(false)
+lrB.optimizer.setNumIterations(numIteration)
 
 val modelA1 = lrA.run(testRDD1, initialWeights)
 val modelA2 = lrA.run(testRDD2, initialWeights)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



spark git commit: [SPARK-5949] HighlyCompressedMapStatus needs more classes registered w/ kryo

2015-03-03 Thread rxin
Repository: spark
Updated Branches:
  refs/heads/branch-1.3 8446ad0eb - 9a0b75cdd


[SPARK-5949] HighlyCompressedMapStatus needs more classes registered w/ kryo

https://issues.apache.org/jira/browse/SPARK-5949

Author: Imran Rashid iras...@cloudera.com

Closes #4877 from squito/SPARK-5949_register_roaring_bitmap and squashes the 
following commits:

7e13316 [Imran Rashid] style style style
5f6bb6d [Imran Rashid] more style
709bfe0 [Imran Rashid] style
a5cb744 [Imran Rashid] update tests to cover both types of 
RoaringBitmapContainers
09610c6 [Imran Rashid] formatting
f9a0b7c [Imran Rashid] put primitive array registrations together
97beaf8 [Imran Rashid] SPARK-5949 HighlyCompressedMapStatus needs more classes 
registered w/ kryo

(cherry picked from commit 1f1fccc5ceb0c5b7656a0594be3a67bd3b432e85)
Signed-off-by: Reynold Xin r...@databricks.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/9a0b75cd
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/9a0b75cd
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/9a0b75cd

Branch: refs/heads/branch-1.3
Commit: 9a0b75cdd85c1933c590480dd233b1803726ed71
Parents: 8446ad0
Author: Imran Rashid iras...@cloudera.com
Authored: Tue Mar 3 15:33:19 2015 -0800
Committer: Reynold Xin r...@databricks.com
Committed: Tue Mar 3 15:33:26 2015 -0800

--
 .../spark/serializer/KryoSerializer.scala   | 15 ++---
 .../spark/serializer/KryoSerializerSuite.scala  | 23 ++--
 2 files changed, 33 insertions(+), 5 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/9a0b75cd/core/src/main/scala/org/apache/spark/serializer/KryoSerializer.scala
--
diff --git 
a/core/src/main/scala/org/apache/spark/serializer/KryoSerializer.scala 
b/core/src/main/scala/org/apache/spark/serializer/KryoSerializer.scala
index 02158aa..9ce64d4 100644
--- a/core/src/main/scala/org/apache/spark/serializer/KryoSerializer.scala
+++ b/core/src/main/scala/org/apache/spark/serializer/KryoSerializer.scala
@@ -20,22 +20,23 @@ package org.apache.spark.serializer
 import java.io.{EOFException, InputStream, OutputStream}
 import java.nio.ByteBuffer
 
+import scala.reflect.ClassTag
+
 import com.esotericsoftware.kryo.{Kryo, KryoException}
 import com.esotericsoftware.kryo.io.{Input = KryoInput, Output = KryoOutput}
 import com.esotericsoftware.kryo.serializers.{JavaSerializer = 
KryoJavaSerializer}
 import com.twitter.chill.{AllScalaRegistrar, EmptyScalaKryoInstantiator}
+import org.roaringbitmap.{ArrayContainer, BitmapContainer, RoaringArray, 
RoaringBitmap}
 
 import org.apache.spark._
 import org.apache.spark.api.python.PythonBroadcast
 import org.apache.spark.broadcast.HttpBroadcast
-import org.apache.spark.network.nio.{PutBlock, GotBlock, GetBlock}
+import org.apache.spark.network.nio.{GetBlock, GotBlock, PutBlock}
 import org.apache.spark.scheduler.{CompressedMapStatus, 
HighlyCompressedMapStatus}
 import org.apache.spark.storage._
 import org.apache.spark.util.BoundedPriorityQueue
 import org.apache.spark.util.collection.CompactBuffer
 
-import scala.reflect.ClassTag
-
 /**
  * A Spark serializer that uses the [[https://code.google.com/p/kryo/ Kryo 
serialization library]].
  *
@@ -202,9 +203,17 @@ private[serializer] object KryoSerializer {
 classOf[GetBlock],
 classOf[CompressedMapStatus],
 classOf[HighlyCompressedMapStatus],
+classOf[RoaringBitmap],
+classOf[RoaringArray],
+classOf[RoaringArray.Element],
+classOf[Array[RoaringArray.Element]],
+classOf[ArrayContainer],
+classOf[BitmapContainer],
 classOf[CompactBuffer[_]],
 classOf[BlockManagerId],
 classOf[Array[Byte]],
+classOf[Array[Short]],
+classOf[Array[Long]],
 classOf[BoundedPriorityQueue[_]],
 classOf[SparkConf]
   )

http://git-wip-us.apache.org/repos/asf/spark/blob/9a0b75cd/core/src/test/scala/org/apache/spark/serializer/KryoSerializerSuite.scala
--
diff --git 
a/core/src/test/scala/org/apache/spark/serializer/KryoSerializerSuite.scala 
b/core/src/test/scala/org/apache/spark/serializer/KryoSerializerSuite.scala
index a70f67a..523d898 100644
--- a/core/src/test/scala/org/apache/spark/serializer/KryoSerializerSuite.scala
+++ b/core/src/test/scala/org/apache/spark/serializer/KryoSerializerSuite.scala
@@ -23,9 +23,10 @@ import scala.reflect.ClassTag
 import com.esotericsoftware.kryo.Kryo
 import org.scalatest.FunSuite
 
-import org.apache.spark.{SparkConf, SharedSparkContext}
+import org.apache.spark.{SharedSparkContext, SparkConf}
+import org.apache.spark.scheduler.HighlyCompressedMapStatus
 import org.apache.spark.serializer.KryoTest._
-
+import org.apache.spark.storage.BlockManagerId
 
 class 

spark git commit: [SPARK-5949] HighlyCompressedMapStatus needs more classes registered w/ kryo

2015-03-03 Thread rxin
Repository: spark
Updated Branches:
  refs/heads/master 6c20f3529 - 1f1fccc5c


[SPARK-5949] HighlyCompressedMapStatus needs more classes registered w/ kryo

https://issues.apache.org/jira/browse/SPARK-5949

Author: Imran Rashid iras...@cloudera.com

Closes #4877 from squito/SPARK-5949_register_roaring_bitmap and squashes the 
following commits:

7e13316 [Imran Rashid] style style style
5f6bb6d [Imran Rashid] more style
709bfe0 [Imran Rashid] style
a5cb744 [Imran Rashid] update tests to cover both types of 
RoaringBitmapContainers
09610c6 [Imran Rashid] formatting
f9a0b7c [Imran Rashid] put primitive array registrations together
97beaf8 [Imran Rashid] SPARK-5949 HighlyCompressedMapStatus needs more classes 
registered w/ kryo


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/1f1fccc5
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/1f1fccc5
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/1f1fccc5

Branch: refs/heads/master
Commit: 1f1fccc5ceb0c5b7656a0594be3a67bd3b432e85
Parents: 6c20f35
Author: Imran Rashid iras...@cloudera.com
Authored: Tue Mar 3 15:33:19 2015 -0800
Committer: Reynold Xin r...@databricks.com
Committed: Tue Mar 3 15:33:19 2015 -0800

--
 .../spark/serializer/KryoSerializer.scala   | 15 ++---
 .../spark/serializer/KryoSerializerSuite.scala  | 23 ++--
 2 files changed, 33 insertions(+), 5 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/1f1fccc5/core/src/main/scala/org/apache/spark/serializer/KryoSerializer.scala
--
diff --git 
a/core/src/main/scala/org/apache/spark/serializer/KryoSerializer.scala 
b/core/src/main/scala/org/apache/spark/serializer/KryoSerializer.scala
index 02158aa..9ce64d4 100644
--- a/core/src/main/scala/org/apache/spark/serializer/KryoSerializer.scala
+++ b/core/src/main/scala/org/apache/spark/serializer/KryoSerializer.scala
@@ -20,22 +20,23 @@ package org.apache.spark.serializer
 import java.io.{EOFException, InputStream, OutputStream}
 import java.nio.ByteBuffer
 
+import scala.reflect.ClassTag
+
 import com.esotericsoftware.kryo.{Kryo, KryoException}
 import com.esotericsoftware.kryo.io.{Input = KryoInput, Output = KryoOutput}
 import com.esotericsoftware.kryo.serializers.{JavaSerializer = 
KryoJavaSerializer}
 import com.twitter.chill.{AllScalaRegistrar, EmptyScalaKryoInstantiator}
+import org.roaringbitmap.{ArrayContainer, BitmapContainer, RoaringArray, 
RoaringBitmap}
 
 import org.apache.spark._
 import org.apache.spark.api.python.PythonBroadcast
 import org.apache.spark.broadcast.HttpBroadcast
-import org.apache.spark.network.nio.{PutBlock, GotBlock, GetBlock}
+import org.apache.spark.network.nio.{GetBlock, GotBlock, PutBlock}
 import org.apache.spark.scheduler.{CompressedMapStatus, 
HighlyCompressedMapStatus}
 import org.apache.spark.storage._
 import org.apache.spark.util.BoundedPriorityQueue
 import org.apache.spark.util.collection.CompactBuffer
 
-import scala.reflect.ClassTag
-
 /**
  * A Spark serializer that uses the [[https://code.google.com/p/kryo/ Kryo 
serialization library]].
  *
@@ -202,9 +203,17 @@ private[serializer] object KryoSerializer {
 classOf[GetBlock],
 classOf[CompressedMapStatus],
 classOf[HighlyCompressedMapStatus],
+classOf[RoaringBitmap],
+classOf[RoaringArray],
+classOf[RoaringArray.Element],
+classOf[Array[RoaringArray.Element]],
+classOf[ArrayContainer],
+classOf[BitmapContainer],
 classOf[CompactBuffer[_]],
 classOf[BlockManagerId],
 classOf[Array[Byte]],
+classOf[Array[Short]],
+classOf[Array[Long]],
 classOf[BoundedPriorityQueue[_]],
 classOf[SparkConf]
   )

http://git-wip-us.apache.org/repos/asf/spark/blob/1f1fccc5/core/src/test/scala/org/apache/spark/serializer/KryoSerializerSuite.scala
--
diff --git 
a/core/src/test/scala/org/apache/spark/serializer/KryoSerializerSuite.scala 
b/core/src/test/scala/org/apache/spark/serializer/KryoSerializerSuite.scala
index a70f67a..523d898 100644
--- a/core/src/test/scala/org/apache/spark/serializer/KryoSerializerSuite.scala
+++ b/core/src/test/scala/org/apache/spark/serializer/KryoSerializerSuite.scala
@@ -23,9 +23,10 @@ import scala.reflect.ClassTag
 import com.esotericsoftware.kryo.Kryo
 import org.scalatest.FunSuite
 
-import org.apache.spark.{SparkConf, SharedSparkContext}
+import org.apache.spark.{SharedSparkContext, SparkConf}
+import org.apache.spark.scheduler.HighlyCompressedMapStatus
 import org.apache.spark.serializer.KryoTest._
-
+import org.apache.spark.storage.BlockManagerId
 
 class KryoSerializerSuite extends FunSuite with SharedSparkContext {
   conf.set(spark.serializer, org.apache.spark.serializer.KryoSerializer)

spark git commit: [SPARK-6141][MLlib] Upgrade Breeze from 0.10 to 0.11 to fix convergence bug

2015-03-03 Thread meng
Repository: spark
Updated Branches:
  refs/heads/master d334bfbcf - 76e20a0a0


[SPARK-6141][MLlib] Upgrade Breeze from 0.10 to 0.11 to fix convergence bug

LBFGS and OWLQN in Breeze 0.10 has convergence check bug.
This is fixed in 0.11, see the description in Breeze project for detail:

https://github.com/scalanlp/breeze/pull/373#issuecomment-76879760

Author: Xiangrui Meng m...@databricks.com
Author: DB Tsai dbt...@alpinenow.com
Author: DB Tsai dbt...@dbtsai.com

Closes #4879 from dbtsai/breeze and squashes the following commits:

d848f65 [DB Tsai] Merge pull request #1 from mengxr/AlpineNow-breeze
c2ca6ac [Xiangrui Meng] upgrade to breeze-0.11.1
35c2f26 [Xiangrui Meng] fix LRSuite
397a208 [DB Tsai] upgrade breeze


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/76e20a0a
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/76e20a0a
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/76e20a0a

Branch: refs/heads/master
Commit: 76e20a0a03cf2c02db35e00271924efb070eaaa5
Parents: d334bfb
Author: Xiangrui Meng m...@databricks.com
Authored: Tue Mar 3 23:52:02 2015 -0800
Committer: Xiangrui Meng m...@databricks.com
Committed: Tue Mar 3 23:52:02 2015 -0800

--
 mllib/pom.xml| 2 +-
 .../spark/mllib/classification/LogisticRegressionSuite.scala | 4 
 2 files changed, 5 insertions(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/76e20a0a/mllib/pom.xml
--
diff --git a/mllib/pom.xml b/mllib/pom.xml
index a8cee3d..4c8f344 100644
--- a/mllib/pom.xml
+++ b/mllib/pom.xml
@@ -63,7 +63,7 @@
 dependency
   groupIdorg.scalanlp/groupId
   artifactIdbreeze_${scala.binary.version}/artifactId
-  version0.10/version
+  version0.11.1/version
   exclusions
 !-- This is included as a compile-scoped dependency by jtransforms, 
which is
  a dependency of breeze. --

http://git-wip-us.apache.org/repos/asf/spark/blob/76e20a0a/mllib/src/test/scala/org/apache/spark/mllib/classification/LogisticRegressionSuite.scala
--
diff --git 
a/mllib/src/test/scala/org/apache/spark/mllib/classification/LogisticRegressionSuite.scala
 
b/mllib/src/test/scala/org/apache/spark/mllib/classification/LogisticRegressionSuite.scala
index d2b40f2..aaa81da 100644
--- 
a/mllib/src/test/scala/org/apache/spark/mllib/classification/LogisticRegressionSuite.scala
+++ 
b/mllib/src/test/scala/org/apache/spark/mllib/classification/LogisticRegressionSuite.scala
@@ -372,8 +372,12 @@ class LogisticRegressionSuite extends FunSuite with 
MLlibTestSparkContext with M
 testRDD2.cache()
 testRDD3.cache()
 
+val numIteration = 10
+
 val lrA = new LogisticRegressionWithLBFGS().setIntercept(true)
+lrA.optimizer.setNumIterations(numIteration)
 val lrB = new 
LogisticRegressionWithLBFGS().setIntercept(true).setFeatureScaling(false)
+lrB.optimizer.setNumIterations(numIteration)
 
 val modelA1 = lrA.run(testRDD1, initialWeights)
 val modelA2 = lrA.run(testRDD2, initialWeights)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



spark git commit: [SPARK-6133] Make sc.stop() idempotent

2015-03-03 Thread andrewor14
Repository: spark
Updated Branches:
  refs/heads/master fe63e8229 - 6c20f3529


[SPARK-6133] Make sc.stop() idempotent

Before we would get the following (benign) error if we called `sc.stop()` 
twice. This is because the listener bus would try to post the end event again 
even after it has already stopped. This happens occasionally when flaky tests 
fail, usually as a result of other sources of error. Either way we shouldn't be 
logging this error when it is not the cause of the failure.
```
ERROR LiveListenerBus: SparkListenerBus has already stopped! Dropping event 
SparkListenerApplicationEnd(1425348445682)
```

Author: Andrew Or and...@databricks.com

Closes #4871 from andrewor14/sc-stop and squashes the following commits:

a14afc5 [Andrew Or] Move code after code
915db16 [Andrew Or] Move code into code


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/6c20f352
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/6c20f352
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/6c20f352

Branch: refs/heads/master
Commit: 6c20f35290e220e4a659a0222d62575ff959d703
Parents: fe63e82
Author: Andrew Or and...@databricks.com
Authored: Tue Mar 3 15:09:57 2015 -0800
Committer: Andrew Or and...@databricks.com
Committed: Tue Mar 3 15:09:57 2015 -0800

--
 core/src/main/scala/org/apache/spark/SparkContext.scala | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/6c20f352/core/src/main/scala/org/apache/spark/SparkContext.scala
--
diff --git a/core/src/main/scala/org/apache/spark/SparkContext.scala 
b/core/src/main/scala/org/apache/spark/SparkContext.scala
index e231e83..1a0bee4 100644
--- a/core/src/main/scala/org/apache/spark/SparkContext.scala
+++ b/core/src/main/scala/org/apache/spark/SparkContext.scala
@@ -1392,10 +1392,10 @@ class SparkContext(config: SparkConf) extends Logging 
with ExecutorAllocationCli
   /** Shut down the SparkContext. */
   def stop() {
 SparkContext.SPARK_CONTEXT_CONSTRUCTOR_LOCK.synchronized {
-  postApplicationEnd()
-  ui.foreach(_.stop())
   if (!stopped) {
 stopped = true
+postApplicationEnd()
+ui.foreach(_.stop())
 env.metricsSystem.report()
 metadataCleaner.cancel()
 cleaner.foreach(_.stop())


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



spark git commit: [SPARK-6133] Make sc.stop() idempotent

2015-03-03 Thread andrewor14
Repository: spark
Updated Branches:
  refs/heads/branch-1.2 eb30fc180 - a91c1c594


[SPARK-6133] Make sc.stop() idempotent

Before we would get the following (benign) error if we called `sc.stop()` 
twice. This is because the listener bus would try to post the end event again 
even after it has already stopped. This happens occasionally when flaky tests 
fail, usually as a result of other sources of error. Either way we shouldn't be 
logging this error when it is not the cause of the failure.
```
ERROR LiveListenerBus: SparkListenerBus has already stopped! Dropping event 
SparkListenerApplicationEnd(1425348445682)
```

Author: Andrew Or and...@databricks.com

Closes #4871 from andrewor14/sc-stop and squashes the following commits:

a14afc5 [Andrew Or] Move code after code
915db16 [Andrew Or] Move code into code

(cherry picked from commit 6c20f35290e220e4a659a0222d62575ff959d703)
Signed-off-by: Andrew Or and...@databricks.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/a91c1c59
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/a91c1c59
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/a91c1c59

Branch: refs/heads/branch-1.2
Commit: a91c1c594c57d4bb0a3283de239ca1d1bd0d7741
Parents: eb30fc1
Author: Andrew Or and...@databricks.com
Authored: Tue Mar 3 15:09:57 2015 -0800
Committer: Andrew Or and...@databricks.com
Committed: Tue Mar 3 15:10:03 2015 -0800

--
 core/src/main/scala/org/apache/spark/SparkContext.scala | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/a91c1c59/core/src/main/scala/org/apache/spark/SparkContext.scala
--
diff --git a/core/src/main/scala/org/apache/spark/SparkContext.scala 
b/core/src/main/scala/org/apache/spark/SparkContext.scala
index b50a541..98e2fed 100644
--- a/core/src/main/scala/org/apache/spark/SparkContext.scala
+++ b/core/src/main/scala/org/apache/spark/SparkContext.scala
@@ -1229,10 +1229,10 @@ class SparkContext(config: SparkConf) extends Logging 
with ExecutorAllocationCli
   /** Shut down the SparkContext. */
   def stop() {
 SparkContext.SPARK_CONTEXT_CONSTRUCTOR_LOCK.synchronized {
-  postApplicationEnd()
-  ui.foreach(_.stop())
   if (!stopped) {
 stopped = true
+postApplicationEnd()
+ui.foreach(_.stop())
 env.metricsSystem.report()
 metadataCleaner.cancel()
 env.actorSystem.stop(heartbeatReceiver)


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



spark git commit: [SPARK-6132][HOTFIX] ContextCleaner InterruptedException should be quiet

2015-03-03 Thread andrewor14
Repository: spark
Updated Branches:
  refs/heads/master 1f1fccc5c - d334bfbcf


[SPARK-6132][HOTFIX] ContextCleaner InterruptedException should be quiet

If the cleaner is stopped, we shouldn't print a huge stack trace when the 
cleaner thread is interrupted because we purposefully did this.

Author: Andrew Or and...@databricks.com

Closes #4882 from andrewor14/cleaner-interrupt and squashes the following 
commits:

8652120 [Andrew Or] Just a hot fix


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/d334bfbc
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/d334bfbc
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/d334bfbc

Branch: refs/heads/master
Commit: d334bfbcf38d0c1c6835e72bc85e42196e295744
Parents: 1f1fccc
Author: Andrew Or and...@databricks.com
Authored: Tue Mar 3 20:49:45 2015 -0800
Committer: Andrew Or and...@databricks.com
Committed: Tue Mar 3 20:49:45 2015 -0800

--
 core/src/main/scala/org/apache/spark/ContextCleaner.scala | 1 +
 1 file changed, 1 insertion(+)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/d334bfbc/core/src/main/scala/org/apache/spark/ContextCleaner.scala
--
diff --git a/core/src/main/scala/org/apache/spark/ContextCleaner.scala 
b/core/src/main/scala/org/apache/spark/ContextCleaner.scala
index 4dab886..0c59a61 100644
--- a/core/src/main/scala/org/apache/spark/ContextCleaner.scala
+++ b/core/src/main/scala/org/apache/spark/ContextCleaner.scala
@@ -168,6 +168,7 @@ private[spark] class ContextCleaner(sc: SparkContext) 
extends Logging {
   }
 }
   } catch {
+case ie: InterruptedException if stopped = // ignore
 case e: Exception = logError(Error in cleaning thread, e)
   }
 }


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



spark git commit: BUILD: Minor tweaks to internal build scripts

2015-03-03 Thread pwendell
Repository: spark
Updated Branches:
  refs/heads/master 165ff3642 - 0c9a8eaed


BUILD: Minor tweaks to internal build scripts

This adds two features:
1. The ability to publish with a different maven version than
   that specified in the release source.
2. Forking of different Zinc instances during the parallel dist
   creation (to help with some stability issues).


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/0c9a8eae
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/0c9a8eae
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/0c9a8eae

Branch: refs/heads/master
Commit: 0c9a8eaed74b2b381a1cc70e6e6a2783bac0912c
Parents: 165ff36
Author: Patrick Wendell patr...@databricks.com
Authored: Tue Mar 3 00:38:12 2015 -0800
Committer: Patrick Wendell patr...@databricks.com
Committed: Tue Mar 3 01:53:48 2015 -0800

--
 dev/create-release/create-release.sh | 24 +++-
 1 file changed, 19 insertions(+), 5 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/0c9a8eae/dev/create-release/create-release.sh
--
diff --git a/dev/create-release/create-release.sh 
b/dev/create-release/create-release.sh
index da15ce3..6f87fcd 100755
--- a/dev/create-release/create-release.sh
+++ b/dev/create-release/create-release.sh
@@ -34,6 +34,9 @@ ASF_PASSWORD=${ASF_PASSWORD:-XXX}
 GPG_PASSPHRASE=${GPG_PASSPHRASE:-XXX}
 GIT_BRANCH=${GIT_BRANCH:-branch-1.0}
 RELEASE_VERSION=${RELEASE_VERSION:-1.2.0}
+# Allows publishing under a different version identifier than
+# was present in the actual release sources (e.g. rc-X)
+PUBLISH_VERSION=${PUBLISH_VERSION:-$RELEASE_VERSION} 
 NEXT_VERSION=${NEXT_VERSION:-1.2.1}
 RC_NAME=${RC_NAME:-rc2}
 
@@ -97,30 +100,35 @@ if [[ ! $@ =~ --skip-publish ]]; then
   pushd spark
   git checkout --force $GIT_TAG 
   
+  # Substitute in case published version is different than released
+  old=^\( \{2,4\}\)version${RELEASE_VERSION}\/version$
+  new=\1version${PUBLISH_VERSION}\/version
+  find . -name pom.xml | grep -v dev | xargs -I {} sed -i \
+-e s/${old}/${new}/ {}
+
   # Using Nexus API documented here:
   # 
https://support.sonatype.com/entries/39720203-Uploading-to-a-Staging-Repository-via-REST-API
   echo Creating Nexus staging repository
-  repo_request=promoteRequestdatadescriptionApache Spark 
$GIT_TAG/description/data/promoteRequest
+  repo_request=promoteRequestdatadescriptionApache Spark $GIT_TAG 
(published as $PUBLISH_VERSION)/description/data/promoteRequest
   out=$(curl -X POST -d $repo_request -u $ASF_USERNAME:$ASF_PASSWORD \
 -H Content-Type:application/xml -v \
 $NEXUS_ROOT/profiles/$NEXUS_PROFILE/start)
   staged_repo_id=$(echo $out | sed -e 
s/.*\(orgapachespark-[0-9]\{4\}\).*/\1/)
   echo Created Nexus staging repository: $staged_repo_id
 
-  rm -rf $SPARK_REPO
-
-  mvn -DskipTests -Dhadoop.version=2.2.0 -Dyarn.version=2.2.0 \
+  build/mvn -DskipTests -Dhadoop.version=2.2.0 -Dyarn.version=2.2.0 \
 -Pyarn -Phive -Phadoop-2.2 -Pspark-ganglia-lgpl -Pkinesis-asl \
 clean install
 
   ./dev/change-version-to-2.11.sh
   
-  mvn -DskipTests -Dhadoop.version=2.2.0 -Dyarn.version=2.2.0 \
+  build/mvn -DskipTests -Dhadoop.version=2.2.0 -Dyarn.version=2.2.0 \
 -Dscala-2.11 -Pyarn -Phive -Phadoop-2.2 -Pspark-ganglia-lgpl -Pkinesis-asl 
\
 clean install
 
   ./dev/change-version-to-2.10.sh
 
+  rm -rf $SPARK_REPO
   pushd $SPARK_REPO
 
   # Remove any extra files generated during install
@@ -197,6 +205,12 @@ if [[ ! $@ =~ --skip-package ]]; then
   ./dev/change-version-to-2.11.sh
 fi
 
+# Create new Zinc instances for each binary release to avoid interference
+# that causes OOM's and random compiler crashes.
+zinc_port=${zinc_port:-3030}
+zinc_port=$[$zinc_port + 1]
+export ZINC_PORT=$zinc_port
+
 ./make-distribution.sh --name $NAME --tgz $FLAGS 21 | tee 
../binary-release-$NAME.log
 cd ..
 cp spark-$RELEASE_VERSION-bin-$NAME/spark-$RELEASE_VERSION-bin-$NAME.tgz .


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[2/2] spark git commit: Preparing Spark release v1.3.0-rc2

2015-03-03 Thread pwendell
Preparing Spark release v1.3.0-rc2


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/3af26870
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/3af26870
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/3af26870

Branch: refs/heads/branch-1.3
Commit: 3af26870e5163438868c4eb2df88380a533bb232
Parents: b012ed1
Author: Patrick Wendell patr...@databricks.com
Authored: Tue Mar 3 10:23:07 2015 +
Committer: Patrick Wendell patr...@databricks.com
Committed: Tue Mar 3 10:23:07 2015 +

--
 assembly/pom.xml  | 2 +-
 bagel/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 examples/pom.xml  | 2 +-
 external/flume-sink/pom.xml   | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka-assembly/pom.xml   | 2 +-
 external/kafka/pom.xml| 2 +-
 external/mqtt/pom.xml | 2 +-
 external/twitter/pom.xml  | 2 +-
 external/zeromq/pom.xml   | 2 +-
 extras/java8-tests/pom.xml| 2 +-
 extras/kinesis-asl/pom.xml| 2 +-
 extras/spark-ganglia-lgpl/pom.xml | 2 +-
 graphx/pom.xml| 2 +-
 mllib/pom.xml | 2 +-
 network/common/pom.xml| 2 +-
 network/shuffle/pom.xml   | 2 +-
 network/yarn/pom.xml  | 2 +-
 pom.xml   | 2 +-
 repl/pom.xml  | 2 +-
 sql/catalyst/pom.xml  | 2 +-
 sql/core/pom.xml  | 2 +-
 sql/hive-thriftserver/pom.xml | 2 +-
 sql/hive/pom.xml  | 2 +-
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 yarn/pom.xml  | 2 +-
 28 files changed, 28 insertions(+), 28 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/3af26870/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index c1bcdbb..0764d2d 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.3.0-SNAPSHOT/version
+version1.3.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/3af26870/bagel/pom.xml
--
diff --git a/bagel/pom.xml b/bagel/pom.xml
index 510e926..4f73cf7 100644
--- a/bagel/pom.xml
+++ b/bagel/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.3.0-SNAPSHOT/version
+version1.3.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/3af26870/core/pom.xml
--
diff --git a/core/pom.xml b/core/pom.xml
index c993781..5612149 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.3.0-SNAPSHOT/version
+version1.3.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/3af26870/examples/pom.xml
--
diff --git a/examples/pom.xml b/examples/pom.xml
index 8caad2b..f7d6030 100644
--- a/examples/pom.xml
+++ b/examples/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.3.0-SNAPSHOT/version
+version1.3.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/3af26870/external/flume-sink/pom.xml
--
diff --git a/external/flume-sink/pom.xml b/external/flume-sink/pom.xml
index 0706f1e..45aa775 100644
--- a/external/flume-sink/pom.xml
+++ b/external/flume-sink/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.3.0-SNAPSHOT/version
+version1.3.0/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/3af26870/external/flume/pom.xml
--
diff --git a/external/flume/pom.xml b/external/flume/pom.xml
index 1f26813..455304f 100644
--- a/external/flume/pom.xml
+++ b/external/flume/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.3.0-SNAPSHOT/version
+version1.3.0/version
 relativePath../../pom.xml/relativePath
   /parent
 


Git Push Summary

2015-03-03 Thread pwendell
Repository: spark
Updated Tags:  refs/tags/v1.3.0-rc2 [created] 3af26870e

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



spark git commit: [SPARK-6118] making package name of deploy.worker.CommandUtils and deploy.CommandUtilsSuite consistent

2015-03-03 Thread srowen
Repository: spark
Updated Branches:
  refs/heads/master 0c9a8eaed - 975643c25


[SPARK-6118] making package name of deploy.worker.CommandUtils and 
deploy.CommandUtilsSuite consistent

https://issues.apache.org/jira/browse/SPARK-6118

I found that the object CommandUtils is placed under deploy.worker package, 
while CommandUtilsSuite is  under deploy

Conventionally, we put the implementation and unit test class under the same 
package

here, to minimize the change, I move CommandUtilsSuite to worker package,

**However, CommandUtils seems to contain some general methods (though only used 
by worker.* classes currently**,  we may also consider to replace CommonUtils

Author: CodingCat zhunans...@gmail.com

Closes #4856 from CodingCat/SPARK-6118 and squashes the following commits:

cb93700 [CodingCat] making package name consistent


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/975643c2
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/975643c2
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/975643c2

Branch: refs/heads/master
Commit: 975643c256e548601bf9015c8840c947df5446bf
Parents: 0c9a8ea
Author: CodingCat zhunans...@gmail.com
Authored: Tue Mar 3 10:32:57 2015 +
Committer: Sean Owen so...@cloudera.com
Committed: Tue Mar 3 10:32:57 2015 +

--
 .../apache/spark/deploy/CommandUtilsSuite.scala | 37 
 .../spark/deploy/worker/CommandUtilsSuite.scala | 36 +++
 2 files changed, 36 insertions(+), 37 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/975643c2/core/src/test/scala/org/apache/spark/deploy/CommandUtilsSuite.scala
--
diff --git 
a/core/src/test/scala/org/apache/spark/deploy/CommandUtilsSuite.scala 
b/core/src/test/scala/org/apache/spark/deploy/CommandUtilsSuite.scala
deleted file mode 100644
index 7915ee7..000
--- a/core/src/test/scala/org/apache/spark/deploy/CommandUtilsSuite.scala
+++ /dev/null
@@ -1,37 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one or more
- * contributor license agreements.  See the NOTICE file distributed with
- * this work for additional information regarding copyright ownership.
- * The ASF licenses this file to You under the Apache License, Version 2.0
- * (the License); you may not use this file except in compliance with
- * the License.  You may obtain a copy of the License at
- *
- *http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an AS IS BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-package org.apache.spark.deploy
-
-import org.apache.spark.deploy.worker.CommandUtils
-import org.apache.spark.util.Utils
-
-import org.scalatest.{FunSuite, Matchers}
-
-class CommandUtilsSuite extends FunSuite with Matchers {
-
-  test(set libraryPath correctly) {
-val appId = 12345-worker321-9876
-val sparkHome = sys.props.getOrElse(spark.test.home, 
fail(spark.test.home is not set!))
-val cmd = new Command(mainClass, Seq(), Map(), Seq(), 
Seq(libraryPathToB), Seq())
-val builder = CommandUtils.buildProcessBuilder(cmd, 512, sparkHome, t = t)
-val libraryPath = Utils.libraryPathEnvName
-val env = builder.environment
-env.keySet should contain(libraryPath)
-assert(env.get(libraryPath).startsWith(libraryPathToB))
-  }
-}

http://git-wip-us.apache.org/repos/asf/spark/blob/975643c2/core/src/test/scala/org/apache/spark/deploy/worker/CommandUtilsSuite.scala
--
diff --git 
a/core/src/test/scala/org/apache/spark/deploy/worker/CommandUtilsSuite.scala 
b/core/src/test/scala/org/apache/spark/deploy/worker/CommandUtilsSuite.scala
new file mode 100644
index 000..1c27d83
--- /dev/null
+++ b/core/src/test/scala/org/apache/spark/deploy/worker/CommandUtilsSuite.scala
@@ -0,0 +1,36 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the License); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an AS IS BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either 

spark git commit: [SPARK-6138][CORE][minor] enhance the `toArray` method in `SizeTrackingVector`

2015-03-03 Thread srowen
Repository: spark
Updated Branches:
  refs/heads/master 975643c25 - e359794ce


[SPARK-6138][CORE][minor] enhance the `toArray` method in `SizeTrackingVector`

Use array copy instead of `Iterator#toArray` to make it more efficient.

Author: Wenchen Fan cloud0...@outlook.com

Closes #4825 from cloud-fan/minor and squashes the following commits:

c933ee5 [Wenchen Fan] make toArray method just in parent
946a35b [Wenchen Fan] minor enhance


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/e359794c
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/e359794c
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/e359794c

Branch: refs/heads/master
Commit: e359794cec7d30ece38752f62dc2a1d3d26b8feb
Parents: 975643c
Author: Wenchen Fan cloud0...@outlook.com
Authored: Tue Mar 3 12:12:23 2015 +
Committer: Sean Owen so...@cloudera.com
Committed: Tue Mar 3 12:12:23 2015 +

--
 .../spark/util/collection/PrimitiveVector.scala  | 15 ---
 .../spark/util/collection/SizeTrackingVector.scala   |  7 ---
 2 files changed, 12 insertions(+), 10 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/e359794c/core/src/main/scala/org/apache/spark/util/collection/PrimitiveVector.scala
--
diff --git 
a/core/src/main/scala/org/apache/spark/util/collection/PrimitiveVector.scala 
b/core/src/main/scala/org/apache/spark/util/collection/PrimitiveVector.scala
index 7e76d06..b6c380a 100644
--- a/core/src/main/scala/org/apache/spark/util/collection/PrimitiveVector.scala
+++ b/core/src/main/scala/org/apache/spark/util/collection/PrimitiveVector.scala
@@ -71,12 +71,21 @@ class PrimitiveVector[@specialized(Long, Int, Double) V: 
ClassTag](initialSize:
 
   /** Resizes the array, dropping elements if the total length decreases. */
   def resize(newLength: Int): PrimitiveVector[V] = {
-val newArray = new Array[V](newLength)
-_array.copyToArray(newArray)
-_array = newArray
+_array = copyArrayWithLength(newLength)
 if (newLength  _numElements) {
   _numElements = newLength
 }
 this
   }
+
+  /** Return a trimmed version of the underlying array. */
+  def toArray: Array[V] = {
+copyArrayWithLength(size)
+  }
+
+  private def copyArrayWithLength(length: Int): Array[V] = {
+val copy = new Array[V](length)
+_array.copyToArray(copy)
+copy
+  }
 }

http://git-wip-us.apache.org/repos/asf/spark/blob/e359794c/core/src/main/scala/org/apache/spark/util/collection/SizeTrackingVector.scala
--
diff --git 
a/core/src/main/scala/org/apache/spark/util/collection/SizeTrackingVector.scala 
b/core/src/main/scala/org/apache/spark/util/collection/SizeTrackingVector.scala
index 65a7b4e..dfcfb66 100644
--- 
a/core/src/main/scala/org/apache/spark/util/collection/SizeTrackingVector.scala
+++ 
b/core/src/main/scala/org/apache/spark/util/collection/SizeTrackingVector.scala
@@ -36,11 +36,4 @@ private[spark] class SizeTrackingVector[T: ClassTag]
 resetSamples()
 this
   }
-
-  /**
-   * Return a trimmed version of the underlying array.
-   */
-  def toArray: Array[T] = {
-super.iterator.toArray
-  }
 }


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



spark git commit: HOTFIX: Bump HBase version in MapR profiles.

2015-03-03 Thread pwendell
Repository: spark
Updated Branches:
  refs/heads/branch-1.3 841d2a27f - 1aa846165


HOTFIX: Bump HBase version in MapR profiles.

After #2982 (SPARK-4048) we rely on the newer HBase packaging format.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/1aa84616
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/1aa84616
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/1aa84616

Branch: refs/heads/branch-1.3
Commit: 1aa8461652ab02c3d5961dfb7b87d44f43d56093
Parents: 841d2a2
Author: Patrick Wendell patr...@databricks.com
Authored: Tue Mar 3 01:38:07 2015 -0800
Committer: Patrick Wendell patr...@databricks.com
Committed: Tue Mar 3 01:39:26 2015 -0800

--
 pom.xml | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/1aa84616/pom.xml
--
diff --git a/pom.xml b/pom.xml
index 34d4d42..ce1e544 100644
--- a/pom.xml
+++ b/pom.xml
@@ -1605,7 +1605,7 @@
   properties
 hadoop.version1.0.3-mapr-3.0.3/hadoop.version
 yarn.version2.4.1-mapr-1408/yarn.version
-hbase.version0.94.17-mapr-1405/hbase.version
+hbase.version0.98.4-mapr-1408/hbase.version
 zookeeper.version3.4.5-mapr-1406/zookeeper.version
   /properties
 /profile
@@ -1615,7 +1615,7 @@
   properties
 hadoop.version2.4.1-mapr-1408/hadoop.version
 yarn.version2.4.1-mapr-1408/yarn.version
-hbase.version0.94.17-mapr-1405-4.0.0-FCS/hbase.version
+hbase.version0.98.4-mapr-1408/hbase.version
 zookeeper.version3.4.5-mapr-1406/zookeeper.version
   /properties
   dependencies


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



spark git commit: HOTFIX: Bump HBase version in MapR profiles.

2015-03-03 Thread pwendell
Repository: spark
Updated Branches:
  refs/heads/master b19605619 - 165ff3642


HOTFIX: Bump HBase version in MapR profiles.

After #2982 (SPARK-4048) we rely on the newer HBase packaging format.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/165ff364
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/165ff364
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/165ff364

Branch: refs/heads/master
Commit: 165ff364265a5328f797e0901fe0dc02f30a7e7a
Parents: b196056
Author: Patrick Wendell patr...@databricks.com
Authored: Tue Mar 3 01:38:07 2015 -0800
Committer: Patrick Wendell patr...@databricks.com
Committed: Tue Mar 3 01:38:50 2015 -0800

--
 pom.xml | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/165ff364/pom.xml
--
diff --git a/pom.xml b/pom.xml
index bb355bf..77d08d6 100644
--- a/pom.xml
+++ b/pom.xml
@@ -1605,7 +1605,7 @@
   properties
 hadoop.version1.0.3-mapr-3.0.3/hadoop.version
 yarn.version2.4.1-mapr-1408/yarn.version
-hbase.version0.94.17-mapr-1405/hbase.version
+hbase.version0.98.4-mapr-1408/hbase.version
 zookeeper.version3.4.5-mapr-1406/zookeeper.version
   /properties
 /profile
@@ -1615,7 +1615,7 @@
   properties
 hadoop.version2.4.1-mapr-1408/hadoop.version
 yarn.version2.4.1-mapr-1408/yarn.version
-hbase.version0.94.17-mapr-1405-4.0.0-FCS/hbase.version
+hbase.version0.98.4-mapr-1408/hbase.version
 zookeeper.version3.4.5-mapr-1406/zookeeper.version
   /properties
   dependencies


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



[1/2] spark git commit: Revert Preparing development version 1.3.1-SNAPSHOT

2015-03-03 Thread pwendell
Repository: spark
Updated Branches:
  refs/heads/branch-1.3 ce7158cf7 - b012ed189


Revert Preparing development version 1.3.1-SNAPSHOT

This reverts commit 2ab0ba04f66683be25cbe0e83cecf2bdcb0f13ba.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/4fee08ef
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/4fee08ef
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/4fee08ef

Branch: refs/heads/branch-1.3
Commit: 4fee08ef0141b1be5684d78b6fe9cb93c98b0bc4
Parents: ce7158c
Author: Patrick Wendell patr...@databricks.com
Authored: Tue Mar 3 02:20:03 2015 -0800
Committer: Patrick Wendell patr...@databricks.com
Committed: Tue Mar 3 02:20:03 2015 -0800

--
 assembly/pom.xml  | 2 +-
 bagel/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 examples/pom.xml  | 2 +-
 external/flume-sink/pom.xml   | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka-assembly/pom.xml   | 2 +-
 external/kafka/pom.xml| 2 +-
 external/mqtt/pom.xml | 2 +-
 external/twitter/pom.xml  | 2 +-
 external/zeromq/pom.xml   | 2 +-
 extras/java8-tests/pom.xml| 2 +-
 extras/kinesis-asl/pom.xml| 2 +-
 extras/spark-ganglia-lgpl/pom.xml | 2 +-
 graphx/pom.xml| 2 +-
 mllib/pom.xml | 2 +-
 network/common/pom.xml| 2 +-
 network/shuffle/pom.xml   | 2 +-
 network/yarn/pom.xml  | 2 +-
 pom.xml   | 2 +-
 repl/pom.xml  | 2 +-
 sql/catalyst/pom.xml  | 2 +-
 sql/core/pom.xml  | 2 +-
 sql/hive-thriftserver/pom.xml | 2 +-
 sql/hive/pom.xml  | 2 +-
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 yarn/pom.xml  | 2 +-
 28 files changed, 28 insertions(+), 28 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/4fee08ef/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index 87b3e6f..0764d2d 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.3.1-SNAPSHOT/version
+version1.3.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/4fee08ef/bagel/pom.xml
--
diff --git a/bagel/pom.xml b/bagel/pom.xml
index 4a13c58..4f73cf7 100644
--- a/bagel/pom.xml
+++ b/bagel/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.3.1-SNAPSHOT/version
+version1.3.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/4fee08ef/core/pom.xml
--
diff --git a/core/pom.xml b/core/pom.xml
index aca0f58..5612149 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.3.1-SNAPSHOT/version
+version1.3.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/4fee08ef/examples/pom.xml
--
diff --git a/examples/pom.xml b/examples/pom.xml
index c424592..f7d6030 100644
--- a/examples/pom.xml
+++ b/examples/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.3.1-SNAPSHOT/version
+version1.3.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/4fee08ef/external/flume-sink/pom.xml
--
diff --git a/external/flume-sink/pom.xml b/external/flume-sink/pom.xml
index ef960a8..45aa775 100644
--- a/external/flume-sink/pom.xml
+++ b/external/flume-sink/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.3.1-SNAPSHOT/version
+version1.3.0/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/4fee08ef/external/flume/pom.xml
--
diff --git a/external/flume/pom.xml b/external/flume/pom.xml
index f01d6e8..455304f 100644
--- a/external/flume/pom.xml
+++ b/external/flume/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
- 

[2/2] spark git commit: Revert Preparing Spark release v1.3.0-rc1

2015-03-03 Thread pwendell
Revert Preparing Spark release v1.3.0-rc1

This reverts commit f97b0d4a6b26504916816d7aefcf3132cd1da6c2.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/b012ed18
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/b012ed18
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/b012ed18

Branch: refs/heads/branch-1.3
Commit: b012ed189844d2a515a882144364921caf89b4c0
Parents: 4fee08e
Author: Patrick Wendell patr...@databricks.com
Authored: Tue Mar 3 02:20:05 2015 -0800
Committer: Patrick Wendell patr...@databricks.com
Committed: Tue Mar 3 02:20:05 2015 -0800

--
 assembly/pom.xml  | 2 +-
 bagel/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 examples/pom.xml  | 2 +-
 external/flume-sink/pom.xml   | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka-assembly/pom.xml   | 2 +-
 external/kafka/pom.xml| 2 +-
 external/mqtt/pom.xml | 2 +-
 external/twitter/pom.xml  | 2 +-
 external/zeromq/pom.xml   | 2 +-
 extras/java8-tests/pom.xml| 2 +-
 extras/kinesis-asl/pom.xml| 2 +-
 extras/spark-ganglia-lgpl/pom.xml | 2 +-
 graphx/pom.xml| 2 +-
 mllib/pom.xml | 2 +-
 network/common/pom.xml| 2 +-
 network/shuffle/pom.xml   | 2 +-
 network/yarn/pom.xml  | 2 +-
 pom.xml   | 2 +-
 repl/pom.xml  | 2 +-
 sql/catalyst/pom.xml  | 2 +-
 sql/core/pom.xml  | 2 +-
 sql/hive-thriftserver/pom.xml | 2 +-
 sql/hive/pom.xml  | 2 +-
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 yarn/pom.xml  | 2 +-
 28 files changed, 28 insertions(+), 28 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/b012ed18/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index 0764d2d..c1bcdbb 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.3.0/version
+version1.3.0-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/b012ed18/bagel/pom.xml
--
diff --git a/bagel/pom.xml b/bagel/pom.xml
index 4f73cf7..510e926 100644
--- a/bagel/pom.xml
+++ b/bagel/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.3.0/version
+version1.3.0-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/b012ed18/core/pom.xml
--
diff --git a/core/pom.xml b/core/pom.xml
index 5612149..c993781 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.3.0/version
+version1.3.0-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/b012ed18/examples/pom.xml
--
diff --git a/examples/pom.xml b/examples/pom.xml
index f7d6030..8caad2b 100644
--- a/examples/pom.xml
+++ b/examples/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.3.0/version
+version1.3.0-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/b012ed18/external/flume-sink/pom.xml
--
diff --git a/external/flume-sink/pom.xml b/external/flume-sink/pom.xml
index 45aa775..0706f1e 100644
--- a/external/flume-sink/pom.xml
+++ b/external/flume-sink/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.3.0/version
+version1.3.0-SNAPSHOT/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/b012ed18/external/flume/pom.xml
--
diff --git a/external/flume/pom.xml b/external/flume/pom.xml
index 455304f..1f26813 100644
--- a/external/flume/pom.xml
+++ b/external/flume/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.3.0/version
+version1.3.0-SNAPSHOT/version