git commit: HOTFIX: Bump spark-ec2 version to 1.1.0

2014-08-29 Thread pwendell
Repository: spark
Updated Branches:
  refs/heads/branch-1.1 a4a7a2414 - c71b5c6db


HOTFIX: Bump spark-ec2 version to 1.1.0


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/c71b5c6d
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/c71b5c6d
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/c71b5c6d

Branch: refs/heads/branch-1.1
Commit: c71b5c6db151cfc63bfeabdc88034c3dd9dc9e60
Parents: a4a7a24
Author: Patrick Wendell pwend...@gmail.com
Authored: Fri Aug 29 11:20:45 2014 -0700
Committer: Patrick Wendell pwend...@gmail.com
Committed: Fri Aug 29 11:20:45 2014 -0700

--
 ec2/spark_ec2.py | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/c71b5c6d/ec2/spark_ec2.py
--
diff --git a/ec2/spark_ec2.py b/ec2/spark_ec2.py
index afef4ef..59356c5 100755
--- a/ec2/spark_ec2.py
+++ b/ec2/spark_ec2.py
@@ -82,7 +82,7 @@ def parse_args():
  between zones applies))
 parser.add_option(-a, --ami, help=Amazon Machine Image ID to use)
 parser.add_option(
--v, --spark-version, default=1.0.0,
+-v, --spark-version, default=1.1.0,
 help=Version of Spark to use: 'X.Y.Z' or a specific git hash)
 parser.add_option(
 --spark-git-repo,


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



git commit: [SPARK-3307] [PySpark] Fix doc string of SparkContext.broadcast()

2014-08-29 Thread joshrosen
Repository: spark
Updated Branches:
  refs/heads/master 27df6ce6a - e248328b3


[SPARK-3307] [PySpark] Fix doc string of SparkContext.broadcast()

 remove invalid docs

Author: Davies Liu davies@gmail.com

Closes #2202 from davies/keep and squashes the following commits:

aa3b44f [Davies Liu] remove invalid docs


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/e248328b
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/e248328b
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/e248328b

Branch: refs/heads/master
Commit: e248328b39f52073422a12fd0388208de41be1c7
Parents: 27df6ce
Author: Davies Liu davies@gmail.com
Authored: Fri Aug 29 11:47:49 2014 -0700
Committer: Josh Rosen joshro...@apache.org
Committed: Fri Aug 29 11:47:49 2014 -0700

--
 python/pyspark/context.py | 2 --
 1 file changed, 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/e248328b/python/pyspark/context.py
--
diff --git a/python/pyspark/context.py b/python/pyspark/context.py
index a90870e..82f76de 100644
--- a/python/pyspark/context.py
+++ b/python/pyspark/context.py
@@ -568,8 +568,6 @@ class SparkContext(object):
 L{Broadcastpyspark.broadcast.Broadcast}
 object for reading it in distributed functions. The variable will
 be sent to each cluster only once.
-
-:keep: Keep the `value` in driver or not.
 
 ser = CompressedSerializer(PickleSerializer())
 # pass large object by py4j is very slow and need much memory


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



git commit: [SPARK-3307] [PySpark] Fix doc string of SparkContext.broadcast()

2014-08-29 Thread joshrosen
Repository: spark
Updated Branches:
  refs/heads/branch-1.1 c71b5c6db - 98d0716a1


[SPARK-3307] [PySpark] Fix doc string of SparkContext.broadcast()

 remove invalid docs

Author: Davies Liu davies@gmail.com

Closes #2202 from davies/keep and squashes the following commits:

aa3b44f [Davies Liu] remove invalid docs

(cherry picked from commit e248328b39f52073422a12fd0388208de41be1c7)
Signed-off-by: Josh Rosen joshro...@apache.org


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/98d0716a
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/98d0716a
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/98d0716a

Branch: refs/heads/branch-1.1
Commit: 98d0716a11e8e5731adfa6031a5d35916f42b6f5
Parents: c71b5c6
Author: Davies Liu davies@gmail.com
Authored: Fri Aug 29 11:47:49 2014 -0700
Committer: Josh Rosen joshro...@apache.org
Committed: Fri Aug 29 11:48:00 2014 -0700

--
 python/pyspark/context.py | 2 --
 1 file changed, 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/98d0716a/python/pyspark/context.py
--
diff --git a/python/pyspark/context.py b/python/pyspark/context.py
index a90870e..82f76de 100644
--- a/python/pyspark/context.py
+++ b/python/pyspark/context.py
@@ -568,8 +568,6 @@ class SparkContext(object):
 L{Broadcastpyspark.broadcast.Broadcast}
 object for reading it in distributed functions. The variable will
 be sent to each cluster only once.
-
-:keep: Keep the `value` in driver or not.
 
 ser = CompressedSerializer(PickleSerializer())
 # pass large object by py4j is very slow and need much memory


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



git commit: [Docs] SQL doc formatting and typo fixes

2014-08-29 Thread marmbrus
Repository: spark
Updated Branches:
  refs/heads/branch-1.1 98d0716a1 - bfa2dc99a


[Docs] SQL doc formatting and typo fixes

As [reported on the dev 
list](http://apache-spark-developers-list.1001551.n3.nabble.com/VOTE-Release-Apache-Spark-1-1-0-RC2-tp8107p8131.html):
* Code fencing with triple-backticks doesn’t seem to work like it does on 
GitHub. Newlines are lost. Instead, use 4-space indent to format small code 
blocks.
* Nested bullets need 2 leading spaces, not 1.
* Spellcheck!

Author: Nicholas Chammas nicholas.cham...@gmail.com
Author: nchammas nicholas.cham...@gmail.com

Closes #2201 from nchammas/sql-doc-fixes and squashes the following commits:

873f889 [Nicholas Chammas] [Docs] fix skip-api flag
5195e0c [Nicholas Chammas] [Docs] SQL doc formatting and typo fixes
3b26c8d [nchammas] [Spark QA] Link to console output on test time out

(cherry picked from commit 53aa8316e88980c6f46d3b9fc90d935a4738a370)
Signed-off-by: Michael Armbrust mich...@databricks.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/bfa2dc99
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/bfa2dc99
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/bfa2dc99

Branch: refs/heads/branch-1.1
Commit: bfa2dc99a22c23dc4b10d1f9e5dd9681f6f48537
Parents: 98d0716
Author: Nicholas Chammas nicholas.cham...@gmail.com
Authored: Fri Aug 29 15:23:32 2014 -0700
Committer: Michael Armbrust mich...@databricks.com
Committed: Fri Aug 29 15:23:41 2014 -0700

--
 docs/README.md|   2 +-
 docs/sql-programming-guide.md | 109 +
 2 files changed, 52 insertions(+), 59 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/bfa2dc99/docs/README.md
--
diff --git a/docs/README.md b/docs/README.md
index fd7ba4e..0a0126c 100644
--- a/docs/README.md
+++ b/docs/README.md
@@ -30,7 +30,7 @@ called `_site` containing index.html as well as the rest of 
the compiled files.
 You can modify the default Jekyll build as follows:
 
 # Skip generating API docs (which takes a while)
-$ SKIP_SCALADOC=1 jekyll build
+$ SKIP_API=1 jekyll build
 # Serve content locally on port 4000
 $ jekyll serve --watch
 # Build the site with extra features used on the live page

http://git-wip-us.apache.org/repos/asf/spark/blob/bfa2dc99/docs/sql-programming-guide.md
--
diff --git a/docs/sql-programming-guide.md b/docs/sql-programming-guide.md
index c41f280..8f7fb54 100644
--- a/docs/sql-programming-guide.md
+++ b/docs/sql-programming-guide.md
@@ -474,10 +474,10 @@ anotherPeople = sqlContext.jsonRDD(anotherPeopleRDD)
 
 Spark SQL also supports reading and writing data stored in [Apache 
Hive](http://hive.apache.org/).
 However, since Hive has a large number of dependencies, it is not included in 
the default Spark assembly.
-In order to use Hive you must first run '`sbt/sbt -Phive assembly/assembly`' 
(or use `-Phive` for maven).
+In order to use Hive you must first run `sbt/sbt -Phive assembly/assembly` 
(or use `-Phive` for maven).
 This command builds a new assembly jar that includes Hive. Note that this Hive 
assembly jar must also be present
 on all of the worker nodes, as they will need access to the Hive serialization 
and deserialization libraries
-(SerDes) in order to acccess data stored in Hive.
+(SerDes) in order to access data stored in Hive.
 
 Configuration of Hive is done by placing your `hive-site.xml` file in `conf/`.
 
@@ -576,9 +576,8 @@ evaluated by the SQL execution engine.  A full list of the 
functions supported c
 
 ## Running the Thrift JDBC server
 
-The Thrift JDBC server implemented here corresponds to the [`HiveServer2`]
-(https://cwiki.apache.org/confluence/display/Hive/Setting+Up+HiveServer2) in 
Hive 0.12. You can test
-the JDBC server with the beeline script comes with either Spark or Hive 0.12.
+The Thrift JDBC server implemented here corresponds to the 
[`HiveServer2`](https://cwiki.apache.org/confluence/display/Hive/Setting+Up+HiveServer2)
+in Hive 0.12. You can test the JDBC server with the beeline script comes with 
either Spark or Hive 0.12.
 
 To start the JDBC server, run the following in the Spark directory:
 
@@ -597,7 +596,7 @@ Connect to the JDBC server in beeline with:
 
 Beeline will ask you for a username and password. In non-secure mode, simply 
enter the username on
 your machine and a blank password. For secure mode, please follow the 
instructions given in the
-[beeline 
documentation](https://cwiki.apache.org/confluence/display/Hive/HiveServer2+Clients)
+[beeline 
documentation](https://cwiki.apache.org/confluence/display/Hive/HiveServer2+Clients).
 
 Configuration of Hive is done by placing your 

git commit: SPARK-2813: [SQL] Implement SQRT() directly in Spark SQL

2014-08-29 Thread marmbrus
Repository: spark
Updated Branches:
  refs/heads/master 53aa8316e - 2f1519def


SPARK-2813:  [SQL] Implement SQRT() directly in Spark SQL

This PR adds a native implementation for SQL SQRT() and thus avoids delegating 
this function to Hive.

Author: William Benton wi...@redhat.com

Closes #1750 from willb/spark-2813 and squashes the following commits:

22c8a79 [William Benton] Fixed missed newline from rebase
d673861 [William Benton] Added string coercions for SQRT and associated test 
case
e125df4 [William Benton] Added ExpressionEvaluationSuite test cases for SQRT
7b84bcd [William Benton] SQL SQRT now properly returns NULL for NULL inputs
8256971 [William Benton] added SQRT test to SqlQuerySuite
504d2e5 [William Benton] Added native SQRT implementation


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/2f1519de
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/2f1519de
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/2f1519de

Branch: refs/heads/master
Commit: 2f1519defaba4f3c7d536669f909bfd9e13e4069
Parents: 53aa831
Author: William Benton wi...@redhat.com
Authored: Fri Aug 29 15:26:59 2014 -0700
Committer: Michael Armbrust mich...@databricks.com
Committed: Fri Aug 29 15:26:59 2014 -0700

--
 .../org/apache/spark/sql/catalyst/SqlParser.scala |  2 ++
 .../sql/catalyst/analysis/HiveTypeCoercion.scala  |  2 ++
 .../spark/sql/catalyst/expressions/arithmetic.scala   | 13 +
 .../expressions/ExpressionEvaluationSuite.scala   | 13 +
 .../scala/org/apache/spark/sql/SQLQuerySuite.scala| 14 ++
 .../main/scala/org/apache/spark/sql/hive/HiveQl.scala |  2 ++
 6 files changed, 46 insertions(+)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/2f1519de/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/SqlParser.scala
--
diff --git 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/SqlParser.scala 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/SqlParser.scala
index 2c73a80..4f166c0 100644
--- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/SqlParser.scala
+++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/SqlParser.scala
@@ -122,6 +122,7 @@ class SqlParser extends StandardTokenParsers with 
PackratParsers {
   protected val EXCEPT = Keyword(EXCEPT)
   protected val SUBSTR = Keyword(SUBSTR)
   protected val SUBSTRING = Keyword(SUBSTRING)
+  protected val SQRT = Keyword(SQRT)
 
   // Use reflection to find the reserved words defined in this class.
   protected val reservedWords =
@@ -323,6 +324,7 @@ class SqlParser extends StandardTokenParsers with 
PackratParsers {
 (SUBSTR | SUBSTRING) ~ ( ~ expression ~ , ~ expression ~ , ~ 
expression ~ ) ^^ {
   case s ~ , ~ p ~ , ~ l = Substring(s,p,l)
 } |
+SQRT ~ ( ~ expression ~ ) ^^ { case exp = Sqrt(exp) } |
 ident ~ ( ~ repsep(expression, ,) ~ ) ^^ {
   case udfName ~ _ ~ exprs = UnresolvedFunction(udfName, exprs)
 }

http://git-wip-us.apache.org/repos/asf/spark/blob/2f1519de/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/HiveTypeCoercion.scala
--
diff --git 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/HiveTypeCoercion.scala
 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/HiveTypeCoercion.scala
index 15eb598..ecfcd62 100644
--- 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/HiveTypeCoercion.scala
+++ 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/HiveTypeCoercion.scala
@@ -227,6 +227,8 @@ trait HiveTypeCoercion {
 Sum(Cast(e, DoubleType))
   case Average(e) if e.dataType == StringType =
 Average(Cast(e, DoubleType))
+  case Sqrt(e) if e.dataType == StringType =
+Sqrt(Cast(e, DoubleType))
 }
   }
 

http://git-wip-us.apache.org/repos/asf/spark/blob/2f1519de/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/arithmetic.scala
--
diff --git 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/arithmetic.scala
 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/arithmetic.scala
index aae86a3..56f0428 100644
--- 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/arithmetic.scala
+++ 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/arithmetic.scala
@@ -33,6 +33,19 @@ case class UnaryMinus(child: Expression) extends 
UnaryExpression {
   }
 }
 
+case class Sqrt(child: Expression) extends UnaryExpression {
+  type EvaluatedType = Any
+  
+  def dataType = child.dataType
+  

git commit: [SPARK-3234][Build] Fixed environment variables that rely on deprecated command line options in make-distribution.sh

2014-08-29 Thread pwendell
Repository: spark
Updated Branches:
  refs/heads/master 2f1519def - 287c0ac77


[SPARK-3234][Build] Fixed environment variables that rely on deprecated command 
line options in make-distribution.sh

Please refer to [SPARK-3234](https://issues.apache.org/jira/browse/SPARK-3234) 
for details.

Author: Cheng Lian lian.cs@gmail.com

Closes #2208 from liancheng/spark-3234 and squashes the following commits:

fb26de8 [Cheng Lian] Fixed SPARK-3234


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/287c0ac7
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/287c0ac7
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/287c0ac7

Branch: refs/heads/master
Commit: 287c0ac7722dd4bc51b921ccc6f0e3c1625b5ff4
Parents: 2f1519d
Author: Cheng Lian lian.cs@gmail.com
Authored: Fri Aug 29 15:29:43 2014 -0700
Committer: Patrick Wendell pwend...@gmail.com
Committed: Fri Aug 29 15:29:43 2014 -0700

--
 make-distribution.sh | 12 +++-
 1 file changed, 11 insertions(+), 1 deletion(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/287c0ac7/make-distribution.sh
--
diff --git a/make-distribution.sh b/make-distribution.sh
index f7a6a9d..ee13990 100755
--- a/make-distribution.sh
+++ b/make-distribution.sh
@@ -113,7 +113,17 @@ if ! which mvn /dev/null; then
 echo -e Download Maven from https://maven.apache.org/;
 exit -1;
 fi
+
 VERSION=$(mvn help:evaluate -Dexpression=project.version 2/dev/null | grep -v 
INFO | tail -n 1)
+SPARK_HADOOP_VERSION=$(mvn help:evaluate -Dexpression=hadoop.version $@ 
2/dev/null\
+| grep -v INFO\
+| tail -n 1)
+SPARK_HIVE=$(mvn help:evaluate -Dexpression=project.activeProfiles $@ 
2/dev/null\
+| grep -v INFO\
+| fgrep --count idhive/id;\
+# Reset exit status to 0, otherwise the script stops here if the last grep 
finds nothing\
+# because we use set -o pipefail
+echo -n)
 
 JAVA_CMD=$JAVA_HOME/bin/java
 JAVA_VERSION=$($JAVA_CMD -version 21)
@@ -175,7 +185,7 @@ cp $FWDIR/examples/target/scala*/spark-examples*.jar 
$DISTDIR/lib/
 mkdir -p $DISTDIR/examples/src/main
 cp -r $FWDIR/examples/src/main $DISTDIR/examples/src/
 
-if [ $SPARK_HIVE == true ]; then
+if [ $SPARK_HIVE == 1 ]; then
   cp $FWDIR/lib_managed/jars/datanucleus*.jar $DISTDIR/lib/
 fi
 


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



git commit: [SPARK-3198] [SQL] Remove the TreeNode.id

2014-08-29 Thread marmbrus
Repository: spark
Updated Branches:
  refs/heads/master 287c0ac77 - dc4d577c6


[SPARK-3198] [SQL] Remove the TreeNode.id

Thus id property of the TreeNode API does save time in a faster way to compare 
2 TreeNodes, it is kind of performance bottleneck during the expression object 
creation in a multi-threading env (because of the memory barrier).
Fortunately, the tree node comparison only happen once in master, so even we 
remove it, the entire performance will not be affected.

Author: Cheng Hao hao.ch...@intel.com

Closes #2155 from chenghao-intel/treenode and squashes the following commits:

7cf2cd2 [Cheng Hao] Remove the implicit keyword for TreeNodeRef and some other 
small issues
5873415 [Cheng Hao] Remove the TreeNode.id


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/dc4d577c
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/dc4d577c
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/dc4d577c

Branch: refs/heads/master
Commit: dc4d577c6549df58f42c0e22cac354554d169896
Parents: 287c0ac
Author: Cheng Hao hao.ch...@intel.com
Authored: Fri Aug 29 15:32:26 2014 -0700
Committer: Michael Armbrust mich...@databricks.com
Committed: Fri Aug 29 15:32:26 2014 -0700

--
 .../spark/sql/catalyst/planning/patterns.scala  | 11 +
 .../spark/sql/catalyst/plans/QueryPlan.scala| 12 +-
 .../spark/sql/catalyst/trees/TreeNode.scala | 24 ++--
 .../spark/sql/catalyst/trees/package.scala  | 11 +
 .../sql/catalyst/trees/TreeNodeSuite.scala  |  5 +++-
 .../sql/execution/GeneratedAggregate.scala  | 10 
 .../spark/sql/execution/debug/package.scala |  7 +++---
 .../apache/spark/sql/execution/pythonUdfs.scala |  2 +-
 8 files changed, 40 insertions(+), 42 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/dc4d577c/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/planning/patterns.scala
--
diff --git 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/planning/patterns.scala
 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/planning/patterns.scala
index 90923fe..f0fd9a8 100644
--- 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/planning/patterns.scala
+++ 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/planning/patterns.scala
@@ -19,8 +19,9 @@ package org.apache.spark.sql.catalyst.planning
 
 import scala.annotation.tailrec
 
-import org.apache.spark.sql.catalyst.expressions._
 import org.apache.spark.Logging
+import org.apache.spark.sql.catalyst.trees.TreeNodeRef
+import org.apache.spark.sql.catalyst.expressions._
 import org.apache.spark.sql.catalyst.plans._
 import org.apache.spark.sql.catalyst.plans.logical._
 
@@ -134,8 +135,8 @@ object PartialAggregation {
   // Only do partial aggregation if supported by all aggregate expressions.
   if (allAggregates.size == partialAggregates.size) {
 // Create a map of expressions to their partial evaluations for all 
aggregate expressions.
-val partialEvaluations: Map[Long, SplitEvaluation] =
-  partialAggregates.map(a = (a.id, a.asPartial)).toMap
+val partialEvaluations: Map[TreeNodeRef, SplitEvaluation] =
+  partialAggregates.map(a = (new TreeNodeRef(a), a.asPartial)).toMap
 
 // We need to pass all grouping expressions though so the grouping can 
happen a second
 // time. However some of them might be unnamed so we alias them 
allowing them to be
@@ -148,8 +149,8 @@ object PartialAggregation {
 // Replace aggregations with a new expression that computes the result 
from the already
 // computed partial evaluations and grouping values.
 val rewrittenAggregateExpressions = 
aggregateExpressions.map(_.transformUp {
-  case e: Expression if partialEvaluations.contains(e.id) =
-partialEvaluations(e.id).finalEvaluation
+  case e: Expression if partialEvaluations.contains(new 
TreeNodeRef(e)) =
+partialEvaluations(new TreeNodeRef(e)).finalEvaluation
   case e: Expression if namedGroupingExpressions.contains(e) =
 namedGroupingExpressions(e).toAttribute
 }).asInstanceOf[Seq[NamedExpression]]

http://git-wip-us.apache.org/repos/asf/spark/blob/dc4d577c/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/plans/QueryPlan.scala
--
diff --git 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/plans/QueryPlan.scala
 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/plans/QueryPlan.scala
index 1e177e2..af9e4d8 100644
--- 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/plans/QueryPlan.scala
+++ 

git commit: [SPARK-3269][SQL] Decreases initial buffer size for row set to prevent OOM

2014-08-29 Thread marmbrus
Repository: spark
Updated Branches:
  refs/heads/master b1eccfc88 - d94a44d7c


[SPARK-3269][SQL] Decreases initial buffer size for row set to prevent OOM

When a large batch size is specified, `SparkSQLOperationManager` OOMs even if 
the whole result set is much smaller than the batch size.

Author: Cheng Lian lian.cs@gmail.com

Closes #2171 from liancheng/jdbc-fetch-size and squashes the following commits:

5e1623b [Cheng Lian] Decreases initial buffer size for row set to prevent OOM


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/d94a44d7
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/d94a44d7
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/d94a44d7

Branch: refs/heads/master
Commit: d94a44d7caaf3fe7559d9ad7b10872fa16cf81ca
Parents: b1eccfc
Author: Cheng Lian lian.cs@gmail.com
Authored: Fri Aug 29 15:36:04 2014 -0700
Committer: Michael Armbrust mich...@databricks.com
Committed: Fri Aug 29 15:36:04 2014 -0700

--
 .../sql/hive/thriftserver/server/SparkSQLOperationManager.scala | 5 +++--
 1 file changed, 3 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/d94a44d7/sql/hive-thriftserver/src/main/scala/org/apache/spark/sql/hive/thriftserver/server/SparkSQLOperationManager.scala
--
diff --git 
a/sql/hive-thriftserver/src/main/scala/org/apache/spark/sql/hive/thriftserver/server/SparkSQLOperationManager.scala
 
b/sql/hive-thriftserver/src/main/scala/org/apache/spark/sql/hive/thriftserver/server/SparkSQLOperationManager.scala
index 6eccb1b..f12b5a6 100644
--- 
a/sql/hive-thriftserver/src/main/scala/org/apache/spark/sql/hive/thriftserver/server/SparkSQLOperationManager.scala
+++ 
b/sql/hive-thriftserver/src/main/scala/org/apache/spark/sql/hive/thriftserver/server/SparkSQLOperationManager.scala
@@ -66,9 +66,10 @@ class SparkSQLOperationManager(hiveContext: HiveContext) 
extends OperationManage
 if (!iter.hasNext) {
   new RowSet()
 } else {
-  val maxRows = maxRowsL.toInt // Do you really want a row batch 
larger than Int Max? No.
+  // maxRowsL here typically maps to java.sql.Statement.getFetchSize, 
which is an int
+  val maxRows = maxRowsL.toInt
   var curRow = 0
-  var rowSet = new ArrayBuffer[Row](maxRows)
+  var rowSet = new ArrayBuffer[Row](maxRows.min(1024))
 
   while (curRow  maxRows  iter.hasNext) {
 val sparkRow = iter.next()


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



git commit: [SPARK-3269][SQL] Decreases initial buffer size for row set to prevent OOM

2014-08-29 Thread marmbrus
Repository: spark
Updated Branches:
  refs/heads/branch-1.1 cf049efdc - 9bae345ab


[SPARK-3269][SQL] Decreases initial buffer size for row set to prevent OOM

When a large batch size is specified, `SparkSQLOperationManager` OOMs even if 
the whole result set is much smaller than the batch size.

Author: Cheng Lian lian.cs@gmail.com

Closes #2171 from liancheng/jdbc-fetch-size and squashes the following commits:

5e1623b [Cheng Lian] Decreases initial buffer size for row set to prevent OOM

(cherry picked from commit d94a44d7caaf3fe7559d9ad7b10872fa16cf81ca)
Signed-off-by: Michael Armbrust mich...@databricks.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/9bae345a
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/9bae345a
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/9bae345a

Branch: refs/heads/branch-1.1
Commit: 9bae345ab543b686e8b32e504711140950d9ea68
Parents: cf049ef
Author: Cheng Lian lian.cs@gmail.com
Authored: Fri Aug 29 15:36:04 2014 -0700
Committer: Michael Armbrust mich...@databricks.com
Committed: Fri Aug 29 15:36:19 2014 -0700

--
 .../sql/hive/thriftserver/server/SparkSQLOperationManager.scala | 5 +++--
 1 file changed, 3 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/9bae345a/sql/hive-thriftserver/src/main/scala/org/apache/spark/sql/hive/thriftserver/server/SparkSQLOperationManager.scala
--
diff --git 
a/sql/hive-thriftserver/src/main/scala/org/apache/spark/sql/hive/thriftserver/server/SparkSQLOperationManager.scala
 
b/sql/hive-thriftserver/src/main/scala/org/apache/spark/sql/hive/thriftserver/server/SparkSQLOperationManager.scala
index 6eccb1b..f12b5a6 100644
--- 
a/sql/hive-thriftserver/src/main/scala/org/apache/spark/sql/hive/thriftserver/server/SparkSQLOperationManager.scala
+++ 
b/sql/hive-thriftserver/src/main/scala/org/apache/spark/sql/hive/thriftserver/server/SparkSQLOperationManager.scala
@@ -66,9 +66,10 @@ class SparkSQLOperationManager(hiveContext: HiveContext) 
extends OperationManage
 if (!iter.hasNext) {
   new RowSet()
 } else {
-  val maxRows = maxRowsL.toInt // Do you really want a row batch 
larger than Int Max? No.
+  // maxRowsL here typically maps to java.sql.Statement.getFetchSize, 
which is an int
+  val maxRows = maxRowsL.toInt
   var curRow = 0
-  var rowSet = new ArrayBuffer[Row](maxRows)
+  var rowSet = new ArrayBuffer[Row](maxRows.min(1024))
 
   while (curRow  maxRows  iter.hasNext) {
 val sparkRow = iter.next()


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



git commit: [SPARK-3291][SQL]TestcaseName in createQueryTest should not contain :

2014-08-29 Thread marmbrus
Repository: spark
Updated Branches:
  refs/heads/branch-1.1 9bae345ab - c1333b852


[SPARK-3291][SQL]TestcaseName in createQueryTest should not contain :

: is not allowed to appear in a file name of Windows system. If file name 
contains :, this file can't be checked out in a Windows system and developers 
using Windows must be careful to not commit the deletion of such files, Which 
is very inconvenient.

Author: qiping.lqp qiping@alibaba-inc.com

Closes #2191 from chouqin/querytest and squashes the following commits:

0e943a1 [qiping.lqp] rename golden file
60a863f [qiping.lqp] TestcaseName in createQueryTest should not contain :

(cherry picked from commit 634d04b87c2744d645e9c26e746ba2006371d9b5)
Signed-off-by: Michael Armbrust mich...@databricks.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/c1333b85
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/c1333b85
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/c1333b85

Branch: refs/heads/branch-1.1
Commit: c1333b85292636c8db24ca08401bd559ed2c71f9
Parents: 9bae345
Author: qiping.lqp qiping@alibaba-inc.com
Authored: Fri Aug 29 15:37:43 2014 -0700
Committer: Michael Armbrust mich...@databricks.com
Committed: Fri Aug 29 15:38:00 2014 -0700

--
 ...uery Hive table-0-5d14d21a239daa42b086cc895215009a | 14 ++
 ...ity: Hive table-0-5d14d21a239daa42b086cc895215009a | 14 --
 .../spark/sql/hive/execution/HiveComparisonTest.scala |  3 +++
 .../spark/sql/hive/execution/HiveQuerySuite.scala |  2 +-
 4 files changed, 18 insertions(+), 15 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/c1333b85/sql/hive/src/test/resources/golden/case
 sensitivity when query Hive table-0-5d14d21a239daa42b086cc895215009a
--
diff --git a/sql/hive/src/test/resources/golden/case sensitivity when query 
Hive table-0-5d14d21a239daa42b086cc895215009a 
b/sql/hive/src/test/resources/golden/case sensitivity when query Hive 
table-0-5d14d21a239daa42b086cc895215009a
new file mode 100644
index 000..4d7127c
--- /dev/null
+++ b/sql/hive/src/test/resources/golden/case sensitivity when query Hive 
table-0-5d14d21a239daa42b086cc895215009a  
@@ -0,0 +1,14 @@
+0  val_0
+4  val_4
+12 val_12
+8  val_8
+0  val_0
+0  val_0
+10 val_10
+5  val_5
+11 val_11
+5  val_5
+2  val_2
+12 val_12
+5  val_5
+9  val_9

http://git-wip-us.apache.org/repos/asf/spark/blob/c1333b85/sql/hive/src/test/resources/golden/case
 sensitivity: Hive table-0-5d14d21a239daa42b086cc895215009a
--
diff --git a/sql/hive/src/test/resources/golden/case sensitivity: Hive 
table-0-5d14d21a239daa42b086cc895215009a 
b/sql/hive/src/test/resources/golden/case sensitivity: Hive 
table-0-5d14d21a239daa42b086cc895215009a
deleted file mode 100644
index 4d7127c..000
--- a/sql/hive/src/test/resources/golden/case sensitivity: Hive 
table-0-5d14d21a239daa42b086cc895215009a
+++ /dev/null
@@ -1,14 +0,0 @@
-0  val_0
-4  val_4
-12 val_12
-8  val_8
-0  val_0
-0  val_0
-10 val_10
-5  val_5
-11 val_11
-5  val_5
-2  val_2
-12 val_12
-5  val_5
-9  val_9

http://git-wip-us.apache.org/repos/asf/spark/blob/c1333b85/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveComparisonTest.scala
--
diff --git 
a/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveComparisonTest.scala
 
b/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveComparisonTest.scala
index 502ce8f..671c3b1 100644
--- 
a/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveComparisonTest.scala
+++ 
b/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveComparisonTest.scala
@@ -195,6 +195,9 @@ abstract class HiveComparisonTest
 
   val installHooksCommand = (?i)SET.*hooks.r
   def createQueryTest(testCaseName: String, sql: String, reset: Boolean = 
true) {
+// testCaseName must not contain ':', which is not allowed to appear in a 
filename of Windows
+assert(!testCaseName.contains(:))
+
 // If test sharding is enable, skip tests that are not in the correct 
shard.
 shardInfo.foreach {
   case (shardId, numShards) if testCaseName.hashCode % numShards != 
shardId = return

http://git-wip-us.apache.org/repos/asf/spark/blob/c1333b85/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveQuerySuite.scala
--
diff --git 
a/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveQuerySuite.scala
 

git commit: [SPARK-3173][SQL] Timestamp support in the parser

2014-08-29 Thread marmbrus
Repository: spark
Updated Branches:
  refs/heads/master 634d04b87 - 98ddbe6cd


 [SPARK-3173][SQL] Timestamp support in the parser

If you have a table with TIMESTAMP column, that column can't be used in WHERE 
clause properly - it is not evaluated properly. 
[More](https://issues.apache.org/jira/browse/SPARK-3173)

Motivation: http://www.aproint.com/aggregation-with-spark-sql/

- [x] modify SqlParser so it supports casting to TIMESTAMP (workaround for item 
2)
- [x] the string literal should be converted into Timestamp if the column is 
Timestamp.

Author: Zdenek Farana zdenek.far...@gmail.com
Author: Zdenek Farana zdenek.far...@aproint.com

Closes #2084 from byF/SPARK-3173 and squashes the following commits:

442b59d [Zdenek Farana] Fixed test merge conflict
2dbf4f6 [Zdenek Farana] Merge remote-tracking branch 'origin/SPARK-3173' into 
SPARK-3173
65b6215 [Zdenek Farana] Fixed timezone sensitivity in the test
47b27b4 [Zdenek Farana] Now works in the case of StringLiteral=TimestampColumn
96a661b [Zdenek Farana] Code style change
491dfcf [Zdenek Farana] Added test cases for SPARK-3173
4446b1e [Zdenek Farana] A string literal is casted into Timestamp when the 
column is Timestamp.
59af397 [Zdenek Farana] Added a new TIMESTAMP keyword; CAST to TIMESTAMP now 
can be used in SQL expression.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/98ddbe6c
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/98ddbe6c
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/98ddbe6c

Branch: refs/heads/master
Commit: 98ddbe6cdbe4141df3d008dcb675ecd682c97492
Parents: 634d04b
Author: Zdenek Farana zdenek.far...@gmail.com
Authored: Fri Aug 29 15:39:15 2014 -0700
Committer: Michael Armbrust mich...@databricks.com
Committed: Fri Aug 29 15:39:15 2014 -0700

--
 .../apache/spark/sql/catalyst/SqlParser.scala   |  3 +-
 .../catalyst/analysis/HiveTypeCoercion.scala| 10 +
 .../org/apache/spark/sql/SQLQuerySuite.scala| 43 +++-
 3 files changed, 54 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/98ddbe6c/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/SqlParser.scala
--
diff --git 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/SqlParser.scala 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/SqlParser.scala
index 4f166c0..a88bd85 100644
--- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/SqlParser.scala
+++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/SqlParser.scala
@@ -114,6 +114,7 @@ class SqlParser extends StandardTokenParsers with 
PackratParsers {
   protected val STRING = Keyword(STRING)
   protected val SUM = Keyword(SUM)
   protected val TABLE = Keyword(TABLE)
+  protected val TIMESTAMP = Keyword(TIMESTAMP)
   protected val TRUE = Keyword(TRUE)
   protected val UNCACHE = Keyword(UNCACHE)
   protected val UNION = Keyword(UNION)
@@ -359,7 +360,7 @@ class SqlParser extends StandardTokenParsers with 
PackratParsers {
 literal
 
   protected lazy val dataType: Parser[DataType] =
-STRING ^^^ StringType
+STRING ^^^ StringType | TIMESTAMP ^^^ TimestampType
 }
 
 class SqlLexical(val keywords: Seq[String]) extends StdLexical {

http://git-wip-us.apache.org/repos/asf/spark/blob/98ddbe6c/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/HiveTypeCoercion.scala
--
diff --git 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/HiveTypeCoercion.scala
 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/HiveTypeCoercion.scala
index ecfcd62..d6758eb 100644
--- 
a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/HiveTypeCoercion.scala
+++ 
b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/HiveTypeCoercion.scala
@@ -218,11 +218,21 @@ trait HiveTypeCoercion {
   case a: BinaryArithmetic if a.right.dataType == StringType =
 a.makeCopy(Array(a.left, Cast(a.right, DoubleType)))
 
+  case p: BinaryPredicate if p.left.dataType == StringType
+ p.right.dataType == TimestampType =
+p.makeCopy(Array(Cast(p.left, TimestampType), p.right))
+  case p: BinaryPredicate if p.left.dataType == TimestampType
+ p.right.dataType == StringType =
+p.makeCopy(Array(p.left, Cast(p.right, TimestampType)))
+
   case p: BinaryPredicate if p.left.dataType == StringType  
p.right.dataType != StringType =
 p.makeCopy(Array(Cast(p.left, DoubleType), p.right))
   case p: BinaryPredicate if p.left.dataType != StringType  
p.right.dataType == StringType =
 p.makeCopy(Array(p.left, Cast(p.right, DoubleType)))
 
+  case i @ In(a,b) if 

[1/2] git commit: Revert [maven-release-plugin] prepare for next development iteration

2014-08-29 Thread pwendell
Repository: spark
Updated Branches:
  refs/heads/branch-1.1 c1333b852 - c4b7ec8dd


Revert [maven-release-plugin] prepare for next development iteration

This reverts commit a4a7a241441489a0d31365e18476ae2e1c34464d.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/926f171c
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/926f171c
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/926f171c

Branch: refs/heads/branch-1.1
Commit: 926f171c90d9c3cb83e99c31bccf79c21bd6f5e0
Parents: c1333b8
Author: Patrick Wendell pwend...@gmail.com
Authored: Fri Aug 29 15:55:26 2014 -0700
Committer: Patrick Wendell pwend...@gmail.com
Committed: Fri Aug 29 15:55:26 2014 -0700

--
 assembly/pom.xml  | 2 +-
 bagel/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 examples/pom.xml  | 2 +-
 external/flume-sink/pom.xml   | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka/pom.xml| 2 +-
 external/mqtt/pom.xml | 2 +-
 external/twitter/pom.xml  | 2 +-
 external/zeromq/pom.xml   | 2 +-
 extras/kinesis-asl/pom.xml| 2 +-
 extras/spark-ganglia-lgpl/pom.xml | 2 +-
 graphx/pom.xml| 2 +-
 mllib/pom.xml | 2 +-
 pom.xml   | 4 ++--
 repl/pom.xml  | 2 +-
 sql/catalyst/pom.xml  | 2 +-
 sql/core/pom.xml  | 2 +-
 sql/hive-thriftserver/pom.xml | 2 +-
 sql/hive/pom.xml  | 2 +-
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 yarn/pom.xml  | 2 +-
 yarn/stable/pom.xml   | 2 +-
 24 files changed, 25 insertions(+), 25 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/926f171c/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index 4709b7d..799f8d9 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.1.1-SNAPSHOT/version
+version1.1.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/926f171c/bagel/pom.xml
--
diff --git a/bagel/pom.xml b/bagel/pom.xml
index f29540b..8eec7e5 100644
--- a/bagel/pom.xml
+++ b/bagel/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.1.1-SNAPSHOT/version
+version1.1.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/926f171c/core/pom.xml
--
diff --git a/core/pom.xml b/core/pom.xml
index debc4dd..83e6026 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.1.1-SNAPSHOT/version
+version1.1.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/926f171c/examples/pom.xml
--
diff --git a/examples/pom.xml b/examples/pom.xml
index f35d3d6..9bde90e 100644
--- a/examples/pom.xml
+++ b/examples/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.1.1-SNAPSHOT/version
+version1.1.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/926f171c/external/flume-sink/pom.xml
--
diff --git a/external/flume-sink/pom.xml b/external/flume-sink/pom.xml
index 17d0fe2..daaae05 100644
--- a/external/flume-sink/pom.xml
+++ b/external/flume-sink/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.1.1-SNAPSHOT/version
+version1.1.0/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/926f171c/external/flume/pom.xml
--
diff --git a/external/flume/pom.xml b/external/flume/pom.xml
index 402af35..830eb32 100644
--- a/external/flume/pom.xml
+++ b/external/flume/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.1.1-SNAPSHOT/version
+version1.1.0/version
 relativePath../../pom.xml/relativePath
   /parent
 


[2/2] git commit: Revert [maven-release-plugin] prepare release v1.1.0-rc2

2014-08-29 Thread pwendell
Revert [maven-release-plugin] prepare release v1.1.0-rc2

This reverts commit 711aebb329ca28046396af1e34395a0df92b5327.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/c4b7ec8d
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/c4b7ec8d
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/c4b7ec8d

Branch: refs/heads/branch-1.1
Commit: c4b7ec8dd8d63255451acf7e62383217f1f63bdc
Parents: 926f171
Author: Patrick Wendell pwend...@gmail.com
Authored: Fri Aug 29 15:55:30 2014 -0700
Committer: Patrick Wendell pwend...@gmail.com
Committed: Fri Aug 29 15:55:30 2014 -0700

--
 assembly/pom.xml  | 6 +++---
 bagel/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 examples/pom.xml  | 2 +-
 external/flume-sink/pom.xml   | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka/pom.xml| 2 +-
 external/mqtt/pom.xml | 2 +-
 external/twitter/pom.xml  | 2 +-
 external/zeromq/pom.xml   | 2 +-
 extras/kinesis-asl/pom.xml| 2 +-
 extras/spark-ganglia-lgpl/pom.xml | 2 +-
 graphx/pom.xml| 2 +-
 mllib/pom.xml | 2 +-
 pom.xml   | 9 +
 repl/pom.xml  | 2 +-
 sql/catalyst/pom.xml  | 5 +++--
 sql/core/pom.xml  | 5 +++--
 sql/hive-thriftserver/pom.xml | 5 +++--
 sql/hive/pom.xml  | 5 +++--
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 yarn/pom.xml  | 2 +-
 yarn/stable/pom.xml   | 2 +-
 24 files changed, 38 insertions(+), 33 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/c4b7ec8d/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index 799f8d9..9fbb037 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.1.0/version
+version1.1.0-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 
@@ -124,8 +124,8 @@
 transformer 
implementation=org.apache.maven.plugins.shade.resource.DontIncludeResourceTransformer
   resourcelog4j.properties/resource
 /transformer
-transformer 
implementation=org.apache.maven.plugins.shade.resource.ApacheLicenseResourceTransformer
 /
-transformer 
implementation=org.apache.maven.plugins.shade.resource.ApacheNoticeResourceTransformer
 /
+transformer 
implementation=org.apache.maven.plugins.shade.resource.ApacheLicenseResourceTransformer/
+transformer 
implementation=org.apache.maven.plugins.shade.resource.ApacheNoticeResourceTransformer/
   /transformers
 /configuration
   /execution

http://git-wip-us.apache.org/repos/asf/spark/blob/c4b7ec8d/bagel/pom.xml
--
diff --git a/bagel/pom.xml b/bagel/pom.xml
index 8eec7e5..bd51b11 100644
--- a/bagel/pom.xml
+++ b/bagel/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.1.0/version
+version1.1.0-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/c4b7ec8d/core/pom.xml
--
diff --git a/core/pom.xml b/core/pom.xml
index 83e6026..6d8be37 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.1.0/version
+version1.1.0-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/c4b7ec8d/examples/pom.xml
--
diff --git a/examples/pom.xml b/examples/pom.xml
index 9bde90e..8c4c128 100644
--- a/examples/pom.xml
+++ b/examples/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.1.0/version
+version1.1.0-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/c4b7ec8d/external/flume-sink/pom.xml
--
diff --git a/external/flume-sink/pom.xml b/external/flume-sink/pom.xml
index daaae05..b345276 100644
--- a/external/flume-sink/pom.xml
+++ b/external/flume-sink/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 

git commit: [SPARK-3296][mllib] spark-example should be run-example in head notation of DenseKMeans and SparseNaiveBayes

2014-08-29 Thread meng
Repository: spark
Updated Branches:
  refs/heads/master 98ddbe6cd - 13901764f


[SPARK-3296][mllib] spark-example should be run-example in head notation of 
DenseKMeans and SparseNaiveBayes

`./bin/spark-example`  should be `./bin/run-example` in DenseKMeans and 
SparseNaiveBayes

Author: wangfei wangfei_he...@126.com

Closes #2193 from scwf/run-example and squashes the following commits:

207eb3a [wangfei] spark-example should be run-example
27a8999 [wangfei] ./bin/spark-example should be ./bin/run-example


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/13901764
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/13901764
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/13901764

Branch: refs/heads/master
Commit: 13901764f4e9ed3de03e420d88ab42bdce5d5140
Parents: 98ddbe6
Author: wangfei wangfei_he...@126.com
Authored: Fri Aug 29 17:37:15 2014 -0700
Committer: Xiangrui Meng m...@databricks.com
Committed: Fri Aug 29 17:37:15 2014 -0700

--
 .../main/scala/org/apache/spark/examples/mllib/DenseKMeans.scala   | 2 +-
 .../scala/org/apache/spark/examples/mllib/SparseNaiveBayes.scala   | 2 +-
 2 files changed, 2 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/13901764/examples/src/main/scala/org/apache/spark/examples/mllib/DenseKMeans.scala
--
diff --git 
a/examples/src/main/scala/org/apache/spark/examples/mllib/DenseKMeans.scala 
b/examples/src/main/scala/org/apache/spark/examples/mllib/DenseKMeans.scala
index f96bc1b..89dfa26 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/DenseKMeans.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/DenseKMeans.scala
@@ -27,7 +27,7 @@ import org.apache.spark.mllib.linalg.Vectors
 /**
  * An example k-means app. Run with
  * {{{
- * ./bin/spark-example org.apache.spark.examples.mllib.DenseKMeans [options] 
input
+ * ./bin/run-example org.apache.spark.examples.mllib.DenseKMeans [options] 
input
  * }}}
  * If you use it as a template to create your own app, please use 
`spark-submit` to submit your app.
  */

http://git-wip-us.apache.org/repos/asf/spark/blob/13901764/examples/src/main/scala/org/apache/spark/examples/mllib/SparseNaiveBayes.scala
--
diff --git 
a/examples/src/main/scala/org/apache/spark/examples/mllib/SparseNaiveBayes.scala
 
b/examples/src/main/scala/org/apache/spark/examples/mllib/SparseNaiveBayes.scala
index 88acd9d..952fa2a 100644
--- 
a/examples/src/main/scala/org/apache/spark/examples/mllib/SparseNaiveBayes.scala
+++ 
b/examples/src/main/scala/org/apache/spark/examples/mllib/SparseNaiveBayes.scala
@@ -27,7 +27,7 @@ import org.apache.spark.mllib.util.MLUtils
 /**
  * An example naive Bayes app. Run with
  * {{{
- * ./bin/spark-example org.apache.spark.examples.mllib.SparseNaiveBayes 
[options] input
+ * ./bin/run-example org.apache.spark.examples.mllib.SparseNaiveBayes 
[options] input
  * }}}
  * If you use it as a template to create your own app, please use 
`spark-submit` to submit your app.
  */


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



git commit: [SPARK-3296][mllib] spark-example should be run-example in head notation of DenseKMeans and SparseNaiveBayes

2014-08-29 Thread meng
Repository: spark
Updated Branches:
  refs/heads/branch-1.1 c4b7ec8dd - b0facb590


[SPARK-3296][mllib] spark-example should be run-example in head notation of 
DenseKMeans and SparseNaiveBayes

`./bin/spark-example`  should be `./bin/run-example` in DenseKMeans and 
SparseNaiveBayes

Author: wangfei wangfei_he...@126.com

Closes #2193 from scwf/run-example and squashes the following commits:

207eb3a [wangfei] spark-example should be run-example
27a8999 [wangfei] ./bin/spark-example should be ./bin/run-example

(cherry picked from commit 13901764f4e9ed3de03e420d88ab42bdce5d5140)
Signed-off-by: Xiangrui Meng m...@databricks.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/b0facb59
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/b0facb59
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/b0facb59

Branch: refs/heads/branch-1.1
Commit: b0facb590eac032f82cea35982c3ed335f2ebbf4
Parents: c4b7ec8
Author: wangfei wangfei_he...@126.com
Authored: Fri Aug 29 17:37:15 2014 -0700
Committer: Xiangrui Meng m...@databricks.com
Committed: Fri Aug 29 17:37:36 2014 -0700

--
 .../main/scala/org/apache/spark/examples/mllib/DenseKMeans.scala   | 2 +-
 .../scala/org/apache/spark/examples/mllib/SparseNaiveBayes.scala   | 2 +-
 2 files changed, 2 insertions(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/b0facb59/examples/src/main/scala/org/apache/spark/examples/mllib/DenseKMeans.scala
--
diff --git 
a/examples/src/main/scala/org/apache/spark/examples/mllib/DenseKMeans.scala 
b/examples/src/main/scala/org/apache/spark/examples/mllib/DenseKMeans.scala
index f96bc1b..89dfa26 100644
--- a/examples/src/main/scala/org/apache/spark/examples/mllib/DenseKMeans.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/mllib/DenseKMeans.scala
@@ -27,7 +27,7 @@ import org.apache.spark.mllib.linalg.Vectors
 /**
  * An example k-means app. Run with
  * {{{
- * ./bin/spark-example org.apache.spark.examples.mllib.DenseKMeans [options] 
input
+ * ./bin/run-example org.apache.spark.examples.mllib.DenseKMeans [options] 
input
  * }}}
  * If you use it as a template to create your own app, please use 
`spark-submit` to submit your app.
  */

http://git-wip-us.apache.org/repos/asf/spark/blob/b0facb59/examples/src/main/scala/org/apache/spark/examples/mllib/SparseNaiveBayes.scala
--
diff --git 
a/examples/src/main/scala/org/apache/spark/examples/mllib/SparseNaiveBayes.scala
 
b/examples/src/main/scala/org/apache/spark/examples/mllib/SparseNaiveBayes.scala
index 88acd9d..952fa2a 100644
--- 
a/examples/src/main/scala/org/apache/spark/examples/mllib/SparseNaiveBayes.scala
+++ 
b/examples/src/main/scala/org/apache/spark/examples/mllib/SparseNaiveBayes.scala
@@ -27,7 +27,7 @@ import org.apache.spark.mllib.util.MLUtils
 /**
  * An example naive Bayes app. Run with
  * {{{
- * ./bin/spark-example org.apache.spark.examples.mllib.SparseNaiveBayes 
[options] input
+ * ./bin/run-example org.apache.spark.examples.mllib.SparseNaiveBayes 
[options] input
  * }}}
  * If you use it as a template to create your own app, please use 
`spark-submit` to submit your app.
  */


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



git commit: [SPARK-3320][SQL] Made batched in-memory column buffer building work for SchemaRDDs with empty partitions

2014-08-29 Thread marmbrus
Repository: spark
Updated Branches:
  refs/heads/master 13901764f - 32b18dd52


[SPARK-3320][SQL] Made batched in-memory column buffer building work for 
SchemaRDDs with empty partitions

Author: Cheng Lian lian.cs@gmail.com

Closes #2213 from liancheng/spark-3320 and squashes the following commits:

45a0139 [Cheng Lian] Fixed typo in InMemoryColumnarQuerySuite
f67067d [Cheng Lian] Fixed SPARK-3320


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/32b18dd5
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/32b18dd5
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/32b18dd5

Branch: refs/heads/master
Commit: 32b18dd52cf8920903819f23e406271ecd8ac6bb
Parents: 1390176
Author: Cheng Lian lian.cs@gmail.com
Authored: Fri Aug 29 18:16:47 2014 -0700
Committer: Michael Armbrust mich...@databricks.com
Committed: Fri Aug 29 18:16:47 2014 -0700

--
 .../columnar/InMemoryColumnarTableScan.scala| 49 
 .../scala/org/apache/spark/sql/TestData.scala   |  5 ++
 .../columnar/InMemoryColumnarQuerySuite.scala   | 19 ++--
 3 files changed, 39 insertions(+), 34 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/32b18dd5/sql/core/src/main/scala/org/apache/spark/sql/columnar/InMemoryColumnarTableScan.scala
--
diff --git 
a/sql/core/src/main/scala/org/apache/spark/sql/columnar/InMemoryColumnarTableScan.scala
 
b/sql/core/src/main/scala/org/apache/spark/sql/columnar/InMemoryColumnarTableScan.scala
index bc36bac..cb055cd 100644
--- 
a/sql/core/src/main/scala/org/apache/spark/sql/columnar/InMemoryColumnarTableScan.scala
+++ 
b/sql/core/src/main/scala/org/apache/spark/sql/columnar/InMemoryColumnarTableScan.scala
@@ -104,40 +104,29 @@ private[sql] case class InMemoryColumnarTableScan(
   override def execute() = {
 relation.cachedColumnBuffers.mapPartitions { iterator =
   // Find the ordinals of the requested columns.  If none are requested, 
use the first.
-  val requestedColumns =
-if (attributes.isEmpty) {
-  Seq(0)
-} else {
-  attributes.map(a = relation.output.indexWhere(_.exprId == a.exprId))
-}
-
-  new Iterator[Row] {
-private[this] var columnBuffers: Array[ByteBuffer] = null
-private[this] var columnAccessors: Seq[ColumnAccessor] = null
-nextBatch()
-
-private[this] val nextRow = new 
GenericMutableRow(columnAccessors.length)
-
-def nextBatch() = {
-  columnBuffers = iterator.next()
-  columnAccessors = 
requestedColumns.map(columnBuffers(_)).map(ColumnAccessor(_))
-}
+  val requestedColumns = if (attributes.isEmpty) {
+Seq(0)
+  } else {
+attributes.map(a = relation.output.indexWhere(_.exprId == a.exprId))
+  }
 
-override def next() = {
-  if (!columnAccessors.head.hasNext) {
-nextBatch()
-  }
+  iterator
+.map(batch = requestedColumns.map(batch(_)).map(ColumnAccessor(_)))
+.flatMap { columnAccessors =
+  val nextRow = new GenericMutableRow(columnAccessors.length)
+  new Iterator[Row] {
+override def next() = {
+  var i = 0
+  while (i  nextRow.length) {
+columnAccessors(i).extractTo(nextRow, i)
+i += 1
+  }
+  nextRow
+}
 
-  var i = 0
-  while (i  nextRow.length) {
-columnAccessors(i).extractTo(nextRow, i)
-i += 1
+override def hasNext = columnAccessors.head.hasNext
   }
-  nextRow
 }
-
-override def hasNext = columnAccessors.head.hasNext || iterator.hasNext
-  }
 }
   }
 }

http://git-wip-us.apache.org/repos/asf/spark/blob/32b18dd5/sql/core/src/test/scala/org/apache/spark/sql/TestData.scala
--
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/TestData.scala 
b/sql/core/src/test/scala/org/apache/spark/sql/TestData.scala
index c3ec82f..eb33a61 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/TestData.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/TestData.scala
@@ -151,4 +151,9 @@ object TestData {
 TimestampField(new Timestamp(i))
   })
   timestamps.registerTempTable(timestamps)
+
+  case class IntField(i: Int)
+  // An RDD with 4 elements and 8 partitions
+  val withEmptyParts = TestSQLContext.sparkContext.parallelize((1 to 
4).map(IntField), 8)
+  withEmptyParts.registerTempTable(withEmptyParts)
 }

http://git-wip-us.apache.org/repos/asf/spark/blob/32b18dd5/sql/core/src/test/scala/org/apache/spark/sql/columnar/InMemoryColumnarQuerySuite.scala

git commit: [SPARK-3320][SQL] Made batched in-memory column buffer building work for SchemaRDDs with empty partitions

2014-08-29 Thread marmbrus
Repository: spark
Updated Branches:
  refs/heads/branch-1.1 b0facb590 - aa9364a03


[SPARK-3320][SQL] Made batched in-memory column buffer building work for 
SchemaRDDs with empty partitions

Author: Cheng Lian lian.cs@gmail.com

Closes #2213 from liancheng/spark-3320 and squashes the following commits:

45a0139 [Cheng Lian] Fixed typo in InMemoryColumnarQuerySuite
f67067d [Cheng Lian] Fixed SPARK-3320

(cherry picked from commit 32b18dd52cf8920903819f23e406271ecd8ac6bb)
Signed-off-by: Michael Armbrust mich...@databricks.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/aa9364a0
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/aa9364a0
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/aa9364a0

Branch: refs/heads/branch-1.1
Commit: aa9364a03ddc793fd2c94981fb168ef8100a507c
Parents: b0facb5
Author: Cheng Lian lian.cs@gmail.com
Authored: Fri Aug 29 18:16:47 2014 -0700
Committer: Michael Armbrust mich...@databricks.com
Committed: Fri Aug 29 18:16:58 2014 -0700

--
 .../columnar/InMemoryColumnarTableScan.scala| 49 
 .../scala/org/apache/spark/sql/TestData.scala   |  5 ++
 .../columnar/InMemoryColumnarQuerySuite.scala   | 19 ++--
 3 files changed, 39 insertions(+), 34 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/aa9364a0/sql/core/src/main/scala/org/apache/spark/sql/columnar/InMemoryColumnarTableScan.scala
--
diff --git 
a/sql/core/src/main/scala/org/apache/spark/sql/columnar/InMemoryColumnarTableScan.scala
 
b/sql/core/src/main/scala/org/apache/spark/sql/columnar/InMemoryColumnarTableScan.scala
index bc36bac..cb055cd 100644
--- 
a/sql/core/src/main/scala/org/apache/spark/sql/columnar/InMemoryColumnarTableScan.scala
+++ 
b/sql/core/src/main/scala/org/apache/spark/sql/columnar/InMemoryColumnarTableScan.scala
@@ -104,40 +104,29 @@ private[sql] case class InMemoryColumnarTableScan(
   override def execute() = {
 relation.cachedColumnBuffers.mapPartitions { iterator =
   // Find the ordinals of the requested columns.  If none are requested, 
use the first.
-  val requestedColumns =
-if (attributes.isEmpty) {
-  Seq(0)
-} else {
-  attributes.map(a = relation.output.indexWhere(_.exprId == a.exprId))
-}
-
-  new Iterator[Row] {
-private[this] var columnBuffers: Array[ByteBuffer] = null
-private[this] var columnAccessors: Seq[ColumnAccessor] = null
-nextBatch()
-
-private[this] val nextRow = new 
GenericMutableRow(columnAccessors.length)
-
-def nextBatch() = {
-  columnBuffers = iterator.next()
-  columnAccessors = 
requestedColumns.map(columnBuffers(_)).map(ColumnAccessor(_))
-}
+  val requestedColumns = if (attributes.isEmpty) {
+Seq(0)
+  } else {
+attributes.map(a = relation.output.indexWhere(_.exprId == a.exprId))
+  }
 
-override def next() = {
-  if (!columnAccessors.head.hasNext) {
-nextBatch()
-  }
+  iterator
+.map(batch = requestedColumns.map(batch(_)).map(ColumnAccessor(_)))
+.flatMap { columnAccessors =
+  val nextRow = new GenericMutableRow(columnAccessors.length)
+  new Iterator[Row] {
+override def next() = {
+  var i = 0
+  while (i  nextRow.length) {
+columnAccessors(i).extractTo(nextRow, i)
+i += 1
+  }
+  nextRow
+}
 
-  var i = 0
-  while (i  nextRow.length) {
-columnAccessors(i).extractTo(nextRow, i)
-i += 1
+override def hasNext = columnAccessors.head.hasNext
   }
-  nextRow
 }
-
-override def hasNext = columnAccessors.head.hasNext || iterator.hasNext
-  }
 }
   }
 }

http://git-wip-us.apache.org/repos/asf/spark/blob/aa9364a0/sql/core/src/test/scala/org/apache/spark/sql/TestData.scala
--
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/TestData.scala 
b/sql/core/src/test/scala/org/apache/spark/sql/TestData.scala
index c3ec82f..eb33a61 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/TestData.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/TestData.scala
@@ -151,4 +151,9 @@ object TestData {
 TimestampField(new Timestamp(i))
   })
   timestamps.registerTempTable(timestamps)
+
+  case class IntField(i: Int)
+  // An RDD with 4 elements and 8 partitions
+  val withEmptyParts = TestSQLContext.sparkContext.parallelize((1 to 
4).map(IntField), 8)
+  withEmptyParts.registerTempTable(withEmptyParts)
 }


git commit: Adding new CHANGES.txt

2014-08-29 Thread pwendell
Repository: spark
Updated Branches:
  refs/heads/branch-1.1 aa9364a03 - 272b4a68d


Adding new CHANGES.txt


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/272b4a68
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/272b4a68
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/272b4a68

Branch: refs/heads/branch-1.1
Commit: 272b4a68d9b4ae2d74df7615122c587bfac2dddf
Parents: aa9364a
Author: Patrick Wendell pwend...@gmail.com
Authored: Fri Aug 29 18:49:51 2014 -0700
Committer: Patrick Wendell pwend...@gmail.com
Committed: Fri Aug 29 18:49:51 2014 -0700

--
 CHANGES.txt | 45 +
 1 file changed, 45 insertions(+)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/272b4a68/CHANGES.txt
--
diff --git a/CHANGES.txt b/CHANGES.txt
index b42721a..3693d59 100644
--- a/CHANGES.txt
+++ b/CHANGES.txt
@@ -3,6 +3,51 @@ Spark Change Log
 
 Release 1.1.0
 
+  [SPARK-3320][SQL] Made batched in-memory column buffer building work for 
SchemaRDDs with empty partitions
+  Cheng Lian lian.cs@gmail.com
+  2014-08-29 18:16:47 -0700
+  Commit: aa9364a, github.com/apache/spark/pull/2213
+
+  [SPARK-3296][mllib] spark-example should be run-example in head notation of 
DenseKMeans and SparseNaiveBayes
+  wangfei wangfei_he...@126.com
+  2014-08-29 17:37:15 -0700
+  Commit: b0facb5, github.com/apache/spark/pull/2193
+
+  [SPARK-3291][SQL]TestcaseName in createQueryTest should not contain :
+  qiping.lqp qiping@alibaba-inc.com
+  2014-08-29 15:37:43 -0700
+  Commit: c1333b8, github.com/apache/spark/pull/2191
+
+  [SPARK-3269][SQL] Decreases initial buffer size for row set to prevent OOM
+  Cheng Lian lian.cs@gmail.com
+  2014-08-29 15:36:04 -0700
+  Commit: 9bae345, github.com/apache/spark/pull/2171
+
+  [SPARK-3234][Build] Fixed environment variables that rely on deprecated 
command line options in make-distribution.sh
+  Cheng Lian lian.cs@gmail.com
+  2014-08-29 15:29:43 -0700
+  Commit: cf049ef, github.com/apache/spark/pull/2208
+
+  [Docs] SQL doc formatting and typo fixes
+  Nicholas Chammas nicholas.cham...@gmail.com, nchammas 
nicholas.cham...@gmail.com
+  2014-08-29 15:23:32 -0700
+  Commit: bfa2dc9, github.com/apache/spark/pull/2201
+
+  [SPARK-3307] [PySpark] Fix doc string of SparkContext.broadcast()
+  Davies Liu davies@gmail.com
+  2014-08-29 11:47:49 -0700
+  Commit: 98d0716, github.com/apache/spark/pull/2202
+
+  HOTFIX: Bump spark-ec2 version to 1.1.0
+  Patrick Wendell pwend...@gmail.com
+  2014-08-29 11:20:45 -0700
+  Commit: c71b5c6
+
+  Adding new CHANGES.txt
+  Patrick Wendell pwend...@gmail.com
+  2014-08-28 17:17:30 -0700
+  Commit: 7db87b3
+
   [SPARK-3277] Fix external spilling with LZ4 assertion error
   Andrew Or andrewo...@gmail.com, Patrick Wendell pwend...@gmail.com
   2014-08-28 17:05:21 -0700


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



git commit: [maven-release-plugin] prepare release v1.1.0-rc3

2014-08-29 Thread pwendell
Repository: spark
Updated Branches:
  refs/heads/branch-1.1 272b4a68d - 2b2e02265


[maven-release-plugin] prepare release v1.1.0-rc3


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/2b2e0226
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/2b2e0226
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/2b2e0226

Branch: refs/heads/branch-1.1
Commit: 2b2e02265f80e4c5172c1e498aa9ba2c6b91c6c9
Parents: 272b4a6
Author: Patrick Wendell pwend...@gmail.com
Authored: Sat Aug 30 02:26:03 2014 +
Committer: Patrick Wendell pwend...@gmail.com
Committed: Sat Aug 30 02:26:03 2014 +

--
 assembly/pom.xml  | 6 +++---
 bagel/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 examples/pom.xml  | 2 +-
 external/flume-sink/pom.xml   | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka/pom.xml| 2 +-
 external/mqtt/pom.xml | 2 +-
 external/twitter/pom.xml  | 2 +-
 external/zeromq/pom.xml   | 2 +-
 extras/kinesis-asl/pom.xml| 2 +-
 extras/spark-ganglia-lgpl/pom.xml | 2 +-
 graphx/pom.xml| 2 +-
 mllib/pom.xml | 2 +-
 pom.xml   | 9 -
 repl/pom.xml  | 2 +-
 sql/catalyst/pom.xml  | 5 ++---
 sql/core/pom.xml  | 5 ++---
 sql/hive-thriftserver/pom.xml | 5 ++---
 sql/hive/pom.xml  | 5 ++---
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 yarn/pom.xml  | 2 +-
 yarn/stable/pom.xml   | 2 +-
 24 files changed, 33 insertions(+), 38 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/2b2e0226/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index 9fbb037..799f8d9 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.1.0-SNAPSHOT/version
+version1.1.0/version
 relativePath../pom.xml/relativePath
   /parent
 
@@ -124,8 +124,8 @@
 transformer 
implementation=org.apache.maven.plugins.shade.resource.DontIncludeResourceTransformer
   resourcelog4j.properties/resource
 /transformer
-transformer 
implementation=org.apache.maven.plugins.shade.resource.ApacheLicenseResourceTransformer/
-transformer 
implementation=org.apache.maven.plugins.shade.resource.ApacheNoticeResourceTransformer/
+transformer 
implementation=org.apache.maven.plugins.shade.resource.ApacheLicenseResourceTransformer
 /
+transformer 
implementation=org.apache.maven.plugins.shade.resource.ApacheNoticeResourceTransformer
 /
   /transformers
 /configuration
   /execution

http://git-wip-us.apache.org/repos/asf/spark/blob/2b2e0226/bagel/pom.xml
--
diff --git a/bagel/pom.xml b/bagel/pom.xml
index bd51b11..8eec7e5 100644
--- a/bagel/pom.xml
+++ b/bagel/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.1.0-SNAPSHOT/version
+version1.1.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/2b2e0226/core/pom.xml
--
diff --git a/core/pom.xml b/core/pom.xml
index 6d8be37..83e6026 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.1.0-SNAPSHOT/version
+version1.1.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/2b2e0226/examples/pom.xml
--
diff --git a/examples/pom.xml b/examples/pom.xml
index 8c4c128..9bde90e 100644
--- a/examples/pom.xml
+++ b/examples/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.1.0-SNAPSHOT/version
+version1.1.0/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/2b2e0226/external/flume-sink/pom.xml
--
diff --git a/external/flume-sink/pom.xml b/external/flume-sink/pom.xml
index b345276..daaae05 100644
--- a/external/flume-sink/pom.xml
+++ b/external/flume-sink/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 

git commit: [maven-release-plugin] prepare for next development iteration

2014-08-29 Thread pwendell
Repository: spark
Updated Branches:
  refs/heads/branch-1.1 2b2e02265 - 8b5f0dbd8


[maven-release-plugin] prepare for next development iteration


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/8b5f0dbd
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/8b5f0dbd
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/8b5f0dbd

Branch: refs/heads/branch-1.1
Commit: 8b5f0dbd8d32a25a4e7ba3ebe1a4c3c6310aeb85
Parents: 2b2e022
Author: Patrick Wendell pwend...@gmail.com
Authored: Sat Aug 30 02:26:11 2014 +
Committer: Patrick Wendell pwend...@gmail.com
Committed: Sat Aug 30 02:26:11 2014 +

--
 assembly/pom.xml  | 2 +-
 bagel/pom.xml | 2 +-
 core/pom.xml  | 2 +-
 examples/pom.xml  | 2 +-
 external/flume-sink/pom.xml   | 2 +-
 external/flume/pom.xml| 2 +-
 external/kafka/pom.xml| 2 +-
 external/mqtt/pom.xml | 2 +-
 external/twitter/pom.xml  | 2 +-
 external/zeromq/pom.xml   | 2 +-
 extras/kinesis-asl/pom.xml| 2 +-
 extras/spark-ganglia-lgpl/pom.xml | 2 +-
 graphx/pom.xml| 2 +-
 mllib/pom.xml | 2 +-
 pom.xml   | 4 ++--
 repl/pom.xml  | 2 +-
 sql/catalyst/pom.xml  | 2 +-
 sql/core/pom.xml  | 2 +-
 sql/hive-thriftserver/pom.xml | 2 +-
 sql/hive/pom.xml  | 2 +-
 streaming/pom.xml | 2 +-
 tools/pom.xml | 2 +-
 yarn/pom.xml  | 2 +-
 yarn/stable/pom.xml   | 2 +-
 24 files changed, 25 insertions(+), 25 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/8b5f0dbd/assembly/pom.xml
--
diff --git a/assembly/pom.xml b/assembly/pom.xml
index 799f8d9..4709b7d 100644
--- a/assembly/pom.xml
+++ b/assembly/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.1.0/version
+version1.1.1-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/8b5f0dbd/bagel/pom.xml
--
diff --git a/bagel/pom.xml b/bagel/pom.xml
index 8eec7e5..f29540b 100644
--- a/bagel/pom.xml
+++ b/bagel/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.1.0/version
+version1.1.1-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/8b5f0dbd/core/pom.xml
--
diff --git a/core/pom.xml b/core/pom.xml
index 83e6026..debc4dd 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.1.0/version
+version1.1.1-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/8b5f0dbd/examples/pom.xml
--
diff --git a/examples/pom.xml b/examples/pom.xml
index 9bde90e..f35d3d6 100644
--- a/examples/pom.xml
+++ b/examples/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.1.0/version
+version1.1.1-SNAPSHOT/version
 relativePath../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/8b5f0dbd/external/flume-sink/pom.xml
--
diff --git a/external/flume-sink/pom.xml b/external/flume-sink/pom.xml
index daaae05..17d0fe2 100644
--- a/external/flume-sink/pom.xml
+++ b/external/flume-sink/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.1.0/version
+version1.1.1-SNAPSHOT/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/8b5f0dbd/external/flume/pom.xml
--
diff --git a/external/flume/pom.xml b/external/flume/pom.xml
index 830eb32..402af35 100644
--- a/external/flume/pom.xml
+++ b/external/flume/pom.xml
@@ -21,7 +21,7 @@
   parent
 groupIdorg.apache.spark/groupId
 artifactIdspark-parent/artifactId
-version1.1.0/version
+version1.1.1-SNAPSHOT/version
 relativePath../../pom.xml/relativePath
   /parent
 

http://git-wip-us.apache.org/repos/asf/spark/blob/8b5f0dbd/external/kafka/pom.xml

Git Push Summary

2014-08-29 Thread pwendell
Repository: spark
Updated Tags:  refs/tags/v1.1.0-rc3 [created] 25596d1da

-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



git commit: BUILD: Adding back CDH4 as per user requests

2014-08-29 Thread pwendell
Repository: spark
Updated Branches:
  refs/heads/master 32b18dd52 - a004a8d87


BUILD: Adding back CDH4 as per user requests


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/a004a8d8
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/a004a8d8
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/a004a8d8

Branch: refs/heads/master
Commit: a004a8d879a85af3be0aefa3f331116d4aabb1e4
Parents: 32b18dd
Author: Patrick Wendell pwend...@gmail.com
Authored: Fri Aug 29 22:24:35 2014 -0700
Committer: Patrick Wendell pwend...@gmail.com
Committed: Fri Aug 29 22:24:35 2014 -0700

--
 dev/create-release/create-release.sh | 1 +
 1 file changed, 1 insertion(+)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/a004a8d8/dev/create-release/create-release.sh
--
diff --git a/dev/create-release/create-release.sh 
b/dev/create-release/create-release.sh
index 7549fbb..281e8d4 100755
--- a/dev/create-release/create-release.sh
+++ b/dev/create-release/create-release.sh
@@ -118,6 +118,7 @@ make_binary_release() {
 }
 
 make_binary_release hadoop1 -Phive -Dhadoop.version=1.0.4 
+make_binary_release cdh4 -Phive -Dhadoop.version=2.0.0-mr1-cdh4.2.0 
 make_binary_release hadoop2.3 -Phadoop-2.3 -Phive -Pyarn 
 make_binary_release hadoop2.4 -Phadoop-2.4 -Phive -Pyarn 
 make_binary_release hadoop2.4-without-hive -Phadoop-2.4 -Pyarn 


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



git commit: BUILD: Adding back CDH4 as per user requests

2014-08-29 Thread pwendell
Repository: spark
Updated Branches:
  refs/heads/branch-1.1 8b5f0dbd8 - d4ce26426


BUILD: Adding back CDH4 as per user requests


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/d4ce2642
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/d4ce2642
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/d4ce2642

Branch: refs/heads/branch-1.1
Commit: d4ce26426832eb09a229c51cb5e64048272cf4b6
Parents: 8b5f0db
Author: Patrick Wendell pwend...@gmail.com
Authored: Fri Aug 29 22:24:35 2014 -0700
Committer: Patrick Wendell pwend...@gmail.com
Committed: Fri Aug 29 22:26:04 2014 -0700

--
 dev/create-release/create-release.sh | 1 +
 1 file changed, 1 insertion(+)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/d4ce2642/dev/create-release/create-release.sh
--
diff --git a/dev/create-release/create-release.sh 
b/dev/create-release/create-release.sh
index 7549fbb..281e8d4 100755
--- a/dev/create-release/create-release.sh
+++ b/dev/create-release/create-release.sh
@@ -118,6 +118,7 @@ make_binary_release() {
 }
 
 make_binary_release hadoop1 -Phive -Dhadoop.version=1.0.4 
+make_binary_release cdh4 -Phive -Dhadoop.version=2.0.0-mr1-cdh4.2.0 
 make_binary_release hadoop2.3 -Phadoop-2.3 -Phive -Pyarn 
 make_binary_release hadoop2.4 -Phadoop-2.4 -Phive -Pyarn 
 make_binary_release hadoop2.4-without-hive -Phadoop-2.4 -Pyarn 


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



git commit: [SPARK-3305] Remove unused import from UI classes.

2014-08-29 Thread rxin
Repository: spark
Updated Branches:
  refs/heads/master a004a8d87 - 7e662af33


[SPARK-3305] Remove unused import from UI classes.

Author: Kousuke Saruta saru...@oss.nttdata.co.jp

Closes #2200 from sarutak/SPARK-3305 and squashes the following commits:

3cbd6ee [Kousuke Saruta] Removed unused import from classes related to UI


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/7e662af3
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/7e662af3
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/7e662af3

Branch: refs/heads/master
Commit: 7e662af332beb171dc89027a2562d0949d69cfa0
Parents: a004a8d
Author: Kousuke Saruta saru...@oss.nttdata.co.jp
Authored: Fri Aug 29 22:52:32 2014 -0700
Committer: Reynold Xin r...@apache.org
Committed: Fri Aug 29 22:52:32 2014 -0700

--
 core/src/main/scala/org/apache/spark/ui/exec/ExecutorsPage.scala   | 1 -
 .../main/scala/org/apache/spark/ui/jobs/JobProgressListener.scala  | 2 +-
 2 files changed, 1 insertion(+), 2 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/7e662af3/core/src/main/scala/org/apache/spark/ui/exec/ExecutorsPage.scala
--
diff --git a/core/src/main/scala/org/apache/spark/ui/exec/ExecutorsPage.scala 
b/core/src/main/scala/org/apache/spark/ui/exec/ExecutorsPage.scala
index 02df4e8..b0e3bb3 100644
--- a/core/src/main/scala/org/apache/spark/ui/exec/ExecutorsPage.scala
+++ b/core/src/main/scala/org/apache/spark/ui/exec/ExecutorsPage.scala
@@ -21,7 +21,6 @@ import javax.servlet.http.HttpServletRequest
 
 import scala.xml.Node
 
-import org.apache.spark.storage.StorageLevel
 import org.apache.spark.ui.{ToolTips, UIUtils, WebUIPage}
 import org.apache.spark.util.Utils
 

http://git-wip-us.apache.org/repos/asf/spark/blob/7e662af3/core/src/main/scala/org/apache/spark/ui/jobs/JobProgressListener.scala
--
diff --git 
a/core/src/main/scala/org/apache/spark/ui/jobs/JobProgressListener.scala 
b/core/src/main/scala/org/apache/spark/ui/jobs/JobProgressListener.scala
index f7f918f..eaeb861 100644
--- a/core/src/main/scala/org/apache/spark/ui/jobs/JobProgressListener.scala
+++ b/core/src/main/scala/org/apache/spark/ui/jobs/JobProgressListener.scala
@@ -17,7 +17,7 @@
 
 package org.apache.spark.ui.jobs
 
-import scala.collection.mutable.{HashMap, ListBuffer, Map}
+import scala.collection.mutable.{HashMap, ListBuffer}
 
 import org.apache.spark._
 import org.apache.spark.annotation.DeveloperApi


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org