spark git commit: [SPARK-6285][SQL]Remove ParquetTestData in SparkBuild.scala and in README.md

2015-03-15 Thread lian
Repository: spark
Updated Branches:
  refs/heads/master c49d15662 - 62ede5383


[SPARK-6285][SQL]Remove ParquetTestData in SparkBuild.scala and in README.md

This is a following clean up PR for #5010
This will resolve issues when launching `hive/console` like below:
```
console:20: error: object ParquetTestData is not a member of package 
org.apache.spark.sql.parquet
   import org.apache.spark.sql.parquet.ParquetTestData
```

Author: OopsOutOfMemory victorshen...@126.com

Closes #5032 from OopsOutOfMemory/SPARK-6285 and squashes the following commits:

2996aeb [OopsOutOfMemory] remove ParquetTestData


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/62ede538
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/62ede538
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/62ede538

Branch: refs/heads/master
Commit: 62ede5383f64b69570a66d46939638f4bf38d1b1
Parents: c49d156
Author: OopsOutOfMemory victorshen...@126.com
Authored: Sun Mar 15 20:44:45 2015 +0800
Committer: Cheng Lian l...@databricks.com
Committed: Sun Mar 15 20:44:45 2015 +0800

--
 project/SparkBuild.scala | 6 ++
 sql/README.md| 1 -
 2 files changed, 2 insertions(+), 5 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/62ede538/project/SparkBuild.scala
--
diff --git a/project/SparkBuild.scala b/project/SparkBuild.scala
index f4c74c4..ac37c60 100644
--- a/project/SparkBuild.scala
+++ b/project/SparkBuild.scala
@@ -272,8 +272,7 @@ object SQL {
 |import org.apache.spark.sql.execution
 |import org.apache.spark.sql.functions._
 |import org.apache.spark.sql.test.TestSQLContext._
-|import org.apache.spark.sql.types._
-|import org.apache.spark.sql.parquet.ParquetTestData.stripMargin,
+|import org.apache.spark.sql.types._.stripMargin,
 cleanupCommands in console := sparkContext.stop()
   )
 }
@@ -304,8 +303,7 @@ object Hive {
 |import org.apache.spark.sql.functions._
 |import org.apache.spark.sql.hive._
 |import org.apache.spark.sql.hive.test.TestHive._
-|import org.apache.spark.sql.types._
-|import org.apache.spark.sql.parquet.ParquetTestData.stripMargin,
+|import org.apache.spark.sql.types._.stripMargin,
 cleanupCommands in console := sparkContext.stop(),
 // Some of our log4j jars make it impossible to submit jobs from this JVM 
to Hive Map/Reduce
 // in order to generate golden files.  This is only required for 
developers who are adding new

http://git-wip-us.apache.org/repos/asf/spark/blob/62ede538/sql/README.md
--
diff --git a/sql/README.md b/sql/README.md
index 48f8334..fbb3200 100644
--- a/sql/README.md
+++ b/sql/README.md
@@ -41,7 +41,6 @@ import org.apache.spark.sql.functions._
 import org.apache.spark.sql.hive._
 import org.apache.spark.sql.hive.test.TestHive._
 import org.apache.spark.sql.types._
-import org.apache.spark.sql.parquet.ParquetTestData
 Type in expressions to have them evaluated.
 Type :help for more information.
 


-
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org



spark git commit: [SPARK-3619] Part 2. Upgrade to Mesos 0.21 to work around MESOS-1688

2015-03-15 Thread srowen
Repository: spark
Updated Branches:
  refs/heads/branch-1.3 ad4756321 - 43fcab01a


[SPARK-3619] Part 2. Upgrade to Mesos 0.21 to work around MESOS-1688

- MESOS_NATIVE_LIBRARY become deprecated
- Chagned MESOS_NATIVE_LIBRARY to MESOS_NATIVE_JAVA_LIBRARY

Author: Jongyoul Lee jongy...@gmail.com

Closes #4361 from jongyoul/SPARK-3619-1 and squashes the following commits:

f1ea91f [Jongyoul Lee] Merge branch 'SPARK-3619-1' of 
https://github.com/jongyoul/spark into SPARK-3619-1
a6a00c2 [Jongyoul Lee] [SPARK-3619] Upgrade to Mesos 0.21 to work around 
MESOS-1688 - Removed 'Known issues' section
2e15a21 [Jongyoul Lee] [SPARK-3619] Upgrade to Mesos 0.21 to work around 
MESOS-1688 - MESOS_NATIVE_LIBRARY become deprecated - Chagned 
MESOS_NATIVE_LIBRARY to MESOS_NATIVE_JAVA_LIBRARY
0dace7b [Jongyoul Lee] [SPARK-3619] Upgrade to Mesos 0.21 to work around 
MESOS-1688 - MESOS_NATIVE_LIBRARY become deprecated - Chagned 
MESOS_NATIVE_LIBRARY to MESOS_NATIVE_JAVA_LIBRARY


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/43fcab01
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/43fcab01
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/43fcab01

Branch: refs/heads/branch-1.3
Commit: 43fcab01a4cb8e0534083da36a9fac022575f4f2
Parents: ad47563
Author: Jongyoul Lee jongy...@gmail.com
Authored: Sun Mar 15 15:46:55 2015 +
Committer: Sean Owen so...@cloudera.com
Committed: Sun Mar 15 15:49:01 2015 +

--
 conf/spark-env.sh.template  | 2 +-
 docs/running-on-mesos.md| 5 +
 .../src/test/scala/org/apache/spark/repl/ReplSuite.scala| 2 +-
 .../src/test/scala/org/apache/spark/repl/ReplSuite.scala| 2 +-
 4 files changed, 4 insertions(+), 7 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/43fcab01/conf/spark-env.sh.template
--
diff --git a/conf/spark-env.sh.template b/conf/spark-env.sh.template
index 0886b02..67f81d3 100755
--- a/conf/spark-env.sh.template
+++ b/conf/spark-env.sh.template
@@ -15,7 +15,7 @@
 # - SPARK_PUBLIC_DNS, to set the public DNS name of the driver program
 # - SPARK_CLASSPATH, default classpath entries to append
 # - SPARK_LOCAL_DIRS, storage directories to use on this node for shuffle and 
RDD data
-# - MESOS_NATIVE_LIBRARY, to point to your libmesos.so if you use Mesos
+# - MESOS_NATIVE_JAVA_LIBRARY, to point to your libmesos.so if you use Mesos
 
 # Options read in YARN client mode
 # - HADOOP_CONF_DIR, to point Spark towards Hadoop configuration files

http://git-wip-us.apache.org/repos/asf/spark/blob/43fcab01/docs/running-on-mesos.md
--
diff --git a/docs/running-on-mesos.md b/docs/running-on-mesos.md
index db1173a..bf39816 100644
--- a/docs/running-on-mesos.md
+++ b/docs/running-on-mesos.md
@@ -110,7 +110,7 @@ cluster, or `mesos://zk://host:2181` for a multi-master 
Mesos cluster using ZooK
 The driver also needs some configuration in `spark-env.sh` to interact 
properly with Mesos:
 
 1. In `spark-env.sh` set some environment variables:
- * `export MESOS_NATIVE_LIBRARY=path to libmesos.so`. This path is typically
+ * `export MESOS_NATIVE_JAVA_LIBRARY=path to libmesos.so`. This path is 
typically
`prefix/lib/libmesos.so` where the prefix is `/usr/local` by default. See 
Mesos installation
instructions above. On Mac OS X, the library is called `libmesos.dylib` 
instead of
`libmesos.so`.
@@ -167,9 +167,6 @@ acquire. By default, it will acquire *all* cores in the 
cluster (that get offere
 only makes sense if you run just one application at a time. You can cap the 
maximum number of cores
 using `conf.set(spark.cores.max, 10)` (for example).
 
-# Known issues
-- When using the fine-grained mode, make sure that your executors always 
leave 32 MB free on the slaves. Otherwise it can happen that your Spark job 
does not proceed anymore. Currently, Apache Mesos only offers resources if 
there are at least 32 MB memory allocatable. But as Spark allocates memory only 
for the executor and cpu only for tasks, it can happen on high slave memory 
usage that no new tasks will be started anymore. More details can be found in 
[MESOS-1688](https://issues.apache.org/jira/browse/MESOS-1688). Alternatively 
use the coarse-gained mode, which is not affected by this issue.
-
 # Running Alongside Hadoop
 
 You can run Spark and Mesos alongside your existing Hadoop cluster by just 
launching them as a

http://git-wip-us.apache.org/repos/asf/spark/blob/43fcab01/repl/scala-2.10/src/test/scala/org/apache/spark/repl/ReplSuite.scala
--
diff --git 

spark git commit: [SPARK-3619] Part 2. Upgrade to Mesos 0.21 to work around MESOS-1688

2015-03-15 Thread srowen
Repository: spark
Updated Branches:
  refs/heads/master 62ede5383 - aa6536fa3


[SPARK-3619] Part 2. Upgrade to Mesos 0.21 to work around MESOS-1688

- MESOS_NATIVE_LIBRARY become deprecated
- Chagned MESOS_NATIVE_LIBRARY to MESOS_NATIVE_JAVA_LIBRARY

Author: Jongyoul Lee jongy...@gmail.com

Closes #4361 from jongyoul/SPARK-3619-1 and squashes the following commits:

f1ea91f [Jongyoul Lee] Merge branch 'SPARK-3619-1' of 
https://github.com/jongyoul/spark into SPARK-3619-1
a6a00c2 [Jongyoul Lee] [SPARK-3619] Upgrade to Mesos 0.21 to work around 
MESOS-1688 - Removed 'Known issues' section
2e15a21 [Jongyoul Lee] [SPARK-3619] Upgrade to Mesos 0.21 to work around 
MESOS-1688 - MESOS_NATIVE_LIBRARY become deprecated - Chagned 
MESOS_NATIVE_LIBRARY to MESOS_NATIVE_JAVA_LIBRARY
0dace7b [Jongyoul Lee] [SPARK-3619] Upgrade to Mesos 0.21 to work around 
MESOS-1688 - MESOS_NATIVE_LIBRARY become deprecated - Chagned 
MESOS_NATIVE_LIBRARY to MESOS_NATIVE_JAVA_LIBRARY


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/aa6536fa
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/aa6536fa
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/aa6536fa

Branch: refs/heads/master
Commit: aa6536fa3c2ed1cac47abc79fc22e273f0814858
Parents: 62ede53
Author: Jongyoul Lee jongy...@gmail.com
Authored: Sun Mar 15 15:46:55 2015 +
Committer: Sean Owen so...@cloudera.com
Committed: Sun Mar 15 15:46:55 2015 +

--
 conf/spark-env.sh.template  | 2 +-
 docs/running-on-mesos.md| 5 +
 .../src/test/scala/org/apache/spark/repl/ReplSuite.scala| 2 +-
 .../src/test/scala/org/apache/spark/repl/ReplSuite.scala| 2 +-
 4 files changed, 4 insertions(+), 7 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/aa6536fa/conf/spark-env.sh.template
--
diff --git a/conf/spark-env.sh.template b/conf/spark-env.sh.template
index 0886b02..67f81d3 100755
--- a/conf/spark-env.sh.template
+++ b/conf/spark-env.sh.template
@@ -15,7 +15,7 @@
 # - SPARK_PUBLIC_DNS, to set the public DNS name of the driver program
 # - SPARK_CLASSPATH, default classpath entries to append
 # - SPARK_LOCAL_DIRS, storage directories to use on this node for shuffle and 
RDD data
-# - MESOS_NATIVE_LIBRARY, to point to your libmesos.so if you use Mesos
+# - MESOS_NATIVE_JAVA_LIBRARY, to point to your libmesos.so if you use Mesos
 
 # Options read in YARN client mode
 # - HADOOP_CONF_DIR, to point Spark towards Hadoop configuration files

http://git-wip-us.apache.org/repos/asf/spark/blob/aa6536fa/docs/running-on-mesos.md
--
diff --git a/docs/running-on-mesos.md b/docs/running-on-mesos.md
index e509e4b..59a3e9d 100644
--- a/docs/running-on-mesos.md
+++ b/docs/running-on-mesos.md
@@ -110,7 +110,7 @@ cluster, or `mesos://zk://host:2181` for a multi-master 
Mesos cluster using ZooK
 The driver also needs some configuration in `spark-env.sh` to interact 
properly with Mesos:
 
 1. In `spark-env.sh` set some environment variables:
- * `export MESOS_NATIVE_LIBRARY=path to libmesos.so`. This path is typically
+ * `export MESOS_NATIVE_JAVA_LIBRARY=path to libmesos.so`. This path is 
typically
`prefix/lib/libmesos.so` where the prefix is `/usr/local` by default. See 
Mesos installation
instructions above. On Mac OS X, the library is called `libmesos.dylib` 
instead of
`libmesos.so`.
@@ -167,9 +167,6 @@ acquire. By default, it will acquire *all* cores in the 
cluster (that get offere
 only makes sense if you run just one application at a time. You can cap the 
maximum number of cores
 using `conf.set(spark.cores.max, 10)` (for example).
 
-# Known issues
-- When using the fine-grained mode, make sure that your executors always 
leave 32 MB free on the slaves. Otherwise it can happen that your Spark job 
does not proceed anymore. Currently, Apache Mesos only offers resources if 
there are at least 32 MB memory allocatable. But as Spark allocates memory only 
for the executor and cpu only for tasks, it can happen on high slave memory 
usage that no new tasks will be started anymore. More details can be found in 
[MESOS-1688](https://issues.apache.org/jira/browse/MESOS-1688). Alternatively 
use the coarse-gained mode, which is not affected by this issue.
-
 # Running Alongside Hadoop
 
 You can run Spark and Mesos alongside your existing Hadoop cluster by just 
launching them as a

http://git-wip-us.apache.org/repos/asf/spark/blob/aa6536fa/repl/scala-2.10/src/test/scala/org/apache/spark/repl/ReplSuite.scala
--
diff --git 

svn commit: r1666875 - /spark/site/docs/1.3.0/sql-programming-guide.html

2015-03-15 Thread pwendell
Author: pwendell
Date: Mon Mar 16 05:30:06 2015
New Revision: 1666875

URL: http://svn.apache.org/r1666875
Log:
Updating to incorperate doc changes in SPARK-6275 and SPARK-5310

Modified:
spark/site/docs/1.3.0/sql-programming-guide.html

Modified: spark/site/docs/1.3.0/sql-programming-guide.html
URL: 
http://svn.apache.org/viewvc/spark/site/docs/1.3.0/sql-programming-guide.html?rev=1666875r1=1666874r2=1666875view=diff
==
--- spark/site/docs/1.3.0/sql-programming-guide.html (original)
+++ spark/site/docs/1.3.0/sql-programming-guide.html Mon Mar 16 05:30:06 2015
@@ -113,7 +113,7 @@
   ul id=markdown-toc
   lia href=#overviewOverview/a/li
   lia href=#dataframesDataFrames/aul
-  lia href=#starting-point-sqlcontextStarting Point: 
SQLContext/a/li
+  lia href=#starting-point-sqlcontextStarting Point: 
codeSQLContext/code/a/li
   lia href=#creating-dataframesCreating DataFrames/a/li
   lia href=#dataframe-operationsDataFrame Operations/a/li
   lia href=#running-sql-queries-programmaticallyRunning SQL Queries 
Programmatically/a/li
@@ -133,6 +133,8 @@
   /li
   lia href=#parquet-filesParquet Files/aul
   lia href=#loading-data-programmaticallyLoading Data 
Programmatically/a/li
+  lia href=#partition-discoveryPartition discovery/a/li
+  lia href=#schema-mergingSchema merging/a/li
   lia href=#configurationConfiguration/a/li
 /ul
   /li
@@ -158,7 +160,7 @@
   lia href=#unification-of-the-java-and-scala-apisUnification of 
the Java and Scala APIs/a/li
   lia 
href=#isolation-of-implicit-conversions-and-removal-of-dsl-package-scala-onlyIsolation
 of Implicit Conversions and Removal of dsl Package (Scala-only)/a/li
   lia 
href=#removal-of-the-type-aliases-in-orgapachesparksql-for-datatype-scala-onlyRemoval
 of the type aliases in org.apache.spark.sql for DataType (Scala-only)/a/li
-  lia 
href=#udf-registration-moved-to-sqlcontextudf-java--scalaUDF Registration 
Moved to sqlContext.udf (Java amp; Scala)/a/li
+  lia 
href=#udf-registration-moved-to-sqlcontextudf-java--scalaUDF Registration 
Moved to codesqlContext.udf/code (Java amp; Scala)/a/li
   lia href=#python-datatypes-no-longer-singletonsPython 
DataTypes No Longer Singletons/a/li
 /ul
   /li
@@ -191,14 +193,14 @@
 
 pAll of the examples on this page use sample data included in the Spark 
distribution and can be run in the codespark-shell/code or the 
codepyspark/code shell./p
 
-h2 id=starting-point-sqlcontextStarting Point: SQLContext/h2
+h2 id=starting-point-sqlcontextStarting Point: codeSQLContext/code/h2
 
 div class=codetabs
 div data-lang=scala
 
 pThe entry point into all functionality in Spark SQL is the
-a href=api/scala/index.html#org.apache.spark.sql.SQLContextSQLContext/a 
class, or one of its
-descendants.  To create a basic SQLContext, all you need is a SparkContext./p
+a 
href=api/scala/index.html#org.apache.spark.sql.`SQLContext`codeSQLContext/code/a
 class, or one of its
+descendants.  To create a basic codeSQLContext/code, all you need is a 
SparkContext./p
 
 div class=highlightprecode class=language-scala 
data-lang=scalaspan class=kval/span span class=nsc/spanspan 
class=k:/span span class=ktSparkContext/span span class=c1// An 
existing SparkContext./span
 span class=kval/span span class=nsqlContext/span span 
class=k=/span span class=knew/span span class=norg/spanspan 
class=o./spanspan class=napache/spanspan class=o./spanspan 
class=nspark/spanspan class=o./spanspan class=nsql/spanspan 
class=o./spanspan class=ncSQLContext/spanspan 
class=o(/spanspan class=nsc/spanspan class=o)/span
@@ -211,8 +213,8 @@ descendants.  To create a basic SQLConte
 div data-lang=java
 
 pThe entry point into all functionality in Spark SQL is the
-a href=api/java/index.html#org.apache.spark.sql.SQLContextSQLContext/a 
class, or one of its
-descendants.  To create a basic SQLContext, all you need is a SparkContext./p
+a 
href=api/java/index.html#org.apache.spark.sql.SQLContextcodeSQLContext/code/a
 class, or one of its
+descendants.  To create a basic codeSQLContext/code, all you need is a 
SparkContext./p
 
 div class=highlightprecode class=language-java 
data-lang=javaspan class=nJavaSparkContext/span span 
class=nsc/span span class=o=/span span class=o...;/span span 
class=c1// An existing JavaSparkContext./span
 span class=nSQLContext/span span class=nsqlContext/span span 
class=o=/span span class=knew/span span class=norg/spanspan 
class=o./spanspan class=naapache/spanspan class=o./spanspan 
class=naspark/spanspan class=o./spanspan class=nasql/spanspan 
class=o./spanspan class=naSQLContext/spanspan 
class=o(/spanspan class=nsc/spanspan 
class=o);/span/code/pre/div
@@ -222,8 +224,8 @@ descendants.  To create a basic SQLConte
 div data-lang=python
 
 pThe entry point into all relational functionality in