spark git commit: [SPARK-6285][SQL]Remove ParquetTestData in SparkBuild.scala and in README.md
Repository: spark Updated Branches: refs/heads/master c49d15662 - 62ede5383 [SPARK-6285][SQL]Remove ParquetTestData in SparkBuild.scala and in README.md This is a following clean up PR for #5010 This will resolve issues when launching `hive/console` like below: ``` console:20: error: object ParquetTestData is not a member of package org.apache.spark.sql.parquet import org.apache.spark.sql.parquet.ParquetTestData ``` Author: OopsOutOfMemory victorshen...@126.com Closes #5032 from OopsOutOfMemory/SPARK-6285 and squashes the following commits: 2996aeb [OopsOutOfMemory] remove ParquetTestData Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/62ede538 Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/62ede538 Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/62ede538 Branch: refs/heads/master Commit: 62ede5383f64b69570a66d46939638f4bf38d1b1 Parents: c49d156 Author: OopsOutOfMemory victorshen...@126.com Authored: Sun Mar 15 20:44:45 2015 +0800 Committer: Cheng Lian l...@databricks.com Committed: Sun Mar 15 20:44:45 2015 +0800 -- project/SparkBuild.scala | 6 ++ sql/README.md| 1 - 2 files changed, 2 insertions(+), 5 deletions(-) -- http://git-wip-us.apache.org/repos/asf/spark/blob/62ede538/project/SparkBuild.scala -- diff --git a/project/SparkBuild.scala b/project/SparkBuild.scala index f4c74c4..ac37c60 100644 --- a/project/SparkBuild.scala +++ b/project/SparkBuild.scala @@ -272,8 +272,7 @@ object SQL { |import org.apache.spark.sql.execution |import org.apache.spark.sql.functions._ |import org.apache.spark.sql.test.TestSQLContext._ -|import org.apache.spark.sql.types._ -|import org.apache.spark.sql.parquet.ParquetTestData.stripMargin, +|import org.apache.spark.sql.types._.stripMargin, cleanupCommands in console := sparkContext.stop() ) } @@ -304,8 +303,7 @@ object Hive { |import org.apache.spark.sql.functions._ |import org.apache.spark.sql.hive._ |import org.apache.spark.sql.hive.test.TestHive._ -|import org.apache.spark.sql.types._ -|import org.apache.spark.sql.parquet.ParquetTestData.stripMargin, +|import org.apache.spark.sql.types._.stripMargin, cleanupCommands in console := sparkContext.stop(), // Some of our log4j jars make it impossible to submit jobs from this JVM to Hive Map/Reduce // in order to generate golden files. This is only required for developers who are adding new http://git-wip-us.apache.org/repos/asf/spark/blob/62ede538/sql/README.md -- diff --git a/sql/README.md b/sql/README.md index 48f8334..fbb3200 100644 --- a/sql/README.md +++ b/sql/README.md @@ -41,7 +41,6 @@ import org.apache.spark.sql.functions._ import org.apache.spark.sql.hive._ import org.apache.spark.sql.hive.test.TestHive._ import org.apache.spark.sql.types._ -import org.apache.spark.sql.parquet.ParquetTestData Type in expressions to have them evaluated. Type :help for more information. - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
spark git commit: [SPARK-3619] Part 2. Upgrade to Mesos 0.21 to work around MESOS-1688
Repository: spark Updated Branches: refs/heads/branch-1.3 ad4756321 - 43fcab01a [SPARK-3619] Part 2. Upgrade to Mesos 0.21 to work around MESOS-1688 - MESOS_NATIVE_LIBRARY become deprecated - Chagned MESOS_NATIVE_LIBRARY to MESOS_NATIVE_JAVA_LIBRARY Author: Jongyoul Lee jongy...@gmail.com Closes #4361 from jongyoul/SPARK-3619-1 and squashes the following commits: f1ea91f [Jongyoul Lee] Merge branch 'SPARK-3619-1' of https://github.com/jongyoul/spark into SPARK-3619-1 a6a00c2 [Jongyoul Lee] [SPARK-3619] Upgrade to Mesos 0.21 to work around MESOS-1688 - Removed 'Known issues' section 2e15a21 [Jongyoul Lee] [SPARK-3619] Upgrade to Mesos 0.21 to work around MESOS-1688 - MESOS_NATIVE_LIBRARY become deprecated - Chagned MESOS_NATIVE_LIBRARY to MESOS_NATIVE_JAVA_LIBRARY 0dace7b [Jongyoul Lee] [SPARK-3619] Upgrade to Mesos 0.21 to work around MESOS-1688 - MESOS_NATIVE_LIBRARY become deprecated - Chagned MESOS_NATIVE_LIBRARY to MESOS_NATIVE_JAVA_LIBRARY Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/43fcab01 Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/43fcab01 Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/43fcab01 Branch: refs/heads/branch-1.3 Commit: 43fcab01a4cb8e0534083da36a9fac022575f4f2 Parents: ad47563 Author: Jongyoul Lee jongy...@gmail.com Authored: Sun Mar 15 15:46:55 2015 + Committer: Sean Owen so...@cloudera.com Committed: Sun Mar 15 15:49:01 2015 + -- conf/spark-env.sh.template | 2 +- docs/running-on-mesos.md| 5 + .../src/test/scala/org/apache/spark/repl/ReplSuite.scala| 2 +- .../src/test/scala/org/apache/spark/repl/ReplSuite.scala| 2 +- 4 files changed, 4 insertions(+), 7 deletions(-) -- http://git-wip-us.apache.org/repos/asf/spark/blob/43fcab01/conf/spark-env.sh.template -- diff --git a/conf/spark-env.sh.template b/conf/spark-env.sh.template index 0886b02..67f81d3 100755 --- a/conf/spark-env.sh.template +++ b/conf/spark-env.sh.template @@ -15,7 +15,7 @@ # - SPARK_PUBLIC_DNS, to set the public DNS name of the driver program # - SPARK_CLASSPATH, default classpath entries to append # - SPARK_LOCAL_DIRS, storage directories to use on this node for shuffle and RDD data -# - MESOS_NATIVE_LIBRARY, to point to your libmesos.so if you use Mesos +# - MESOS_NATIVE_JAVA_LIBRARY, to point to your libmesos.so if you use Mesos # Options read in YARN client mode # - HADOOP_CONF_DIR, to point Spark towards Hadoop configuration files http://git-wip-us.apache.org/repos/asf/spark/blob/43fcab01/docs/running-on-mesos.md -- diff --git a/docs/running-on-mesos.md b/docs/running-on-mesos.md index db1173a..bf39816 100644 --- a/docs/running-on-mesos.md +++ b/docs/running-on-mesos.md @@ -110,7 +110,7 @@ cluster, or `mesos://zk://host:2181` for a multi-master Mesos cluster using ZooK The driver also needs some configuration in `spark-env.sh` to interact properly with Mesos: 1. In `spark-env.sh` set some environment variables: - * `export MESOS_NATIVE_LIBRARY=path to libmesos.so`. This path is typically + * `export MESOS_NATIVE_JAVA_LIBRARY=path to libmesos.so`. This path is typically `prefix/lib/libmesos.so` where the prefix is `/usr/local` by default. See Mesos installation instructions above. On Mac OS X, the library is called `libmesos.dylib` instead of `libmesos.so`. @@ -167,9 +167,6 @@ acquire. By default, it will acquire *all* cores in the cluster (that get offere only makes sense if you run just one application at a time. You can cap the maximum number of cores using `conf.set(spark.cores.max, 10)` (for example). -# Known issues -- When using the fine-grained mode, make sure that your executors always leave 32 MB free on the slaves. Otherwise it can happen that your Spark job does not proceed anymore. Currently, Apache Mesos only offers resources if there are at least 32 MB memory allocatable. But as Spark allocates memory only for the executor and cpu only for tasks, it can happen on high slave memory usage that no new tasks will be started anymore. More details can be found in [MESOS-1688](https://issues.apache.org/jira/browse/MESOS-1688). Alternatively use the coarse-gained mode, which is not affected by this issue. - # Running Alongside Hadoop You can run Spark and Mesos alongside your existing Hadoop cluster by just launching them as a http://git-wip-us.apache.org/repos/asf/spark/blob/43fcab01/repl/scala-2.10/src/test/scala/org/apache/spark/repl/ReplSuite.scala -- diff --git
spark git commit: [SPARK-3619] Part 2. Upgrade to Mesos 0.21 to work around MESOS-1688
Repository: spark Updated Branches: refs/heads/master 62ede5383 - aa6536fa3 [SPARK-3619] Part 2. Upgrade to Mesos 0.21 to work around MESOS-1688 - MESOS_NATIVE_LIBRARY become deprecated - Chagned MESOS_NATIVE_LIBRARY to MESOS_NATIVE_JAVA_LIBRARY Author: Jongyoul Lee jongy...@gmail.com Closes #4361 from jongyoul/SPARK-3619-1 and squashes the following commits: f1ea91f [Jongyoul Lee] Merge branch 'SPARK-3619-1' of https://github.com/jongyoul/spark into SPARK-3619-1 a6a00c2 [Jongyoul Lee] [SPARK-3619] Upgrade to Mesos 0.21 to work around MESOS-1688 - Removed 'Known issues' section 2e15a21 [Jongyoul Lee] [SPARK-3619] Upgrade to Mesos 0.21 to work around MESOS-1688 - MESOS_NATIVE_LIBRARY become deprecated - Chagned MESOS_NATIVE_LIBRARY to MESOS_NATIVE_JAVA_LIBRARY 0dace7b [Jongyoul Lee] [SPARK-3619] Upgrade to Mesos 0.21 to work around MESOS-1688 - MESOS_NATIVE_LIBRARY become deprecated - Chagned MESOS_NATIVE_LIBRARY to MESOS_NATIVE_JAVA_LIBRARY Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/aa6536fa Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/aa6536fa Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/aa6536fa Branch: refs/heads/master Commit: aa6536fa3c2ed1cac47abc79fc22e273f0814858 Parents: 62ede53 Author: Jongyoul Lee jongy...@gmail.com Authored: Sun Mar 15 15:46:55 2015 + Committer: Sean Owen so...@cloudera.com Committed: Sun Mar 15 15:46:55 2015 + -- conf/spark-env.sh.template | 2 +- docs/running-on-mesos.md| 5 + .../src/test/scala/org/apache/spark/repl/ReplSuite.scala| 2 +- .../src/test/scala/org/apache/spark/repl/ReplSuite.scala| 2 +- 4 files changed, 4 insertions(+), 7 deletions(-) -- http://git-wip-us.apache.org/repos/asf/spark/blob/aa6536fa/conf/spark-env.sh.template -- diff --git a/conf/spark-env.sh.template b/conf/spark-env.sh.template index 0886b02..67f81d3 100755 --- a/conf/spark-env.sh.template +++ b/conf/spark-env.sh.template @@ -15,7 +15,7 @@ # - SPARK_PUBLIC_DNS, to set the public DNS name of the driver program # - SPARK_CLASSPATH, default classpath entries to append # - SPARK_LOCAL_DIRS, storage directories to use on this node for shuffle and RDD data -# - MESOS_NATIVE_LIBRARY, to point to your libmesos.so if you use Mesos +# - MESOS_NATIVE_JAVA_LIBRARY, to point to your libmesos.so if you use Mesos # Options read in YARN client mode # - HADOOP_CONF_DIR, to point Spark towards Hadoop configuration files http://git-wip-us.apache.org/repos/asf/spark/blob/aa6536fa/docs/running-on-mesos.md -- diff --git a/docs/running-on-mesos.md b/docs/running-on-mesos.md index e509e4b..59a3e9d 100644 --- a/docs/running-on-mesos.md +++ b/docs/running-on-mesos.md @@ -110,7 +110,7 @@ cluster, or `mesos://zk://host:2181` for a multi-master Mesos cluster using ZooK The driver also needs some configuration in `spark-env.sh` to interact properly with Mesos: 1. In `spark-env.sh` set some environment variables: - * `export MESOS_NATIVE_LIBRARY=path to libmesos.so`. This path is typically + * `export MESOS_NATIVE_JAVA_LIBRARY=path to libmesos.so`. This path is typically `prefix/lib/libmesos.so` where the prefix is `/usr/local` by default. See Mesos installation instructions above. On Mac OS X, the library is called `libmesos.dylib` instead of `libmesos.so`. @@ -167,9 +167,6 @@ acquire. By default, it will acquire *all* cores in the cluster (that get offere only makes sense if you run just one application at a time. You can cap the maximum number of cores using `conf.set(spark.cores.max, 10)` (for example). -# Known issues -- When using the fine-grained mode, make sure that your executors always leave 32 MB free on the slaves. Otherwise it can happen that your Spark job does not proceed anymore. Currently, Apache Mesos only offers resources if there are at least 32 MB memory allocatable. But as Spark allocates memory only for the executor and cpu only for tasks, it can happen on high slave memory usage that no new tasks will be started anymore. More details can be found in [MESOS-1688](https://issues.apache.org/jira/browse/MESOS-1688). Alternatively use the coarse-gained mode, which is not affected by this issue. - # Running Alongside Hadoop You can run Spark and Mesos alongside your existing Hadoop cluster by just launching them as a http://git-wip-us.apache.org/repos/asf/spark/blob/aa6536fa/repl/scala-2.10/src/test/scala/org/apache/spark/repl/ReplSuite.scala -- diff --git
svn commit: r1666875 - /spark/site/docs/1.3.0/sql-programming-guide.html
Author: pwendell Date: Mon Mar 16 05:30:06 2015 New Revision: 1666875 URL: http://svn.apache.org/r1666875 Log: Updating to incorperate doc changes in SPARK-6275 and SPARK-5310 Modified: spark/site/docs/1.3.0/sql-programming-guide.html Modified: spark/site/docs/1.3.0/sql-programming-guide.html URL: http://svn.apache.org/viewvc/spark/site/docs/1.3.0/sql-programming-guide.html?rev=1666875r1=1666874r2=1666875view=diff == --- spark/site/docs/1.3.0/sql-programming-guide.html (original) +++ spark/site/docs/1.3.0/sql-programming-guide.html Mon Mar 16 05:30:06 2015 @@ -113,7 +113,7 @@ ul id=markdown-toc lia href=#overviewOverview/a/li lia href=#dataframesDataFrames/aul - lia href=#starting-point-sqlcontextStarting Point: SQLContext/a/li + lia href=#starting-point-sqlcontextStarting Point: codeSQLContext/code/a/li lia href=#creating-dataframesCreating DataFrames/a/li lia href=#dataframe-operationsDataFrame Operations/a/li lia href=#running-sql-queries-programmaticallyRunning SQL Queries Programmatically/a/li @@ -133,6 +133,8 @@ /li lia href=#parquet-filesParquet Files/aul lia href=#loading-data-programmaticallyLoading Data Programmatically/a/li + lia href=#partition-discoveryPartition discovery/a/li + lia href=#schema-mergingSchema merging/a/li lia href=#configurationConfiguration/a/li /ul /li @@ -158,7 +160,7 @@ lia href=#unification-of-the-java-and-scala-apisUnification of the Java and Scala APIs/a/li lia href=#isolation-of-implicit-conversions-and-removal-of-dsl-package-scala-onlyIsolation of Implicit Conversions and Removal of dsl Package (Scala-only)/a/li lia href=#removal-of-the-type-aliases-in-orgapachesparksql-for-datatype-scala-onlyRemoval of the type aliases in org.apache.spark.sql for DataType (Scala-only)/a/li - lia href=#udf-registration-moved-to-sqlcontextudf-java--scalaUDF Registration Moved to sqlContext.udf (Java amp; Scala)/a/li + lia href=#udf-registration-moved-to-sqlcontextudf-java--scalaUDF Registration Moved to codesqlContext.udf/code (Java amp; Scala)/a/li lia href=#python-datatypes-no-longer-singletonsPython DataTypes No Longer Singletons/a/li /ul /li @@ -191,14 +193,14 @@ pAll of the examples on this page use sample data included in the Spark distribution and can be run in the codespark-shell/code or the codepyspark/code shell./p -h2 id=starting-point-sqlcontextStarting Point: SQLContext/h2 +h2 id=starting-point-sqlcontextStarting Point: codeSQLContext/code/h2 div class=codetabs div data-lang=scala pThe entry point into all functionality in Spark SQL is the -a href=api/scala/index.html#org.apache.spark.sql.SQLContextSQLContext/a class, or one of its -descendants. To create a basic SQLContext, all you need is a SparkContext./p +a href=api/scala/index.html#org.apache.spark.sql.`SQLContext`codeSQLContext/code/a class, or one of its +descendants. To create a basic codeSQLContext/code, all you need is a SparkContext./p div class=highlightprecode class=language-scala data-lang=scalaspan class=kval/span span class=nsc/spanspan class=k:/span span class=ktSparkContext/span span class=c1// An existing SparkContext./span span class=kval/span span class=nsqlContext/span span class=k=/span span class=knew/span span class=norg/spanspan class=o./spanspan class=napache/spanspan class=o./spanspan class=nspark/spanspan class=o./spanspan class=nsql/spanspan class=o./spanspan class=ncSQLContext/spanspan class=o(/spanspan class=nsc/spanspan class=o)/span @@ -211,8 +213,8 @@ descendants. To create a basic SQLConte div data-lang=java pThe entry point into all functionality in Spark SQL is the -a href=api/java/index.html#org.apache.spark.sql.SQLContextSQLContext/a class, or one of its -descendants. To create a basic SQLContext, all you need is a SparkContext./p +a href=api/java/index.html#org.apache.spark.sql.SQLContextcodeSQLContext/code/a class, or one of its +descendants. To create a basic codeSQLContext/code, all you need is a SparkContext./p div class=highlightprecode class=language-java data-lang=javaspan class=nJavaSparkContext/span span class=nsc/span span class=o=/span span class=o...;/span span class=c1// An existing JavaSparkContext./span span class=nSQLContext/span span class=nsqlContext/span span class=o=/span span class=knew/span span class=norg/spanspan class=o./spanspan class=naapache/spanspan class=o./spanspan class=naspark/spanspan class=o./spanspan class=nasql/spanspan class=o./spanspan class=naSQLContext/spanspan class=o(/spanspan class=nsc/spanspan class=o);/span/code/pre/div @@ -222,8 +224,8 @@ descendants. To create a basic SQLConte div data-lang=python pThe entry point into all relational functionality in