spark git commit: [SPARK-10778] [MLLIB] Implement toString for AssociationRules.Rule
Repository: spark Updated Branches: refs/heads/master 6f94d56a9 -> 299b43992 [SPARK-10778] [MLLIB] Implement toString for AssociationRules.Rule I implemented toString for AssociationRules.Rule, format like `[x, y] => {z}: 1.0` Author: y-shimizuCloses #8904 from y-shimizu/master. Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/299b4399 Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/299b4399 Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/299b4399 Branch: refs/heads/master Commit: 299b439920f980cce4c4f4e4a8436a5145efeaa3 Parents: 6f94d56 Author: y-shimizu Authored: Sun Sep 27 16:36:03 2015 +0100 Committer: Sean Owen Committed: Sun Sep 27 16:36:03 2015 +0100 -- .../scala/org/apache/spark/mllib/fpm/AssociationRules.scala | 5 + 1 file changed, 5 insertions(+) -- http://git-wip-us.apache.org/repos/asf/spark/blob/299b4399/mllib/src/main/scala/org/apache/spark/mllib/fpm/AssociationRules.scala -- diff --git a/mllib/src/main/scala/org/apache/spark/mllib/fpm/AssociationRules.scala b/mllib/src/main/scala/org/apache/spark/mllib/fpm/AssociationRules.scala index 95c688c..07eb750 100644 --- a/mllib/src/main/scala/org/apache/spark/mllib/fpm/AssociationRules.scala +++ b/mllib/src/main/scala/org/apache/spark/mllib/fpm/AssociationRules.scala @@ -142,5 +142,10 @@ object AssociationRules { def javaConsequent: java.util.List[Item] = { consequent.toList.asJava } + +override def toString: String = { + s"${antecedent.mkString("{", ",", "}")} => " + +s"${consequent.mkString("{", ",", "}")}: ${confidence}" +} } } - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
spark git commit: [SPARK-10741] [SQL] Hive Query Having/OrderBy against Parquet table is not working
Repository: spark Updated Branches: refs/heads/branch-1.5 3fb011a48 -> e0c3212a9 [SPARK-10741] [SQL] Hive Query Having/OrderBy against Parquet table is not working https://issues.apache.org/jira/browse/SPARK-10741 I choose the second approach: do not change output exprIds when convert MetastoreRelation to LogicalRelation Author: Wenchen FanCloses #8889 from cloud-fan/hot-bug. (cherry picked from commit 418e5e4cbdaab87addb91ac0bb2245ff0213ac81) Signed-off-by: Yin Huai Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/e0c3212a Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/e0c3212a Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/e0c3212a Branch: refs/heads/branch-1.5 Commit: e0c3212a9b42e3e704b070da4ac25b68c584427f Parents: 3fb011a Author: Wenchen Fan Authored: Sun Sep 27 09:08:38 2015 -0700 Committer: Yin Huai Committed: Sun Sep 27 09:08:51 2015 -0700 -- .../analysis/MultiInstanceRelation.scala| 1 - .../catalyst/expressions/namedExpressions.scala | 8 +++ .../scala/org/apache/spark/sql/DataFrame.scala | 2 +- .../datasources/DataSourceStrategy.scala| 18 +++--- .../execution/datasources/LogicalRelation.scala | 33 +++--- .../spark/sql/execution/datasources/rules.scala | 14 ++--- .../parquet/ParquetFilterSuite.scala| 2 +- .../ParquetPartitionDiscoverySuite.scala| 2 +- .../spark/sql/hive/HiveMetastoreCatalog.scala | 64 +--- .../spark/sql/hive/execution/commands.scala | 2 +- .../sql/hive/InsertIntoHiveTableSuite.scala | 4 +- .../sql/hive/MetastoreDataSourcesSuite.scala| 2 +- .../sql/hive/execution/SQLQuerySuite.scala | 27 - .../apache/spark/sql/hive/parquetSuites.scala | 8 +-- .../sql/sources/hadoopFsRelationSuites.scala| 2 +- 15 files changed, 103 insertions(+), 86 deletions(-) -- http://git-wip-us.apache.org/repos/asf/spark/blob/e0c3212a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/MultiInstanceRelation.scala -- diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/MultiInstanceRelation.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/MultiInstanceRelation.scala index 35b7402..394be47 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/MultiInstanceRelation.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/MultiInstanceRelation.scala @@ -17,7 +17,6 @@ package org.apache.spark.sql.catalyst.analysis -import org.apache.spark.sql.catalyst.rules.Rule import org.apache.spark.sql.catalyst.plans.logical.LogicalPlan /** http://git-wip-us.apache.org/repos/asf/spark/blob/e0c3212a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/namedExpressions.scala -- diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/namedExpressions.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/namedExpressions.scala index 6f173b5..5768c60 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/namedExpressions.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/namedExpressions.scala @@ -236,6 +236,14 @@ case class AttributeReference( } } + def withExprId(newExprId: ExprId): AttributeReference = { +if (exprId == newExprId) { + this +} else { + AttributeReference(name, dataType, nullable, metadata)(newExprId, qualifiers) +} + } + override def toString: String = s"$name#${exprId.id}$typeSuffix" } http://git-wip-us.apache.org/repos/asf/spark/blob/e0c3212a/sql/core/src/main/scala/org/apache/spark/sql/DataFrame.scala -- diff --git a/sql/core/src/main/scala/org/apache/spark/sql/DataFrame.scala b/sql/core/src/main/scala/org/apache/spark/sql/DataFrame.scala index 49b40a3..f024db7 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/DataFrame.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/DataFrame.scala @@ -1562,7 +1562,7 @@ class DataFrame private[sql]( */ def inputFiles: Array[String] = { val files: Seq[String] = logicalPlan.collect { - case LogicalRelation(fsBasedRelation: FileRelation) => + case LogicalRelation(fsBasedRelation: FileRelation, _) => fsBasedRelation.inputFiles case fr: FileRelation => fr.inputFiles
svn commit: r10668 - /dev/spark/spark-1.5.1-rc1/
Author: rxin Date: Sun Sep 27 21:45:30 2015 New Revision: 10668 Log: Add spark-1.5.1-rc1 Added: dev/spark/spark-1.5.1-rc1/ dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-cdh4.tgz (with props) dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-cdh4.tgz.asc dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-cdh4.tgz.md5 dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-cdh4.tgz.sha dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-hadoop1-scala2.11.tgz (with props) dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-hadoop1-scala2.11.tgz.asc dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-hadoop1-scala2.11.tgz.md5 dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-hadoop1-scala2.11.tgz.sha dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-hadoop2.3.tgz (with props) dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-hadoop2.3.tgz.asc dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-hadoop2.3.tgz.md5 dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-hadoop2.3.tgz.sha dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-hadoop2.4.tgz (with props) dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-hadoop2.4.tgz.asc dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-hadoop2.4.tgz.md5 dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-hadoop2.4.tgz.sha dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-hadoop2.6.tgz (with props) dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-hadoop2.6.tgz.asc dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-hadoop2.6.tgz.md5.txt dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-hadoop2.6.tgz.sha dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-without-hadoop.tgz (with props) dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-without-hadoop.tgz.asc dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-without-hadoop.tgz.md5.txt dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-without-hadoop.tgz.sha dev/spark/spark-1.5.1-rc1/spark-1.5.1.tgz (with props) dev/spark/spark-1.5.1-rc1/spark-1.5.1.tgz.asc dev/spark/spark-1.5.1-rc1/spark-1.5.1.tgz.md5.txt dev/spark/spark-1.5.1-rc1/spark-1.5.1.tgz.sha Added: dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-cdh4.tgz == Binary file - no diff available. Propchange: dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-cdh4.tgz -- svn:mime-type = application/octet-stream Added: dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-cdh4.tgz.asc == --- dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-cdh4.tgz.asc (added) +++ dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-cdh4.tgz.asc Sun Sep 27 21:45:30 2015 @@ -0,0 +1,11 @@ +-BEGIN PGP SIGNATURE- +Version: GnuPG v2.0.14 (GNU/Linux) + +iQEcBAABAgAGBQJWA5RhAAoJEHxsEF/8jtCJ6awH/RNikGEyrSCW7xeKSXJoUbRr +uATtgH1yF9bRnQomHw0oZr1rIXLeuUWJ64IvEAJ6fexnuuPiI7xFOEawLEXJd6hZ +QI96SMOhS5VgVBz8Ox9TfsXushR0MFbuRgvY/TbUrX0ySvxQNMhFdb4ygZOq377O +gwz4qIkOMQvBCzXW9tW47HpMIgHy+JDScHhd+R8Mf6EK4utbB086qKkpr8FUx7wH +ODm6rVbTBYu+HdGPRnZIUjQNEhO6q4/HtHmSegQQziNFDWMy1I5X2uFsJF9H5O4K +rsAIUzNDP8Xgy/i9rbanIVsjLBC6SA8mQUrnFga+FeOmqirOfpnIuDjJtn4QwEY= +=klAd +-END PGP SIGNATURE- Added: dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-cdh4.tgz.md5 == --- dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-cdh4.tgz.md5 (added) +++ dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-cdh4.tgz.md5 Sun Sep 27 21:45:30 2015 @@ -0,0 +1 @@ +spark-1.5.1-bin-cdh4.tgz: 87 9B DD 56 34 D5 E0 B0 5E 85 54 7D E6 10 A8 37 Added: dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-cdh4.tgz.sha == --- dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-cdh4.tgz.sha (added) +++ dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-cdh4.tgz.sha Sun Sep 27 21:45:30 2015 @@ -0,0 +1,3 @@ +spark-1.5.1-bin-cdh4.tgz: 0ED48B83 96B8E4A2 15ED9E31 E3459F7C EA743C9C 1882AD30 + 81645904 E4E356C6 C5ECDF2A 97791F5C 50A8A8C6 06FD1F22 + C8BB5EFE AF8164B3 52B3CC0F 17A39756 Added: dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-hadoop1-scala2.11.tgz == Binary file - no diff available. Propchange: dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-hadoop1-scala2.11.tgz -- svn:mime-type = application/octet-stream Added: dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-hadoop1-scala2.11.tgz.asc == --- dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-hadoop1-scala2.11.tgz.asc (added) +++ dev/spark/spark-1.5.1-rc1/spark-1.5.1-bin-hadoop1-scala2.11.tgz.asc Sun Sep 27 21:45:30 2015 @@ -0,0 +1,11 @@ +-BEGIN PGP SIGNATURE- +Version: GnuPG v2.0.14 (GNU/Linux) + +iQEcBAABAgAGBQJWA5QdAAoJEHxsEF/8jtCJl/gH+wbrAAEivL+tjbp+4zZ9b9Ko
svn commit: r10669 - /dev/spark/spark-1.5.1-rc1/ /release/spark/spark-1.5.1/
Author: rxin Date: Sun Sep 27 21:47:42 2015 New Revision: 10669 Log: Spark 1.5.1 binary Added: release/spark/spark-1.5.1/ - copied from r10668, dev/spark/spark-1.5.1-rc1/ Removed: dev/spark/spark-1.5.1-rc1/ - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
svn commit: r1705570 - in /spark: documentation.md site/documentation.html
Author: pwendell Date: Sun Sep 27 20:22:29 2015 New Revision: 1705570 URL: http://svn.apache.org/viewvc?rev=1705570=rev Log: Adding new Spark books to documentation page. Modified: spark/documentation.md spark/site/documentation.html Modified: spark/documentation.md URL: http://svn.apache.org/viewvc/spark/documentation.md?rev=1705570=1705569=1705570=diff == --- spark/documentation.md (original) +++ spark/documentation.md Sun Sep 27 20:22:29 2015 @@ -146,6 +146,8 @@ Slides, videos and EC2-based exercises f http://shop.oreilly.com/product/0636920035091.do;>Advanced Analytics with Spark, by Sandy Ryza, Uri Laserson, Sean Owen and Josh Wills (O'Reilly Media) http://manning.com/malak/;>Spark GraphX in Action, by Michael Malak (Manning) https://www.packtpub.com/big-data-and-business-intelligence/fast-data-processing-spark-second-edition;>Fast Data Processing with Spark, by Krishna Sankar and Holden Karau (Packt Publishing) + https://www.packtpub.com/big-data-and-business-intelligence/machine-learning-spark;>Machine Learning with Spark, by Nick Pentreath (Packt Publishing) + https://www.packtpub.com/big-data-and-business-intelligence/spark-cookbook;>Spark Cookbook, by Rishi Yadav (Packt Publishing) Examples Modified: spark/site/documentation.html URL: http://svn.apache.org/viewvc/spark/site/documentation.html?rev=1705570=1705569=1705570=diff == --- spark/site/documentation.html (original) +++ spark/site/documentation.html Sun Sep 27 20:22:29 2015 @@ -304,6 +304,8 @@ Slides, videos and EC2-based exercises f http://shop.oreilly.com/product/0636920035091.do;>Advanced Analytics with Spark, by Sandy Ryza, Uri Laserson, Sean Owen and Josh Wills (O'Reilly Media) http://manning.com/malak/;>Spark GraphX in Action, by Michael Malak (Manning) https://www.packtpub.com/big-data-and-business-intelligence/fast-data-processing-spark-second-edition;>Fast Data Processing with Spark, by Krishna Sankar and Holden Karau (Packt Publishing) + https://www.packtpub.com/big-data-and-business-intelligence/machine-learning-spark;>Machine Learning with Spark, by Nick Pentreath (Packt Publishing) + https://www.packtpub.com/big-data-and-business-intelligence/spark-cookbook;>Spark Cookbook, by Rishi Yadav (Packt Publishing) Examples - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
spark git commit: [SPARK-10720] [SQL] [JAVA] Add a java wrapper to create a dataframe from a local list of java beans
Repository: spark Updated Branches: refs/heads/master 418e5e4cb -> 8ecba3e86 [SPARK-10720] [SQL] [JAVA] Add a java wrapper to create a dataframe from a local list of java beans Similar to SPARK-10630 it would be nice if Java users didn't have to parallelize there data explicitly (as Scala users already can skip). Issue came up in http://stackoverflow.com/questions/32613413/apache-spark-machine-learning-cant-get-estimator-example-to-work Author: Holden KarauCloses #8879 from holdenk/SPARK-10720-add-a-java-wrapper-to-create-a-dataframe-from-a-local-list-of-java-beans. Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/8ecba3e8 Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/8ecba3e8 Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/8ecba3e8 Branch: refs/heads/master Commit: 8ecba3e86e53834413da8b4299f5791545cae12e Parents: 418e5e4 Author: Holden Karau Authored: Sun Sep 27 21:16:15 2015 +0100 Committer: Sean Owen Committed: Sun Sep 27 21:16:15 2015 +0100 -- .../scala/org/apache/spark/sql/SQLContext.scala | 51 +++- .../apache/spark/sql/JavaDataFrameSuite.java| 22 +++-- 2 files changed, 56 insertions(+), 17 deletions(-) -- http://git-wip-us.apache.org/repos/asf/spark/blob/8ecba3e8/sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala -- diff --git a/sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala b/sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala index 1bd4e26..cb0a3e3 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala @@ -17,7 +17,7 @@ package org.apache.spark.sql -import java.beans.Introspector +import java.beans.{BeanInfo, Introspector} import java.util.Properties import java.util.concurrent.atomic.AtomicReference @@ -499,21 +499,12 @@ class SQLContext(@transient val sparkContext: SparkContext) * @since 1.3.0 */ def createDataFrame(rdd: RDD[_], beanClass: Class[_]): DataFrame = { -val attributeSeq = getSchema(beanClass) +val attributeSeq: Seq[AttributeReference] = getSchema(beanClass) val className = beanClass.getName val rowRdd = rdd.mapPartitions { iter => // BeanInfo is not serializable so we must rediscover it remotely for each partition. val localBeanInfo = Introspector.getBeanInfo(Utils.classForName(className)) - val extractors = -localBeanInfo.getPropertyDescriptors.filterNot(_.getName == "class").map(_.getReadMethod) - val methodsToConverts = extractors.zip(attributeSeq).map { case (e, attr) => -(e, CatalystTypeConverters.createToCatalystConverter(attr.dataType)) - } - iter.map { row => -new GenericInternalRow( - methodsToConverts.map { case (e, convert) => convert(e.invoke(row)) }.toArray[Any] -): InternalRow - } + SQLContext.beansToRows(iter, localBeanInfo, attributeSeq) } DataFrame(this, LogicalRDD(attributeSeq, rowRdd)(this)) } @@ -531,6 +522,23 @@ class SQLContext(@transient val sparkContext: SparkContext) } /** + * Applies a schema to an List of Java Beans. + * + * WARNING: Since there is no guaranteed ordering for fields in a Java Bean, + * SELECT * queries will return the columns in an undefined order. + * @group dataframes + * @since 1.6.0 + */ + def createDataFrame(data: java.util.List[_], beanClass: Class[_]): DataFrame = { +val attrSeq = getSchema(beanClass) +val className = beanClass.getName +val beanInfo = Introspector.getBeanInfo(beanClass) +val rows = SQLContext.beansToRows(data.asScala.iterator, beanInfo, attrSeq) +DataFrame(self, LocalRelation(attrSeq, rows.toSeq)) + } + + + /** * :: Experimental :: * Returns a [[DataFrameReader]] that can be used to read data in as a [[DataFrame]]. * {{{ @@ -1229,4 +1237,23 @@ object SQLContext { lastInstantiatedContext.set(sqlContext) } } + + /** + * Converts an iterator of Java Beans to InternalRow using the provided + * bean info & schema. This is not related to the singleton, but is a static + * method for internal use. + */ + private def beansToRows(data: Iterator[_], beanInfo: BeanInfo, attrs: Seq[AttributeReference]): + Iterator[InternalRow] = { +val extractors = + beanInfo.getPropertyDescriptors.filterNot(_.getName == "class").map(_.getReadMethod) +val methodsToConverts = extractors.zip(attrs).map { case (e, attr) => + (e, CatalystTypeConverters.createToCatalystConverter(attr.dataType)) +} +data.map{ element => + new GenericInternalRow( +
Git Push Summary
Repository: spark Updated Tags: refs/tags/v1.5.1-rc1 [deleted] 4df97937d - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
Git Push Summary
Repository: spark Updated Tags: refs/tags/v1.5.1 [created] 4f894dd69 - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
spark git commit: add doc for spark.streaming.stopGracefullyOnShutdown
Repository: spark Updated Branches: refs/heads/master 8ecba3e86 -> fb4c7be74 add doc for spark.streaming.stopGracefullyOnShutdown Author: Bin WangCloses #8898 from wb14123/doc. Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/fb4c7be7 Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/fb4c7be7 Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/fb4c7be7 Branch: refs/heads/master Commit: fb4c7be747a35ea733df6010194cd76db449adc9 Parents: 8ecba3e Author: Bin Wang Authored: Sun Sep 27 21:26:54 2015 +0100 Committer: Sean Owen Committed: Sun Sep 27 21:26:54 2015 +0100 -- docs/configuration.md | 8 1 file changed, 8 insertions(+) -- http://git-wip-us.apache.org/repos/asf/spark/blob/fb4c7be7/docs/configuration.md -- diff --git a/docs/configuration.md b/docs/configuration.md index 284f97a..d99092e 100644 --- a/docs/configuration.md +++ b/docs/configuration.md @@ -1491,6 +1491,14 @@ Apart from these, the following properties are also available, and may be useful + spark.streaming.stopGracefullyOnShutdown + false + +If true, Spark shuts down the StreamingContext gracefully on JVM +shutdown rather than immediately. + + + spark.streaming.kafka.maxRatePerPartition not set - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
svn commit: r1705591 - in /spark/site/docs: ./ 1.5.1/ 1.5.1/api/ 1.5.1/api/R/ 1.5.1/api/java/ 1.5.1/api/java/lib/ 1.5.1/api/java/org/ 1.5.1/api/java/org/apache/ 1.5.1/api/java/org/apache/spark/ 1.5.1/
Author: rxin Date: Sun Sep 27 22:59:57 2015 New Revision: 1705591 URL: http://svn.apache.org/viewvc?rev=1705591=rev Log: Add 1.5.1 docs [This commit notification would consist of 805 parts, which exceeds the limit of 50 ones, so it was shortened to the summary.] - To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org
spark git commit: [SPARK-10741] [SQL] Hive Query Having/OrderBy against Parquet table is not working
Repository: spark Updated Branches: refs/heads/master 299b43992 -> 418e5e4cb [SPARK-10741] [SQL] Hive Query Having/OrderBy against Parquet table is not working https://issues.apache.org/jira/browse/SPARK-10741 I choose the second approach: do not change output exprIds when convert MetastoreRelation to LogicalRelation Author: Wenchen FanCloses #8889 from cloud-fan/hot-bug. Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/418e5e4c Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/418e5e4c Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/418e5e4c Branch: refs/heads/master Commit: 418e5e4cbdaab87addb91ac0bb2245ff0213ac81 Parents: 299b439 Author: Wenchen Fan Authored: Sun Sep 27 09:08:38 2015 -0700 Committer: Yin Huai Committed: Sun Sep 27 09:08:38 2015 -0700 -- .../analysis/MultiInstanceRelation.scala| 1 - .../catalyst/expressions/namedExpressions.scala | 8 +++ .../scala/org/apache/spark/sql/DataFrame.scala | 2 +- .../datasources/DataSourceStrategy.scala| 18 +++--- .../execution/datasources/LogicalRelation.scala | 33 +++--- .../spark/sql/execution/datasources/rules.scala | 14 ++--- .../parquet/ParquetFilterSuite.scala| 2 +- .../ParquetPartitionDiscoverySuite.scala| 2 +- .../spark/sql/hive/HiveMetastoreCatalog.scala | 64 +--- .../spark/sql/hive/execution/commands.scala | 2 +- .../sql/hive/InsertIntoHiveTableSuite.scala | 4 +- .../sql/hive/MetastoreDataSourcesSuite.scala| 2 +- .../sql/hive/execution/SQLQuerySuite.scala | 27 - .../apache/spark/sql/hive/parquetSuites.scala | 8 +-- .../sql/sources/hadoopFsRelationSuites.scala| 2 +- 15 files changed, 103 insertions(+), 86 deletions(-) -- http://git-wip-us.apache.org/repos/asf/spark/blob/418e5e4c/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/MultiInstanceRelation.scala -- diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/MultiInstanceRelation.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/MultiInstanceRelation.scala index 35b7402..394be47 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/MultiInstanceRelation.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/MultiInstanceRelation.scala @@ -17,7 +17,6 @@ package org.apache.spark.sql.catalyst.analysis -import org.apache.spark.sql.catalyst.rules.Rule import org.apache.spark.sql.catalyst.plans.logical.LogicalPlan /** http://git-wip-us.apache.org/repos/asf/spark/blob/418e5e4c/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/namedExpressions.scala -- diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/namedExpressions.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/namedExpressions.scala index 6f173b5..5768c60 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/namedExpressions.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/expressions/namedExpressions.scala @@ -236,6 +236,14 @@ case class AttributeReference( } } + def withExprId(newExprId: ExprId): AttributeReference = { +if (exprId == newExprId) { + this +} else { + AttributeReference(name, dataType, nullable, metadata)(newExprId, qualifiers) +} + } + override def toString: String = s"$name#${exprId.id}$typeSuffix" } http://git-wip-us.apache.org/repos/asf/spark/blob/418e5e4c/sql/core/src/main/scala/org/apache/spark/sql/DataFrame.scala -- diff --git a/sql/core/src/main/scala/org/apache/spark/sql/DataFrame.scala b/sql/core/src/main/scala/org/apache/spark/sql/DataFrame.scala index f9995da..9c67ad1 100644 --- a/sql/core/src/main/scala/org/apache/spark/sql/DataFrame.scala +++ b/sql/core/src/main/scala/org/apache/spark/sql/DataFrame.scala @@ -1595,7 +1595,7 @@ class DataFrame private[sql]( */ def inputFiles: Array[String] = { val files: Seq[String] = logicalPlan.collect { - case LogicalRelation(fsBasedRelation: FileRelation) => + case LogicalRelation(fsBasedRelation: FileRelation, _) => fsBasedRelation.inputFiles case fr: FileRelation => fr.inputFiles http://git-wip-us.apache.org/repos/asf/spark/blob/418e5e4c/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/DataSourceStrategy.scala -- diff --git