spark git commit: [SPARK-6781] [SQL] use sqlContext in python shell
Repository: spark Updated Branches: refs/heads/master 66159c350 - 6ada4f6f5 [SPARK-6781] [SQL] use sqlContext in python shell Use `sqlContext` in PySpark shell, make it consistent with SQL programming guide. `sqlCtx` is also kept for compatibility. Author: Davies Liu dav...@databricks.com Closes #5425 from davies/sqlCtx and squashes the following commits: af67340 [Davies Liu] sqlCtx - sqlContext 15a278f [Davies Liu] use sqlContext in python shell Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/6ada4f6f Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/6ada4f6f Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/6ada4f6f Branch: refs/heads/master Commit: 6ada4f6f52cf1d992c7ab0c32318790cf08b0a0d Parents: 66159c3 Author: Davies Liu dav...@databricks.com Authored: Wed Apr 8 13:31:45 2015 -0700 Committer: Reynold Xin r...@databricks.com Committed: Wed Apr 8 13:31:45 2015 -0700 -- docs/ml-guide.md| 2 +- docs/sql-programming-guide.md | 4 +- .../apache/spark/examples/sql/JavaSparkSQL.java | 20 ++--- .../ml/simple_text_classification_pipeline.py | 2 +- .../src/main/python/mllib/dataset_example.py| 6 +- python/pyspark/ml/classification.py | 4 +- python/pyspark/ml/feature.py| 4 +- python/pyspark/shell.py | 6 +- python/pyspark/sql/context.py | 79 ++-- python/pyspark/sql/dataframe.py | 6 +- python/pyspark/sql/functions.py | 2 +- python/pyspark/sql/types.py | 4 +- 12 files changed, 69 insertions(+), 70 deletions(-) -- http://git-wip-us.apache.org/repos/asf/spark/blob/6ada4f6f/docs/ml-guide.md -- diff --git a/docs/ml-guide.md b/docs/ml-guide.md index c08c76d..771a071 100644 --- a/docs/ml-guide.md +++ b/docs/ml-guide.md @@ -493,7 +493,7 @@ from pyspark.ml.feature import HashingTF, Tokenizer from pyspark.sql import Row, SQLContext sc = SparkContext(appName=SimpleTextClassificationPipeline) -sqlCtx = SQLContext(sc) +sqlContext = SQLContext(sc) # Prepare training documents, which are labeled. LabeledDocument = Row(id, text, label) http://git-wip-us.apache.org/repos/asf/spark/blob/6ada4f6f/docs/sql-programming-guide.md -- diff --git a/docs/sql-programming-guide.md b/docs/sql-programming-guide.md index 4441d6a..663f656 100644 --- a/docs/sql-programming-guide.md +++ b/docs/sql-programming-guide.md @@ -1642,7 +1642,7 @@ moved into the udf object in `SQLContext`. div data-lang=scala markdown=1 {% highlight java %} -sqlCtx.udf.register(strLen, (s: String) = s.length()) +sqlContext.udf.register(strLen, (s: String) = s.length()) {% endhighlight %} /div @@ -1650,7 +1650,7 @@ sqlCtx.udf.register(strLen, (s: String) = s.length()) div data-lang=java markdown=1 {% highlight java %} -sqlCtx.udf().register(strLen, (String s) - { s.length(); }); +sqlContext.udf().register(strLen, (String s) - { s.length(); }); {% endhighlight %} /div http://git-wip-us.apache.org/repos/asf/spark/blob/6ada4f6f/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java -- diff --git a/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java b/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java index dee7948..8159ffb 100644 --- a/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java +++ b/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java @@ -55,7 +55,7 @@ public class JavaSparkSQL { public static void main(String[] args) throws Exception { SparkConf sparkConf = new SparkConf().setAppName(JavaSparkSQL); JavaSparkContext ctx = new JavaSparkContext(sparkConf); -SQLContext sqlCtx = new SQLContext(ctx); +SQLContext sqlContext = new SQLContext(ctx); System.out.println(=== Data source: RDD ===); // Load a text file and convert each line to a Java Bean. @@ -74,11 +74,11 @@ public class JavaSparkSQL { }); // Apply a schema to an RDD of Java Beans and register it as a table. -DataFrame schemaPeople = sqlCtx.createDataFrame(people, Person.class); +DataFrame schemaPeople = sqlContext.createDataFrame(people, Person.class); schemaPeople.registerTempTable(people); // SQL can be run over RDDs that have been registered as tables. -DataFrame teenagers = sqlCtx.sql(SELECT name FROM people WHERE age = 13 AND age = 19); +DataFrame teenagers = sqlContext.sql(SELECT name FROM people WHERE age = 13 AND age = 19); // The results of SQL queries are
spark git commit: [SPARK-6781] [SQL] use sqlContext in python shell
Repository: spark Updated Branches: refs/heads/branch-1.3 3b655680c - e1afd479b [SPARK-6781] [SQL] use sqlContext in python shell Use `sqlContext` in PySpark shell, make it consistent with SQL programming guide. `sqlCtx` is also kept for compatibility. Author: Davies Liu dav...@databricks.com Closes #5425 from davies/sqlCtx and squashes the following commits: af67340 [Davies Liu] sqlCtx - sqlContext 15a278f [Davies Liu] use sqlContext in python shell (cherry picked from commit 6ada4f6f52cf1d992c7ab0c32318790cf08b0a0d) Signed-off-by: Reynold Xin r...@databricks.com Project: http://git-wip-us.apache.org/repos/asf/spark/repo Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/e1afd479 Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/e1afd479 Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/e1afd479 Branch: refs/heads/branch-1.3 Commit: e1afd479b3446483e6e1626afdec549cc214d80e Parents: 3b65568 Author: Davies Liu dav...@databricks.com Authored: Wed Apr 8 13:31:45 2015 -0700 Committer: Reynold Xin r...@databricks.com Committed: Wed Apr 8 13:32:00 2015 -0700 -- docs/ml-guide.md| 2 +- docs/sql-programming-guide.md | 4 +- .../apache/spark/examples/sql/JavaSparkSQL.java | 20 ++--- .../ml/simple_text_classification_pipeline.py | 2 +- .../src/main/python/mllib/dataset_example.py| 6 +- python/pyspark/ml/classification.py | 4 +- python/pyspark/ml/feature.py| 4 +- python/pyspark/shell.py | 6 +- python/pyspark/sql/context.py | 79 ++-- python/pyspark/sql/dataframe.py | 6 +- python/pyspark/sql/functions.py | 2 +- python/pyspark/sql/types.py | 4 +- 12 files changed, 69 insertions(+), 70 deletions(-) -- http://git-wip-us.apache.org/repos/asf/spark/blob/e1afd479/docs/ml-guide.md -- diff --git a/docs/ml-guide.md b/docs/ml-guide.md index c08c76d..771a071 100644 --- a/docs/ml-guide.md +++ b/docs/ml-guide.md @@ -493,7 +493,7 @@ from pyspark.ml.feature import HashingTF, Tokenizer from pyspark.sql import Row, SQLContext sc = SparkContext(appName=SimpleTextClassificationPipeline) -sqlCtx = SQLContext(sc) +sqlContext = SQLContext(sc) # Prepare training documents, which are labeled. LabeledDocument = Row(id, text, label) http://git-wip-us.apache.org/repos/asf/spark/blob/e1afd479/docs/sql-programming-guide.md -- diff --git a/docs/sql-programming-guide.md b/docs/sql-programming-guide.md index 4441d6a..663f656 100644 --- a/docs/sql-programming-guide.md +++ b/docs/sql-programming-guide.md @@ -1642,7 +1642,7 @@ moved into the udf object in `SQLContext`. div data-lang=scala markdown=1 {% highlight java %} -sqlCtx.udf.register(strLen, (s: String) = s.length()) +sqlContext.udf.register(strLen, (s: String) = s.length()) {% endhighlight %} /div @@ -1650,7 +1650,7 @@ sqlCtx.udf.register(strLen, (s: String) = s.length()) div data-lang=java markdown=1 {% highlight java %} -sqlCtx.udf().register(strLen, (String s) - { s.length(); }); +sqlContext.udf().register(strLen, (String s) - { s.length(); }); {% endhighlight %} /div http://git-wip-us.apache.org/repos/asf/spark/blob/e1afd479/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java -- diff --git a/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java b/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java index dee7948..8159ffb 100644 --- a/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java +++ b/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java @@ -55,7 +55,7 @@ public class JavaSparkSQL { public static void main(String[] args) throws Exception { SparkConf sparkConf = new SparkConf().setAppName(JavaSparkSQL); JavaSparkContext ctx = new JavaSparkContext(sparkConf); -SQLContext sqlCtx = new SQLContext(ctx); +SQLContext sqlContext = new SQLContext(ctx); System.out.println(=== Data source: RDD ===); // Load a text file and convert each line to a Java Bean. @@ -74,11 +74,11 @@ public class JavaSparkSQL { }); // Apply a schema to an RDD of Java Beans and register it as a table. -DataFrame schemaPeople = sqlCtx.createDataFrame(people, Person.class); +DataFrame schemaPeople = sqlContext.createDataFrame(people, Person.class); schemaPeople.registerTempTable(people); // SQL can be run over RDDs that have been registered as tables. -DataFrame teenagers = sqlCtx.sql(SELECT name FROM people WHERE age = 13 AND age = 19); +