spark git commit: [SPARK-6781] [SQL] use sqlContext in python shell

2015-04-08 Thread rxin
Repository: spark
Updated Branches:
  refs/heads/master 66159c350 - 6ada4f6f5


[SPARK-6781] [SQL] use sqlContext in python shell

Use `sqlContext` in PySpark shell, make it consistent with SQL programming 
guide. `sqlCtx` is also kept for compatibility.

Author: Davies Liu dav...@databricks.com

Closes #5425 from davies/sqlCtx and squashes the following commits:

af67340 [Davies Liu] sqlCtx - sqlContext
15a278f [Davies Liu] use sqlContext in python shell


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/6ada4f6f
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/6ada4f6f
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/6ada4f6f

Branch: refs/heads/master
Commit: 6ada4f6f52cf1d992c7ab0c32318790cf08b0a0d
Parents: 66159c3
Author: Davies Liu dav...@databricks.com
Authored: Wed Apr 8 13:31:45 2015 -0700
Committer: Reynold Xin r...@databricks.com
Committed: Wed Apr 8 13:31:45 2015 -0700

--
 docs/ml-guide.md|  2 +-
 docs/sql-programming-guide.md   |  4 +-
 .../apache/spark/examples/sql/JavaSparkSQL.java | 20 ++---
 .../ml/simple_text_classification_pipeline.py   |  2 +-
 .../src/main/python/mllib/dataset_example.py|  6 +-
 python/pyspark/ml/classification.py |  4 +-
 python/pyspark/ml/feature.py|  4 +-
 python/pyspark/shell.py |  6 +-
 python/pyspark/sql/context.py   | 79 ++--
 python/pyspark/sql/dataframe.py |  6 +-
 python/pyspark/sql/functions.py |  2 +-
 python/pyspark/sql/types.py |  4 +-
 12 files changed, 69 insertions(+), 70 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/6ada4f6f/docs/ml-guide.md
--
diff --git a/docs/ml-guide.md b/docs/ml-guide.md
index c08c76d..771a071 100644
--- a/docs/ml-guide.md
+++ b/docs/ml-guide.md
@@ -493,7 +493,7 @@ from pyspark.ml.feature import HashingTF, Tokenizer
 from pyspark.sql import Row, SQLContext
 
 sc = SparkContext(appName=SimpleTextClassificationPipeline)
-sqlCtx = SQLContext(sc)
+sqlContext = SQLContext(sc)
 
 # Prepare training documents, which are labeled.
 LabeledDocument = Row(id, text, label)

http://git-wip-us.apache.org/repos/asf/spark/blob/6ada4f6f/docs/sql-programming-guide.md
--
diff --git a/docs/sql-programming-guide.md b/docs/sql-programming-guide.md
index 4441d6a..663f656 100644
--- a/docs/sql-programming-guide.md
+++ b/docs/sql-programming-guide.md
@@ -1642,7 +1642,7 @@ moved into the udf object in `SQLContext`.
 div data-lang=scala  markdown=1
 {% highlight java %}
 
-sqlCtx.udf.register(strLen, (s: String) = s.length())
+sqlContext.udf.register(strLen, (s: String) = s.length())
 
 {% endhighlight %}
 /div
@@ -1650,7 +1650,7 @@ sqlCtx.udf.register(strLen, (s: String) = s.length())
 div data-lang=java  markdown=1
 {% highlight java %}
 
-sqlCtx.udf().register(strLen, (String s) - { s.length(); });
+sqlContext.udf().register(strLen, (String s) - { s.length(); });
 
 {% endhighlight %}
 /div

http://git-wip-us.apache.org/repos/asf/spark/blob/6ada4f6f/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java
--
diff --git 
a/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java 
b/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java
index dee7948..8159ffb 100644
--- a/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java
+++ b/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java
@@ -55,7 +55,7 @@ public class JavaSparkSQL {
   public static void main(String[] args) throws Exception {
 SparkConf sparkConf = new SparkConf().setAppName(JavaSparkSQL);
 JavaSparkContext ctx = new JavaSparkContext(sparkConf);
-SQLContext sqlCtx = new SQLContext(ctx);
+SQLContext sqlContext = new SQLContext(ctx);
 
 System.out.println(=== Data source: RDD ===);
 // Load a text file and convert each line to a Java Bean.
@@ -74,11 +74,11 @@ public class JavaSparkSQL {
   });
 
 // Apply a schema to an RDD of Java Beans and register it as a table.
-DataFrame schemaPeople = sqlCtx.createDataFrame(people, Person.class);
+DataFrame schemaPeople = sqlContext.createDataFrame(people, Person.class);
 schemaPeople.registerTempTable(people);
 
 // SQL can be run over RDDs that have been registered as tables.
-DataFrame teenagers = sqlCtx.sql(SELECT name FROM people WHERE age = 13 
AND age = 19);
+DataFrame teenagers = sqlContext.sql(SELECT name FROM people WHERE age = 
13 AND age = 19);
 
 // The results of SQL queries are 

spark git commit: [SPARK-6781] [SQL] use sqlContext in python shell

2015-04-08 Thread rxin
Repository: spark
Updated Branches:
  refs/heads/branch-1.3 3b655680c - e1afd479b


[SPARK-6781] [SQL] use sqlContext in python shell

Use `sqlContext` in PySpark shell, make it consistent with SQL programming 
guide. `sqlCtx` is also kept for compatibility.

Author: Davies Liu dav...@databricks.com

Closes #5425 from davies/sqlCtx and squashes the following commits:

af67340 [Davies Liu] sqlCtx - sqlContext
15a278f [Davies Liu] use sqlContext in python shell

(cherry picked from commit 6ada4f6f52cf1d992c7ab0c32318790cf08b0a0d)
Signed-off-by: Reynold Xin r...@databricks.com


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/e1afd479
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/e1afd479
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/e1afd479

Branch: refs/heads/branch-1.3
Commit: e1afd479b3446483e6e1626afdec549cc214d80e
Parents: 3b65568
Author: Davies Liu dav...@databricks.com
Authored: Wed Apr 8 13:31:45 2015 -0700
Committer: Reynold Xin r...@databricks.com
Committed: Wed Apr 8 13:32:00 2015 -0700

--
 docs/ml-guide.md|  2 +-
 docs/sql-programming-guide.md   |  4 +-
 .../apache/spark/examples/sql/JavaSparkSQL.java | 20 ++---
 .../ml/simple_text_classification_pipeline.py   |  2 +-
 .../src/main/python/mllib/dataset_example.py|  6 +-
 python/pyspark/ml/classification.py |  4 +-
 python/pyspark/ml/feature.py|  4 +-
 python/pyspark/shell.py |  6 +-
 python/pyspark/sql/context.py   | 79 ++--
 python/pyspark/sql/dataframe.py |  6 +-
 python/pyspark/sql/functions.py |  2 +-
 python/pyspark/sql/types.py |  4 +-
 12 files changed, 69 insertions(+), 70 deletions(-)
--


http://git-wip-us.apache.org/repos/asf/spark/blob/e1afd479/docs/ml-guide.md
--
diff --git a/docs/ml-guide.md b/docs/ml-guide.md
index c08c76d..771a071 100644
--- a/docs/ml-guide.md
+++ b/docs/ml-guide.md
@@ -493,7 +493,7 @@ from pyspark.ml.feature import HashingTF, Tokenizer
 from pyspark.sql import Row, SQLContext
 
 sc = SparkContext(appName=SimpleTextClassificationPipeline)
-sqlCtx = SQLContext(sc)
+sqlContext = SQLContext(sc)
 
 # Prepare training documents, which are labeled.
 LabeledDocument = Row(id, text, label)

http://git-wip-us.apache.org/repos/asf/spark/blob/e1afd479/docs/sql-programming-guide.md
--
diff --git a/docs/sql-programming-guide.md b/docs/sql-programming-guide.md
index 4441d6a..663f656 100644
--- a/docs/sql-programming-guide.md
+++ b/docs/sql-programming-guide.md
@@ -1642,7 +1642,7 @@ moved into the udf object in `SQLContext`.
 div data-lang=scala  markdown=1
 {% highlight java %}
 
-sqlCtx.udf.register(strLen, (s: String) = s.length())
+sqlContext.udf.register(strLen, (s: String) = s.length())
 
 {% endhighlight %}
 /div
@@ -1650,7 +1650,7 @@ sqlCtx.udf.register(strLen, (s: String) = s.length())
 div data-lang=java  markdown=1
 {% highlight java %}
 
-sqlCtx.udf().register(strLen, (String s) - { s.length(); });
+sqlContext.udf().register(strLen, (String s) - { s.length(); });
 
 {% endhighlight %}
 /div

http://git-wip-us.apache.org/repos/asf/spark/blob/e1afd479/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java
--
diff --git 
a/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java 
b/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java
index dee7948..8159ffb 100644
--- a/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java
+++ b/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java
@@ -55,7 +55,7 @@ public class JavaSparkSQL {
   public static void main(String[] args) throws Exception {
 SparkConf sparkConf = new SparkConf().setAppName(JavaSparkSQL);
 JavaSparkContext ctx = new JavaSparkContext(sparkConf);
-SQLContext sqlCtx = new SQLContext(ctx);
+SQLContext sqlContext = new SQLContext(ctx);
 
 System.out.println(=== Data source: RDD ===);
 // Load a text file and convert each line to a Java Bean.
@@ -74,11 +74,11 @@ public class JavaSparkSQL {
   });
 
 // Apply a schema to an RDD of Java Beans and register it as a table.
-DataFrame schemaPeople = sqlCtx.createDataFrame(people, Person.class);
+DataFrame schemaPeople = sqlContext.createDataFrame(people, Person.class);
 schemaPeople.registerTempTable(people);
 
 // SQL can be run over RDDs that have been registered as tables.
-DataFrame teenagers = sqlCtx.sql(SELECT name FROM people WHERE age = 13 
AND age = 19);
+