Repository: spark
Updated Branches:
  refs/heads/master 84e2c8bfe -> f92cde24e


[SPARK-3447][SQL] Remove explicit conversion with JListWrapper to avoid NPE

Author: Michael Armbrust <mich...@databricks.com>

Closes #2323 from marmbrus/kryoJListNPE and squashes the following commits:

9634f11 [Michael Armbrust] Rollback JSON RDD changes
4d4d93c [Michael Armbrust] Merge remote-tracking branch 'origin/master' into 
kryoJListNPE
646976b [Michael Armbrust] Fix JSON RDD Conversion too
59065bc [Michael Armbrust] Remove explicit conversion to avoid NPE


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/f92cde24
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/f92cde24
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/f92cde24

Branch: refs/heads/master
Commit: f92cde24e8f30dddd5bcec71bb3687498c1406da
Parents: 84e2c8b
Author: Michael Armbrust <mich...@databricks.com>
Authored: Wed Sep 10 20:59:40 2014 -0700
Committer: Michael Armbrust <mich...@databricks.com>
Committed: Wed Sep 10 20:59:40 2014 -0700

----------------------------------------------------------------------
 sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala | 4 +---
 1 file changed, 1 insertion(+), 3 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/spark/blob/f92cde24/sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala
----------------------------------------------------------------------
diff --git a/sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala 
b/sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala
index a2f334a..c551c7c 100644
--- a/sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala
+++ b/sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala
@@ -460,7 +460,6 @@ class SQLContext(@transient val sparkContext: SparkContext)
       rdd: RDD[Array[Any]],
       schema: StructType): SchemaRDD = {
     import scala.collection.JavaConversions._
-    import scala.collection.convert.Wrappers.{JListWrapper, JMapWrapper}
 
     def needsConversion(dataType: DataType): Boolean = dataType match {
       case ByteType => true
@@ -482,8 +481,7 @@ class SQLContext(@transient val sparkContext: SparkContext)
       case (null, _) => null
 
       case (c: java.util.List[_], ArrayType(elementType, _)) =>
-        val converted = c.map { e => convert(e, elementType)}
-        JListWrapper(converted)
+        c.map { e => convert(e, elementType)}: Seq[Any]
 
       case (c, ArrayType(elementType, _)) if c.getClass.isArray =>
         c.asInstanceOf[Array[_]].map(e => convert(e, elementType)): Seq[Any]


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org
For additional commands, e-mail: commits-h...@spark.apache.org

Reply via email to