[ 
https://issues.apache.org/jira/browse/SPARK-12783?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15102351#comment-15102351
 ] 

Muthu Jayakumar edited comment on SPARK-12783 at 1/15/16 7:34 PM:
------------------------------------------------------------------

I tried the following, but got similar error...

{code}
case class MyMap(map: scala.collection.immutable.Map[String, String])

case class TestCaseClass(a: String, b: String){
  def toMyMap: MyMap = {
    MyMap(Map(a->b))
  }


  def toStr: String = {
    a
  }
}

//main thread...
val df1 = sqlContext.createDataset(Seq(TestCaseClass("2015-05-01", "data1"), 
TestCaseClass("2015-05-01", "data2"))).toDF() 
//.withColumn("swh_date_to_common_request_id_map", f1(col("_1"), 
col("_2"))).drop("_1").drop("_2")
  df1.as[TestCaseClass].map(_.toStr).show() //works fine
  df1.as[TestCaseClass].map(_.toMyMap).show() //error
  df1.as[TestCaseClass].map(each=> each.a -> each.b).show() //works fine
{code}

{quote}
Serialization stack:
        - object not serializable (class: 
scala.reflect.runtime.SynchronizedSymbols$SynchronizedSymbol$$anon$1, value: 
package lang)
        - field (class: scala.reflect.internal.Types$ThisType, name: sym, type: 
class scala.reflect.internal.Symbols$Symbol)
        - object (class scala.reflect.internal.Types$UniqueThisType, 
java.lang.type)
        - field (class: scala.reflect.internal.Types$TypeRef, name: pre, type: 
class scala.reflect.internal.Types$Type)
        - object (class scala.reflect.internal.Types$ClassNoArgsTypeRef, String)
        - field (class: scala.reflect.internal.Types$TypeRef, name: normalized, 
type: class scala.reflect.internal.Types$Type)
        - object (class scala.reflect.internal.Types$AliasNoArgsTypeRef, String)
        - field (class: 
org.apache.spark.sql.catalyst.ScalaReflection$$anonfun$6, name: keyType$1, 
type: class scala.reflect.api.Types$TypeApi)
        - object (class 
org.apache.spark.sql.catalyst.ScalaReflection$$anonfun$6, <function1>)
        - field (class: org.apache.spark.sql.catalyst.expressions.MapObjects, 
name: function, type: interface scala.Function1)
        - object (class org.apache.spark.sql.catalyst.expressions.MapObjects, 
mapobjects(<function1>,invoke(upcast('map,MapType(StringType,StringType,true),- 
field (class: "scala.collection.immutable.Map", name: "map"),- root class: 
"collector.MyMap"),keyArray,ArrayType(StringType,true)),StringType))
        - field (class: org.apache.spark.sql.catalyst.expressions.Invoke, name: 
targetObject, type: class org.apache.spark.sql.catalyst.expressions.Expression)
        - object (class org.apache.spark.sql.catalyst.expressions.Invoke, 
invoke(mapobjects(<function1>,invoke(upcast('map,MapType(StringType,StringType,true),-
 field (class: "scala.collection.immutable.Map", name: "map"),- root class: 
"collector.MyMap"),keyArray,ArrayType(StringType,true)),StringType),array,ObjectType(class
 [Ljava.lang.Object;)))
        - writeObject data (class: 
scala.collection.immutable.List$SerializationProxy)
        - object (class scala.collection.immutable.List$SerializationProxy, 
scala.collection.immutable.List$SerializationProxy@2660f093)
        - writeReplace data (class: 
scala.collection.immutable.List$SerializationProxy)
        - object (class scala.collection.immutable.$colon$colon, 
List(invoke(mapobjects(<function1>,invoke(upcast('map,MapType(StringType,StringType,true),-
 field (class: "scala.collection.immutable.Map", name: "map"),- root class: 
"collector.MyMap"),keyArray,ArrayType(StringType,true)),StringType),array,ObjectType(class
 [Ljava.lang.Object;)), 
invoke(mapobjects(<function1>,invoke(upcast('map,MapType(StringType,StringType,true),-
 field (class: "scala.collection.immutable.Map", name: "map"),- root class: 
"collector.MyMap"),valueArray,ArrayType(StringType,true)),StringType),array,ObjectType(class
 [Ljava.lang.Object;))))
        - field (class: org.apache.spark.sql.catalyst.expressions.StaticInvoke, 
name: arguments, type: interface scala.collection.Seq)
        - object (class org.apache.spark.sql.catalyst.expressions.StaticInvoke, 
staticinvoke(class 
org.apache.spark.sql.catalyst.util.ArrayBasedMapData$,ObjectType(interface 
scala.collection.Map),toScalaMap,invoke(mapobjects(<function1>,invoke(upcast('map,MapType(StringType,StringType,true),-
 field (class: "scala.collection.immutable.Map", name: "map"),- root class: 
"collector.MyMap"),keyArray,ArrayType(StringType,true)),StringType),array,ObjectType(class
 
[Ljava.lang.Object;)),invoke(mapobjects(<function1>,invoke(upcast('map,MapType(StringType,StringType,true),-
 field (class: "scala.collection.immutable.Map", name: "map"),- root class: 
"collector.MyMap"),valueArray,ArrayType(StringType,true)),StringType),array,ObjectType(class
 [Ljava.lang.Object;)),true))
        - writeObject data (class: 
scala.collection.immutable.List$SerializationProxy)
        - object (class scala.collection.immutable.List$SerializationProxy, 
scala.collection.immutable.List$SerializationProxy@72af5ac7)
        - writeReplace data (class: 
scala.collection.immutable.List$SerializationProxy)
        - object (class scala.collection.immutable.$colon$colon, 
List(staticinvoke(class 
org.apache.spark.sql.catalyst.util.ArrayBasedMapData$,ObjectType(interface 
scala.collection.Map),toScalaMap,invoke(mapobjects(<function1>,invoke(upcast('map,MapType(StringType,StringType,true),-
 field (class: "scala.collection.immutable.Map", name: "map"),- root class: 
"collector.MyMap"),keyArray,ArrayType(StringType,true)),StringType),array,ObjectType(class
 
[Ljava.lang.Object;)),invoke(mapobjects(<function1>,invoke(upcast('map,MapType(StringType,StringType,true),-
 field (class: "scala.collection.immutable.Map", name: "map"),- root class: 
"collector.MyMap"),valueArray,ArrayType(StringType,true)),StringType),array,ObjectType(class
 [Ljava.lang.Object;)),true)))
        - field (class: org.apache.spark.sql.catalyst.expressions.NewInstance, 
name: arguments, type: interface scala.collection.Seq)
        - object (class org.apache.spark.sql.catalyst.expressions.NewInstance, 
newinstance(class collector.MyMap,staticinvoke(class 
org.apache.spark.sql.catalyst.util.ArrayBasedMapData$,ObjectType(interface 
scala.collection.Map),toScalaMap,invoke(mapobjects(<function1>,invoke(upcast('map,MapType(StringType,StringType,true),-
 field (class: "scala.collection.immutable.Map", name: "map"),- root class: 
"collector.MyMap"),keyArray,ArrayType(StringType,true)),StringType),array,ObjectType(class
 
[Ljava.lang.Object;)),invoke(mapobjects(<function1>,invoke(upcast('map,MapType(StringType,StringType,true),-
 field (class: "scala.collection.immutable.Map", name: "map"),- root class: 
"collector.MyMap"),valueArray,ArrayType(StringType,true)),StringType),array,ObjectType(class
 [Ljava.lang.Object;)),true),false,ObjectType(class collector.MyMap),None))
        - field (class: 
org.apache.spark.sql.catalyst.encoders.ExpressionEncoder, name: 
fromRowExpression, type: class 
org.apache.spark.sql.catalyst.expressions.Expression)
        - object (class 
org.apache.spark.sql.catalyst.encoders.ExpressionEncoder, 
class[map#ExprId(15,584e7368-7577-4fd1-bd2b-2541f51a2cd5): map<string,string>])
        - field (class: org.apache.spark.sql.execution.MapPartitions, name: 
uEncoder, type: class org.apache.spark.sql.catalyst.encoders.ExpressionEncoder)
        - object (class org.apache.spark.sql.execution.MapPartitions, 
!MapPartitions <function1>, class[a[0]: string, b[0]: string], 
class[map#ExprId(15,584e7368-7577-4fd1-bd2b-2541f51a2cd5): map<string,string>], 
[map#19]
+- LocalTableScan [a#2,b#3], 
[[0,180000000a,2800000005,2d35302d35313032,3130,3161746164],[0,180000000a,2800000005,2d35302d35313032,3130,3261746164]]
)
        - field (class: 
org.apache.spark.sql.execution.MapPartitions$$anonfun$8, name: $outer, type: 
class org.apache.spark.sql.execution.MapPartitions)
        - object (class 
org.apache.spark.sql.execution.MapPartitions$$anonfun$8, <function1>)
        - field (class: 
org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1, name: f$22, type: 
interface scala.Function1)
        - object (class 
org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1, <function0>)
        - field (class: 
org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1$$anonfun$apply$21, 
name: $outer, type: class 
org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1)
        - object (class 
org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1$$anonfun$apply$21, 
<function3>)
        - field (class: org.apache.spark.rdd.MapPartitionsRDD, name: f, type: 
interface scala.Function3)
        - object (class org.apache.spark.rdd.MapPartitionsRDD, 
MapPartitionsRDD[4] at show at CollectorSparkTest.scala:53)
        - field (class: org.apache.spark.NarrowDependency, name: _rdd, type: 
class org.apache.spark.rdd.RDD)
        - object (class org.apache.spark.OneToOneDependency, 
org.apache.spark.OneToOneDependency@322a3ac9)
        - writeObject data (class: 
scala.collection.immutable.List$SerializationProxy)
        - object (class scala.collection.immutable.List$SerializationProxy, 
scala.collection.immutable.List$SerializationProxy@4b0f9525)
        - writeReplace data (class: 
scala.collection.immutable.List$SerializationProxy)
        - object (class scala.collection.immutable.$colon$colon, 
List(org.apache.spark.OneToOneDependency@322a3ac9))
        - field (class: org.apache.spark.rdd.RDD, name: 
org$apache$spark$rdd$RDD$$dependencies_, type: interface scala.collection.Seq)
        - object (class org.apache.spark.rdd.MapPartitionsRDD, 
MapPartitionsRDD[5] at show at CollectorSparkTest.scala:53)
        - field (class: scala.Tuple2, name: _1, type: class java.lang.Object)
        - object (class scala.Tuple2, (MapPartitionsRDD[5] at show at 
CollectorSparkTest.scala:53,<function2>))
        at 
org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1431)
        at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1419)
        at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1418)
        at 
scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
        at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
        at 
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1418)
        at 
org.apache.spark.scheduler.DAGScheduler.submitMissingTasks(DAGScheduler.scala:1010)
        at 
org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$submitStage(DAGScheduler.scala:921)
        at 
org.apache.spark.scheduler.DAGScheduler.handleJobSubmitted(DAGScheduler.scala:861)
        at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1607)
        at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1599)
        at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1588)
        at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
        at 
org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:620)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:1832)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:1845)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:1858)
        at 
org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:212)
        at 
org.apache.spark.sql.execution.Limit.executeCollect(basicOperators.scala:165)
        at 
org.apache.spark.sql.execution.SparkPlan.executeCollectPublic(SparkPlan.scala:174)
        at 
org.apache.spark.sql.DataFrame$$anonfun$org$apache$spark$sql$DataFrame$$execute$1$1.apply(DataFrame.scala:1538)
        at 
org.apache.spark.sql.DataFrame$$anonfun$org$apache$spark$sql$DataFrame$$execute$1$1.apply(DataFrame.scala:1538)
        at 
org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:56)
        at 
org.apache.spark.sql.DataFrame.withNewExecutionId(DataFrame.scala:2125)
        at 
org.apache.spark.sql.DataFrame.org$apache$spark$sql$DataFrame$$execute$1(DataFrame.scala:1537)
        at 
org.apache.spark.sql.DataFrame.org$apache$spark$sql$DataFrame$$collect(DataFrame.scala:1544)
        at 
org.apache.spark.sql.DataFrame$$anonfun$head$1.apply(DataFrame.scala:1414)
        at 
org.apache.spark.sql.DataFrame$$anonfun$head$1.apply(DataFrame.scala:1413)
        at org.apache.spark.sql.DataFrame.withCallback(DataFrame.scala:2138)
        at org.apache.spark.sql.DataFrame.head(DataFrame.scala:1413)
        at org.apache.spark.sql.DataFrame.take(DataFrame.scala:1495)
        at org.apache.spark.sql.DataFrame.showString(DataFrame.scala:171)
        at org.apache.spark.sql.DataFrame.show(DataFrame.scala:394)
        at org.apache.spark.sql.Dataset.show(Dataset.scala:228)
        at org.apache.spark.sql.Dataset.show(Dataset.scala:192)
        at org.apache.spark.sql.Dataset.show(Dataset.scala:200)
{quote}



was (Author: babloo80):
I tried the following, but got similar error...

{code}
case class MyMap(map: scala.collection.immutable.Map[String, String])

case class TestCaseClass(a: String, b: String){
  def toMyMap: MyMap = {
    MyMap(Map(a->b))
  }

  def toStr: String = {
    a
  }
}

//main thread...
val df1 = sqlContext.createDataset(Seq(TestCaseClass("2015-05-01", "data1"), 
TestCaseClass("2015-05-01", "data2"))).toDF() 
//.withColumn("swh_date_to_common_request_id_map", f1(col("_1"), 
col("_2"))).drop("_1").drop("_2")
  df1.as[TestCaseClass].map(_.toStr).show() //works fine
  df1.as[TestCaseClass].map(_.toMyMap).show() //error
  df1.as[TestCaseClass].map(each=> each.a -> each.b).show() //works fine
{code}

> Dataset map serialization error
> -------------------------------
>
>                 Key: SPARK-12783
>                 URL: https://issues.apache.org/jira/browse/SPARK-12783
>             Project: Spark
>          Issue Type: Bug
>          Components: SQL
>    Affects Versions: 1.6.0
>            Reporter: Muthu Jayakumar
>            Assignee: Wenchen Fan
>            Priority: Critical
>
> When Dataset API is used to map to another case class, an error is thrown.
> {code}
> case class MyMap(map: Map[String, String])
> case class TestCaseClass(a: String, b: String){
>   def toMyMap: MyMap = {
>     MyMap(Map(a->b))
>   }
>   def toStr: String = {
>     a
>   }
> }
> //Main method section below
> import sqlContext.implicits._
> val df1 = sqlContext.createDataset(Seq(TestCaseClass("2015-05-01", "data1"), 
> TestCaseClass("2015-05-01", "data2"))).toDF()
> df1.as[TestCaseClass].map(_.toStr).show() //works fine
> df1.as[TestCaseClass].map(_.toMyMap).show() //fails
> {code}
> Error message:
> {quote}
> Caused by: java.io.NotSerializableException: 
> scala.reflect.runtime.SynchronizedSymbols$SynchronizedSymbol$$anon$1
> Serialization stack:
>       - object not serializable (class: 
> scala.reflect.runtime.SynchronizedSymbols$SynchronizedSymbol$$anon$1, value: 
> package lang)
>       - field (class: scala.reflect.internal.Types$ThisType, name: sym, type: 
> class scala.reflect.internal.Symbols$Symbol)
>       - object (class scala.reflect.internal.Types$UniqueThisType, 
> java.lang.type)
>       - field (class: scala.reflect.internal.Types$TypeRef, name: pre, type: 
> class scala.reflect.internal.Types$Type)
>       - object (class scala.reflect.internal.Types$ClassNoArgsTypeRef, String)
>       - field (class: scala.reflect.internal.Types$TypeRef, name: normalized, 
> type: class scala.reflect.internal.Types$Type)
>       - object (class scala.reflect.internal.Types$AliasNoArgsTypeRef, String)
>       - field (class: 
> org.apache.spark.sql.catalyst.ScalaReflection$$anonfun$6, name: keyType$1, 
> type: class scala.reflect.api.Types$TypeApi)
>       - object (class 
> org.apache.spark.sql.catalyst.ScalaReflection$$anonfun$6, <function1>)
>       - field (class: org.apache.spark.sql.catalyst.expressions.MapObjects, 
> name: function, type: interface scala.Function1)
>       - object (class org.apache.spark.sql.catalyst.expressions.MapObjects, 
> mapobjects(<function1>,invoke(upcast('map,MapType(StringType,StringType,true),-
>  field (class: "scala.collection.immutable.Map", name: "map"),- root class: 
> "collector.MyMap"),keyArray,ArrayType(StringType,true)),StringType))
>       - field (class: org.apache.spark.sql.catalyst.expressions.Invoke, name: 
> targetObject, type: class 
> org.apache.spark.sql.catalyst.expressions.Expression)
>       - object (class org.apache.spark.sql.catalyst.expressions.Invoke, 
> invoke(mapobjects(<function1>,invoke(upcast('map,MapType(StringType,StringType,true),-
>  field (class: "scala.collection.immutable.Map", name: "map"),- root class: 
> "collector.MyMap"),keyArray,ArrayType(StringType,true)),StringType),array,ObjectType(class
>  [Ljava.lang.Object;)))
>       - writeObject data (class: 
> scala.collection.immutable.List$SerializationProxy)
>       - object (class scala.collection.immutable.List$SerializationProxy, 
> scala.collection.immutable.List$SerializationProxy@4c7e3aab)
>       - writeReplace data (class: 
> scala.collection.immutable.List$SerializationProxy)
>       - object (class scala.collection.immutable.$colon$colon, 
> List(invoke(mapobjects(<function1>,invoke(upcast('map,MapType(StringType,StringType,true),-
>  field (class: "scala.collection.immutable.Map", name: "map"),- root class: 
> "collector.MyMap"),keyArray,ArrayType(StringType,true)),StringType),array,ObjectType(class
>  [Ljava.lang.Object;)), 
> invoke(mapobjects(<function1>,invoke(upcast('map,MapType(StringType,StringType,true),-
>  field (class: "scala.collection.immutable.Map", name: "map"),- root class: 
> "collector.MyMap"),valueArray,ArrayType(StringType,true)),StringType),array,ObjectType(class
>  [Ljava.lang.Object;))))
>       - field (class: org.apache.spark.sql.catalyst.expressions.StaticInvoke, 
> name: arguments, type: interface scala.collection.Seq)
>       - object (class org.apache.spark.sql.catalyst.expressions.StaticInvoke, 
> staticinvoke(class 
> org.apache.spark.sql.catalyst.util.ArrayBasedMapData$,ObjectType(interface 
> scala.collection.Map),toScalaMap,invoke(mapobjects(<function1>,invoke(upcast('map,MapType(StringType,StringType,true),-
>  field (class: "scala.collection.immutable.Map", name: "map"),- root class: 
> "collector.MyMap"),keyArray,ArrayType(StringType,true)),StringType),array,ObjectType(class
>  
> [Ljava.lang.Object;)),invoke(mapobjects(<function1>,invoke(upcast('map,MapType(StringType,StringType,true),-
>  field (class: "scala.collection.immutable.Map", name: "map"),- root class: 
> "collector.MyMap"),valueArray,ArrayType(StringType,true)),StringType),array,ObjectType(class
>  [Ljava.lang.Object;)),true))
>       - writeObject data (class: 
> scala.collection.immutable.List$SerializationProxy)
>       - object (class scala.collection.immutable.List$SerializationProxy, 
> scala.collection.immutable.List$SerializationProxy@78d9820)
>       - writeReplace data (class: 
> scala.collection.immutable.List$SerializationProxy)
>       - object (class scala.collection.immutable.$colon$colon, 
> List(staticinvoke(class 
> org.apache.spark.sql.catalyst.util.ArrayBasedMapData$,ObjectType(interface 
> scala.collection.Map),toScalaMap,invoke(mapobjects(<function1>,invoke(upcast('map,MapType(StringType,StringType,true),-
>  field (class: "scala.collection.immutable.Map", name: "map"),- root class: 
> "collector.MyMap"),keyArray,ArrayType(StringType,true)),StringType),array,ObjectType(class
>  
> [Ljava.lang.Object;)),invoke(mapobjects(<function1>,invoke(upcast('map,MapType(StringType,StringType,true),-
>  field (class: "scala.collection.immutable.Map", name: "map"),- root class: 
> "collector.MyMap"),valueArray,ArrayType(StringType,true)),StringType),array,ObjectType(class
>  [Ljava.lang.Object;)),true)))
>       - field (class: org.apache.spark.sql.catalyst.expressions.NewInstance, 
> name: arguments, type: interface scala.collection.Seq)
>       - object (class org.apache.spark.sql.catalyst.expressions.NewInstance, 
> newinstance(class collector.MyMap,staticinvoke(class 
> org.apache.spark.sql.catalyst.util.ArrayBasedMapData$,ObjectType(interface 
> scala.collection.Map),toScalaMap,invoke(mapobjects(<function1>,invoke(upcast('map,MapType(StringType,StringType,true),-
>  field (class: "scala.collection.immutable.Map", name: "map"),- root class: 
> "collector.MyMap"),keyArray,ArrayType(StringType,true)),StringType),array,ObjectType(class
>  
> [Ljava.lang.Object;)),invoke(mapobjects(<function1>,invoke(upcast('map,MapType(StringType,StringType,true),-
>  field (class: "scala.collection.immutable.Map", name: "map"),- root class: 
> "collector.MyMap"),valueArray,ArrayType(StringType,true)),StringType),array,ObjectType(class
>  [Ljava.lang.Object;)),true),false,ObjectType(class collector.MyMap),None))
>       - field (class: 
> org.apache.spark.sql.catalyst.encoders.ExpressionEncoder, name: 
> fromRowExpression, type: class 
> org.apache.spark.sql.catalyst.expressions.Expression)
>       - object (class 
> org.apache.spark.sql.catalyst.encoders.ExpressionEncoder, 
> class[map#ExprId(9,5d198984-4022-43b2-a2a3-ddbb214ba0ef): map<string,string>])
>       - field (class: org.apache.spark.sql.execution.MapPartitions, name: 
> uEncoder, type: class 
> org.apache.spark.sql.catalyst.encoders.ExpressionEncoder)
>       - object (class org.apache.spark.sql.execution.MapPartitions, 
> !MapPartitions <function1>, class[a[0]: string, b[0]: string], 
> class[map#ExprId(9,5d198984-4022-43b2-a2a3-ddbb214ba0ef): 
> map<string,string>], [map#13]
> +- LocalTableScan [a#2,b#3], 
> [[0,180000000a,2800000005,2d35302d35313032,3130,3161746164],[0,180000000a,2800000005,2d35302d35313032,3130,3261746164]]
> )
>       - field (class: 
> org.apache.spark.sql.execution.MapPartitions$$anonfun$8, name: $outer, type: 
> class org.apache.spark.sql.execution.MapPartitions)
>       - object (class 
> org.apache.spark.sql.execution.MapPartitions$$anonfun$8, <function1>)
>       - field (class: 
> org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1, name: f$22, type: 
> interface scala.Function1)
>       - object (class 
> org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1, <function0>)
>       - field (class: 
> org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1$$anonfun$apply$21, 
> name: $outer, type: class 
> org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1)
>       - object (class 
> org.apache.spark.rdd.RDD$$anonfun$mapPartitionsInternal$1$$anonfun$apply$21, 
> <function3>)
>       - field (class: org.apache.spark.rdd.MapPartitionsRDD, name: f, type: 
> interface scala.Function3)
>       - object (class org.apache.spark.rdd.MapPartitionsRDD, 
> MapPartitionsRDD[1] at show at CollectorSparkTest.scala:50)
>       - field (class: org.apache.spark.NarrowDependency, name: _rdd, type: 
> class org.apache.spark.rdd.RDD)
>       - object (class org.apache.spark.OneToOneDependency, 
> org.apache.spark.OneToOneDependency@4d60c27d)
>       - writeObject data (class: 
> scala.collection.immutable.List$SerializationProxy)
>       - object (class scala.collection.immutable.List$SerializationProxy, 
> scala.collection.immutable.List$SerializationProxy@6c436651)
>       - writeReplace data (class: 
> scala.collection.immutable.List$SerializationProxy)
>       - object (class scala.collection.immutable.$colon$colon, 
> List(org.apache.spark.OneToOneDependency@4d60c27d))
>       - field (class: org.apache.spark.rdd.RDD, name: 
> org$apache$spark$rdd$RDD$$dependencies_, type: interface scala.collection.Seq)
>       - object (class org.apache.spark.rdd.MapPartitionsRDD, 
> MapPartitionsRDD[2] at show at CollectorSparkTest.scala:50)
>       - field (class: scala.Tuple2, name: _1, type: class java.lang.Object)
>       - object (class scala.Tuple2, (MapPartitionsRDD[2] at show at 
> CollectorSparkTest.scala:50,<function2>))
>       at 
> org.apache.spark.serializer.SerializationDebugger$.improveException(SerializationDebugger.scala:40)
>       at 
> org.apache.spark.serializer.JavaSerializationStream.writeObject(JavaSerializer.scala:47)
>       at 
> org.apache.spark.serializer.JavaSerializerInstance.serialize(JavaSerializer.scala:101)
>       at 
> org.apache.spark.scheduler.DAGScheduler.submitMissingTasks(DAGScheduler.scala:1003)
>       at 
> org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$submitStage(DAGScheduler.scala:921)
>       at 
> org.apache.spark.scheduler.DAGScheduler.handleJobSubmitted(DAGScheduler.scala:861)
>       at 
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1607)
>       at 
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1599)
>       at 
> org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1588)
>       at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
> {quote}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to