In Spark 1.1, I'm seeing tasks with callbacks that don't involve my code at all! I'd seen something like this before in 1.0.0, but the behavior seems to be back
apply at Option.scala:120 <http://localhost:4040/stages/stage?id=52&attempt=0> org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:202) scala.Option.getOrElse(Option.scala:120) org.apache.spark.rdd.RDD.partitions(RDD.scala:202) org.apache.spark.rdd.FilteredRDD.getPartitions(FilteredRDD.scala:29) org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:204) org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:202) scala.Option.getOrElse(Option.scala:120) org.apache.spark.rdd.RDD.partitions(RDD.scala:202) org.apache.spark.rdd.MappedRDD.getPartitions(MappedRDD.scala:28) org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:204) org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:202) scala.Option.getOrElse(Option.scala:120) org.apache.spark.rdd.RDD.partitions(RDD.scala:202) org.apache.spark.rdd.FilteredRDD.getPartitions(FilteredRDD.scala:29) org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:204) org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:202) scala.Option.getOrElse(Option.scala:120) org.apache.spark.rdd.RDD.partitions(RDD.scala:202) org.apache.spark.rdd.MappedRDD.getPartitions(MappedRDD.scala:28) org.apache.spark.rdd.RDD$$anonfun$partitions$2.apply(RDD.scala:204) Ideas on what might be going on?