[ https://issues.apache.org/jira/browse/SPARK-18855?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15753120#comment-15753120 ]
Linbo commented on SPARK-18855: ------------------------------- Tried several ways, the more "Spark" way is trying to create TraversableRDDFunctions file and implicit def rddToTraversableRDDFunctions[U](rdd: RDD[TraversableRDDFunctions[U]]) inside RDD object. But it's hard to make this method generic because class RDD is invariant. I will close this issue. It's more impactful that this should go on Dataset. > Add RDD flatten function > ------------------------ > > Key: SPARK-18855 > URL: https://issues.apache.org/jira/browse/SPARK-18855 > Project: Spark > Issue Type: New Feature > Components: Spark Core > Reporter: Linbo > Priority: Minor > Labels: flatten, rdd > > A new RDD flatten function is similar to flatten function of scala > collections: > {code:title=spark-shell|borderStyle=solid} > scala> val rdd = sc.makeRDD(List(List(1, 2, 3), List(4, 5), List(6))) > rdd: org.apache.spark.rdd.RDD[List[Int]] = ParallelCollectionRDD[0] at > makeRDD at <console>:24 > scala> rdd.flatten.collect > res0: Array[Int] = Array(1, 2, 3, 4, 5, 6) > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org