Wow, it really was that easy! The implicit joining works a treat.
Many thanks,
Jon
On 13 October 2014 22:58, Stephen Boesch java...@gmail.com wrote:
is the following what you are looking for?
scala sc.parallelize(myMap.map{ case (k,v) = (k,v) }.toSeq)
res2:
is the following what you are looking for?
scala sc.parallelize(myMap.map{ case (k,v) = (k,v) }.toSeq)
res2: org.apache.spark.rdd.RDD[(String, Int)] = ParallelCollectionRDD[0] at
parallelize at console:21
2014-10-13 14:02 GMT-07:00 jon.g.massey jon.g.mas...@gmail.com:
Hi guys,
Just
Map.toSeq already does that even. You can skip the map. You can put
together Maps with ++ too. You should have an RDD of pairs then, but
to get the special RDD functions you're looking for remember to import
SparkContext._
On Mon, Oct 13, 2014 at 10:58 PM, Stephen Boesch java...@gmail.com wrote: