Yup, i meant union only. On Wed, Jan 7, 2015, 16:19 Sean Owen <so...@cloudera.com> wrote:
> I think you mean union(). Yes, you could also simply make an RDD for each > file, and use SparkContext.union() to put them together. > > On Wed, Jan 7, 2015 at 9:51 AM, Raghavendra Pandey < > raghavendra.pan...@gmail.com> wrote: > >> You can also use join function of rdd. This is actually kind of append >> funtion that add up all the rdds and create one uber rdd. >> >> On Wed, Jan 7, 2015, 14:30 rkgurram <rkgur...@gmail.com> wrote: >> >>> Thank you for the response, sure will try that out. >>> >>> Currently I changed my code such that the first map "files.map" to >>> "files.flatMap", which I guess will do similar what you are saying, it >>> gives >>> me a List[] of elements (in this case LabeledPoints, I could also do >>> RDDs) >>> which I then turned into a mega RDD. The current problem seems to be >>> gone, I >>> no longer get the NPE but further down I am getting a indexOutOfBounds, >>> so >>> trying to figure out if the original problem is manifesting itself as a >>> new >>> one. >>> >>> >>> Regards >>> -Ravi >>> >>> >>> >>> >>> -- >>> View this message in context: http://apache-spark-user-list. >>> 1001560.n3.nabble.com/How-to-merge-a-RDD-of-RDDs-into-one- >>> uber-RDD-tp20986p21012.html >>> Sent from the Apache Spark User List mailing list archive at Nabble.com. >>> >>> --------------------------------------------------------------------- >>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >>> For additional commands, e-mail: user-h...@spark.apache.org >>> >>> >