An RDD is a fault-tolerant distributed structure. It is the primary abstraction in Spark.
I would strongly suggest that you have a look at the following to get a basic idea. http://www.cs.berkeley.edu/~pwendell/strataconf/api/core/spark/RDD.html http://spark.apache.org/docs/latest/quick-start.html#basics https://www.usenix.org/conference/nsdi12/technical-sessions/presentation/zaharia On Sat, Sep 13, 2014 at 12:06 AM, Deep Pradhan <pradhandeep1...@gmail.com> wrote: > Take for example this: > I have declared one queue *val queue = Queue.empty[Int]*, which is a pure > scala line in the program. I actually want the queue to be an RDD but there > are no direct methods to create RDD which is a queue right? What say do you > have on this? > Does there exist something like: *Create and RDD which is a queue *? > > On Sat, Sep 13, 2014 at 8:43 AM, Hari Shreedharan < > hshreedha...@cloudera.com> wrote: > >> No, Scala primitives remain primitives. Unless you create an RDD using >> one of the many methods - you would not be able to access any of the RDD >> methods. There is no automatic porting. Spark is an application as far as >> scala is concerned - there is no compilation (except of course, the scala, >> JIT compilation etc). >> >> On Fri, Sep 12, 2014 at 8:04 PM, Deep Pradhan <pradhandeep1...@gmail.com> >> wrote: >> >>> I know that unpersist is a method on RDD. >>> But my confusion is that, when we port our Scala programs to Spark, >>> doesn't everything change to RDDs? >>> >>> On Fri, Sep 12, 2014 at 10:16 PM, Nicholas Chammas < >>> nicholas.cham...@gmail.com> wrote: >>> >>>> unpersist is a method on RDDs. RDDs are abstractions introduced by >>>> Spark. >>>> >>>> An Int is just a Scala Int. You can't call unpersist on Int in Scala, >>>> and that doesn't change in Spark. >>>> >>>> On Fri, Sep 12, 2014 at 12:33 PM, Deep Pradhan < >>>> pradhandeep1...@gmail.com> wrote: >>>> >>>>> There is one thing that I am confused about. >>>>> Spark has codes that have been implemented in Scala. Now, can we run >>>>> any Scala code on the Spark framework? What will be the difference in the >>>>> execution of the scala code in normal systems and on Spark? >>>>> The reason for my question is the following: >>>>> I had a variable >>>>> *val temp = <some operations>* >>>>> This temp was being created inside the loop, so as to manually throw >>>>> it out of the cache, every time the loop ends I was calling >>>>> *temp.unpersist()*, this was returning an error saying that *value >>>>> unpersist is not a method of Int*, which means that temp is an Int. >>>>> Can some one explain to me why I was not able to call *unpersist* on >>>>> *temp*? >>>>> >>>>> Thank You >>>>> >>>> >>>> >>> >> >