No, Scala primitives remain primitives. Unless you create an RDD using one of the many methods - you would not be able to access any of the RDD methods. There is no automatic porting. Spark is an application as far as scala is concerned - there is no compilation (except of course, the scala, JIT compilation etc).
On Fri, Sep 12, 2014 at 8:04 PM, Deep Pradhan <pradhandeep1...@gmail.com> wrote: > I know that unpersist is a method on RDD. > But my confusion is that, when we port our Scala programs to Spark, > doesn't everything change to RDDs? > > On Fri, Sep 12, 2014 at 10:16 PM, Nicholas Chammas < > nicholas.cham...@gmail.com> wrote: > >> unpersist is a method on RDDs. RDDs are abstractions introduced by Spark. >> >> An Int is just a Scala Int. You can't call unpersist on Int in Scala, and >> that doesn't change in Spark. >> >> On Fri, Sep 12, 2014 at 12:33 PM, Deep Pradhan <pradhandeep1...@gmail.com >> > wrote: >> >>> There is one thing that I am confused about. >>> Spark has codes that have been implemented in Scala. Now, can we run any >>> Scala code on the Spark framework? What will be the difference in the >>> execution of the scala code in normal systems and on Spark? >>> The reason for my question is the following: >>> I had a variable >>> *val temp = <some operations>* >>> This temp was being created inside the loop, so as to manually throw it >>> out of the cache, every time the loop ends I was calling >>> *temp.unpersist()*, this was returning an error saying that *value >>> unpersist is not a method of Int*, which means that temp is an Int. >>> Can some one explain to me why I was not able to call *unpersist* on >>> *temp*? >>> >>> Thank You >>> >> >> >