Hi, 

Do you mean you are running the script with 
https://github.com/amplab-extras/SparkR-pkg and spark 1.2? I am afraid that 
currently there is no development effort and support on  the SparkR-pkg since 
it has been integrated into Spark since Spark 1.4.

Unfortunately, the RDD API and RDD-like API of DataFrame of SparkR is not 
exposed in Spark 1.4 for some considerations. Although not exposed, some 
RDD-like API of DataFrame are actually implemented which you can find in the 
SparkR source code, including 
lapply/lapplyPartition/flatMap/foreach/foreachPartition. Though not 
recommended, but if you really want to use them, you can use SparkR::: to 
access them as a temporary workaround.

There is on-going investigation and discussion on whether to expose a subset of 
RDD API or not, you can refer to 
https://issues.apache.org/jira/browse/SPARK-7264 if you are interested.

-----Original Message-----
From: Jennifer15 [mailto:bsabe...@purdue.edu] 
Sent: Monday, July 27, 2015 1:47 PM
To: user@spark.apache.org
Subject: unserialize error in sparkR

Hi,
I have a newbie question; I get the following error by increasing the number of 
samples in my sample script  samplescript.R 
<http://apache-spark-user-list.1001560.n3.nabble.com/file/n24002/samplescript.R>
, which is written in Spark1.2 (no error for small sample of error):

Error in unserialize(obj) : 
ReadItem: unknown type 0, perhaps written by later version of R
Calls: assetForecast ... convertJListToRList -> lapply -> lapply -> FUN   ->
unserialize
Execution halted

I tried using Spark1.4 though I could not find lapply or any similar functions 
for dataframes.
I am not sure if this error is because of using spark1.2 though if it is, what 
is the equivalent of lapply/map to work on dataframes?




--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/unserialize-error-in-sparkR-tp24002.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional 
commands, e-mail: user-h...@spark.apache.org


---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org

Reply via email to