Have you set spark.local.dir (I think this is the config setting)?

It needs to point to a volume with plenty of space.

By default if I recall it point to /tmp

Sent from my iPhone

> On 19 Sep 2014, at 23:35, "jw.cmu" <jinliangw...@gmail.com> wrote:
> 
> I'm trying to run Spark ALS using the netflix dataset but failed due to "No
> space on device" exception. It seems the exception is thrown after the
> training phase. It's not clear to me what is being written and where is the
> output directory.
> 
> I was able to run the same code on the provided test.data dataset.
> 
> I'm new to Spark and I'd like to get some hints for resolving this problem.
> 
> The code I ran was got from
> https://spark.apache.org/docs/latest/mllib-collaborative-filtering.html (the
> Java version).
> 
> Relevant info:
> 
> Spark version: 1.0.2 (Standalone deployment)
> # slaves/workers/exectuors: 8
> Core per worker: 64
> memory per executor: 100g
> 
> Application parameters are left as default.
> 
> 
> 
> 
> 
> 
> 
> --
> View this message in context: 
> http://apache-spark-user-list.1001560.n3.nabble.com/Failed-running-Spark-ALS-tp14704.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.
> 
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> For additional commands, e-mail: user-h...@spark.apache.org
> 

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org

Reply via email to