Thanks! Setting java.io.tmpdir did the trick. Sadly, I still ran into an
issue with the amount of RAM pyspark was grabbing. In fact I got a message
from my web provider warning that I was exceeding the memory limit for my
(entry level) account. So I won't be pursuing it farther. Oh well, it
I'm trying to get pyspark running on a shared web host. I can get into the
pyspark shell but whenever I run a simple command like
sc.parallelize([1,2,3,4]).sum() I get an error that seems to stem from some
kind of permission issue with libsnappyjava.so:
Caused by: java.lang.UnsatisfiedLinkError:
This is due to the snappy-java library; I think that you'll have to
configure either java.io.tmpdir or org.xerial.snappy.tempdir; see
https://github.com/xerial/snappy-java/blob/1198363176ad671d933fdaf0938b8b9e609c0d8a/src/main/java/org/xerial/snappy/SnappyLoader.java#L335
On Mon, Jan 11, 2016