I don't think this is the same issue as it works just fine in pyspark
v1.3.1.
Are you aware of any workaround? I was hoping to start testing one of my
apps in Spark 1.4 and I use the CSV exports as a safety valve to easily
debug my data flow.
-Don
On Sun, Jun 14, 2015 at 7:18 PM, Burak Yavuz
I looked at this again, and when I use the Scala spark-shell and load a CSV
using the same package it works just fine, so this seems specific to
pyspark.
I've created the following JIRA:
https://issues.apache.org/jira/browse/SPARK-8365
-Don
On Sat, Jun 13, 2015 at 11:46 AM, Don Drake
Hi Don,
This seems related to a known issue, where the classpath on the driver is
missing the related classes. This is a bug in py4j as py4j uses the System
Classloader rather than Spark's Context Classloader. However, this problem
existed in 1.3.0 as well, therefore I'm curious whether it's the
I downloaded the pre-compiled Spark 1.4.0 and attempted to run an existing
Python Spark application against it and got the following error:
py4j.protocol.Py4JJavaError: An error occurred while calling o90.save.
: java.lang.RuntimeException: Failed to load class for data source: