It turned out to be a problem with `SerializationUtils` from Apache Commons
Lang. There is an open issue where the class will throw a
`ClassNotFoundException` even if the class is in the classpath in a
multiple-classloader environment:
https://issues.apache.org/jira/browse/LANG-1049
We moved away
You also need to have library hadoop-aws in your classpath. From Hadoop 2.6,
the AWS libraries come in that separate library.
Also, you will need this line in your hadoop configuration:
hadoopConf.set("fs.s3.impl",
"org.apache.hadoop.fs.s3native.NativeS3FileSystem")
--
View this message in
Hello,
I am facing an issue where I cannot run my Spark job in a cluster
environment (standalone or EMR) but it works successfully if I run it
locally using local[*] as master.
I am getting ClassNotFoundException: com.mycompany.folder.MyObject on the
slave executors. I don't really understand