[ https://issues.apache.org/jira/browse/SPARK-7819?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14572255#comment-14572255 ]
Fi commented on SPARK-7819: --------------------------- Actually, maybe the InvalidClassCastException might be a little too flaky. I am running a spark job which queries an ORC table daily partition via the HiveContext. A single context is created, and I spin up about ten threads (one for each day). I am seeing plenty of the same errors in various tasks, enough to kill the job for that day. Curiously, the exact same serialVersionUID are logged in each failure: org.apache.spark.sql.hive.MetastoreRelation; local class incompatible: stream classdesc serialVersionUID = 2590680563934099718, local class serialVersionUID = -8650941563091306200 So the interesting thing is that some jobs (for a particular day) work perfectly fine, but others fail. I tried running this multi-threaded job again, and the same error occurs, but in different places. This job works perfectly fine on our Spark 1.3 builds, unfortunately, this issue is occurring too often in a larger job in Spark 1.4 :( > Isolated Hive Client Loader appears to cause Native Library > libMapRClient.4.0.2-mapr.so already loaded in another classloader error > ----------------------------------------------------------------------------------------------------------------------------------- > > Key: SPARK-7819 > URL: https://issues.apache.org/jira/browse/SPARK-7819 > Project: Spark > Issue Type: Bug > Components: SQL > Affects Versions: 1.4.0 > Reporter: Fi > Priority: Critical > Attachments: invalidClassException.log, stacktrace.txt, test.py > > > In reference to the pull request: https://github.com/apache/spark/pull/5876 > I have been running the Spark 1.3 branch for some time with no major hiccups, > and recently switched to the Spark 1.4 branch. > I build my spark distribution with the following build command: > {noformat} > make-distribution.sh --tgz --skip-java-test --with-tachyon -Phive > -Phive-0.13.1 -Pmapr4 -Pspark-ganglia-lgpl -Pkinesis-asl -Phive-thriftserver > {noformat} > When running a python script containing a series of smoke tests I use to > validate the build, I encountered an error under the following conditions: > * start a spark context > * start a hive context > * run any hive query > * stop the spark context > * start a second spark context > * run any hive query > ** ERROR > From what I can tell, the Isolated Class Loader is hitting a MapR class that > is loading its native library (presumedly as part of a static initializer). > Unfortunately, the JVM prohibits this the second time around. > I would think that shutting down the SparkContext would clear out any > vestigials of the JVM, so I'm surprised that this would even be a problem. > Note: all other smoke tests we are running passes fine. > I will attach the stacktrace and a python script reproducing the issue (at > least for my environment and build). -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org