Hi Sebastian, That exception generally means you have the class loaded by two different class loaders, and some code is trying to mix instances created by the two different loaded classes.
Do you happen to have that class both in the spark jars and in your app's uber-jar? That might explain the problem, although I'm not terribly familiar with Spark's class loader hierarchy. On Thu, May 29, 2014 at 5:51 AM, Sebastian Schelter <s...@apache.org> wrote: > Hi, > > I have trouble running some custom code on Spark 0.9.1 in standalone mode on > a cluster. I built a fat jar (excluding Spark) that I'm adding to the > classpath with ADD_JARS=... When I start the Spark shell, I can instantiate > classes, but when I run Spark code, I get strange ClassCastExceptions like > this: > > 14/05/29 14:48:10 INFO TaskSetManager: Loss was due to > java.lang.ClassCastException: io.ssc.sampling.matrix.DenseBlock cannot be > cast to io.ssc.sampling.matrix.DenseBlock [duplicate 1] > > What am I doing wrong? > > Thx, > Sebastian -- Marcelo