I'm almost certain the problem is the ClassLoader. So adding
fork := true solves problems for test and run. The problem is how can I fork a JVM for sbt console? fork in console := true seems not working... Jianshi On Fri, Mar 13, 2015 at 4:35 PM, Jianshi Huang <jianshi.hu...@gmail.com> wrote: > I guess it's a ClassLoader issue. But I have no idea how to debug it. Any > hints? > > Jianshi > > On Fri, Mar 13, 2015 at 3:00 PM, Eric Charles <e...@apache.org> wrote: > >> i have the same issue running spark sql code from eclipse workspace. If >> you run your code from the command line (with a packaged jar) or from >> Intellij, I bet it should work. >> >> IMHO This is some how related to eclipse env, but would love to know how >> to fix it (whether via eclipse conf, or via a patch in spark). >> >> >> >> On 03/01/2015 02:32 AM, Michael Armbrust wrote: >> >> I think its possible that the problem is that the scala compiler is not >> being loaded by the primordial classloader (but instead by some child >> classloader) and thus the scala reflection mirror is failing to initialize >> when it can't find it. Unfortunately, the only solution that I know of is >> to load all required jars when the JVM starts. >> >> On Sat, Feb 28, 2015 at 5:26 PM, Ashish Nigam <ashnigamt...@gmail.com> >> wrote: >> >>> Also, can scala version play any role here? >>> I am using scala 2.11.5 but all spark packages have dependency to scala >>> 2.11.2 >>> Just wanted to make sure that scala version is not an issue here. >>> >>> On Sat, Feb 28, 2015 at 9:18 AM, Ashish Nigam <ashnigamt...@gmail.com> >>> wrote: >>> >>>> Hi, >>>> I wrote a very simple program in scala to convert an existing RDD to >>>> SchemaRDD. >>>> But createSchemaRDD function is throwing exception >>>> >>>> Exception in thread "main" scala.ScalaReflectionException: class >>>> org.apache.spark.sql.catalyst.ScalaReflection in JavaMirror with primordial >>>> classloader with boot classpath [.....] not found >>>> >>>> >>>> Here's more info on the versions I am using - >>>> >>>> <scala.binary.version>2.11</scala.binary.version> >>>> <spark.version>1.2.1</spark.version> >>>> <scala.version>2.11.5</scala.version> >>>> >>>> Please let me know how can I resolve this problem. >>>> >>>> Thanks >>>> Ashish >>>> >>> >>> >> > > > -- > Jianshi Huang > > LinkedIn: jianshi > Twitter: @jshuang > Github & Blog: http://huangjs.github.com/ > -- Jianshi Huang LinkedIn: jianshi Twitter: @jshuang Github & Blog: http://huangjs.github.com/