I guess it's a ClassLoader issue. But I have no idea how to debug it. Any hints?
Jianshi On Fri, Mar 13, 2015 at 3:00 PM, Eric Charles <e...@apache.org> wrote: > i have the same issue running spark sql code from eclipse workspace. If > you run your code from the command line (with a packaged jar) or from > Intellij, I bet it should work. > > IMHO This is some how related to eclipse env, but would love to know how > to fix it (whether via eclipse conf, or via a patch in spark). > > > > On 03/01/2015 02:32 AM, Michael Armbrust wrote: > > I think its possible that the problem is that the scala compiler is not > being loaded by the primordial classloader (but instead by some child > classloader) and thus the scala reflection mirror is failing to initialize > when it can't find it. Unfortunately, the only solution that I know of is > to load all required jars when the JVM starts. > > On Sat, Feb 28, 2015 at 5:26 PM, Ashish Nigam <ashnigamt...@gmail.com> > wrote: > >> Also, can scala version play any role here? >> I am using scala 2.11.5 but all spark packages have dependency to scala >> 2.11.2 >> Just wanted to make sure that scala version is not an issue here. >> >> On Sat, Feb 28, 2015 at 9:18 AM, Ashish Nigam <ashnigamt...@gmail.com> >> wrote: >> >>> Hi, >>> I wrote a very simple program in scala to convert an existing RDD to >>> SchemaRDD. >>> But createSchemaRDD function is throwing exception >>> >>> Exception in thread "main" scala.ScalaReflectionException: class >>> org.apache.spark.sql.catalyst.ScalaReflection in JavaMirror with primordial >>> classloader with boot classpath [.....] not found >>> >>> >>> Here's more info on the versions I am using - >>> >>> <scala.binary.version>2.11</scala.binary.version> >>> <spark.version>1.2.1</spark.version> >>> <scala.version>2.11.5</scala.version> >>> >>> Please let me know how can I resolve this problem. >>> >>> Thanks >>> Ashish >>> >> >> > -- Jianshi Huang LinkedIn: jianshi Twitter: @jshuang Github & Blog: http://huangjs.github.com/