I guess it's a ClassLoader issue. But I have no idea how to debug it. Any
hints?

Jianshi

On Fri, Mar 13, 2015 at 3:00 PM, Eric Charles <e...@apache.org> wrote:

>  i have the same issue running spark sql code from eclipse workspace. If
> you run your code from the command line (with a packaged jar) or from
> Intellij, I bet it should work.
>
> IMHO This is some how related to eclipse env, but would love to know how
> to fix it (whether via eclipse conf, or via a patch in spark).
>
>
>
> On 03/01/2015 02:32 AM, Michael Armbrust wrote:
>
> I think its possible that the problem is that the scala compiler is not
> being loaded by the primordial classloader (but instead by some child
> classloader) and thus the scala reflection mirror is failing to initialize
> when it can't find it. Unfortunately, the only solution that I know of is
> to load all required jars when the JVM starts.
>
> On Sat, Feb 28, 2015 at 5:26 PM, Ashish Nigam <ashnigamt...@gmail.com>
> wrote:
>
>>  Also, can scala version play any role here?
>>  I am using scala 2.11.5 but all spark packages have dependency to scala
>> 2.11.2
>>  Just wanted to make sure that scala version is not an issue here.
>>
>> On Sat, Feb 28, 2015 at 9:18 AM, Ashish Nigam <ashnigamt...@gmail.com>
>> wrote:
>>
>>>    Hi,
>>>  I wrote a very simple program in scala to convert an existing RDD to
>>> SchemaRDD.
>>>  But createSchemaRDD function is throwing exception
>>>
>>> Exception in thread "main" scala.ScalaReflectionException: class
>>> org.apache.spark.sql.catalyst.ScalaReflection in JavaMirror with primordial
>>> classloader with boot classpath [.....] not found
>>>
>>>
>>>  Here's more info on the versions I am using -
>>>
>>> <scala.binary.version>2.11</scala.binary.version>
>>>     <spark.version>1.2.1</spark.version>
>>>     <scala.version>2.11.5</scala.version>
>>>
>>>  Please let me know how can I resolve this problem.
>>>
>>>  Thanks
>>>  Ashish
>>>
>>
>>
>


-- 
Jianshi Huang

LinkedIn: jianshi
Twitter: @jshuang
Github & Blog: http://huangjs.github.com/

Reply via email to