ahhh never mind… I didn’t notice that a spark-assembly jar file gets produced 
after compiling the whole spark suite… So no more manual editing of the jar 
file of the AMI for now!

Christos 


On Oct 10, 2014, at 12:15 AM, Christos Kozanitis wrote:

> Hi
> 
> I have written a few extensions for sparkSQL (for version 1.1.0) and I am 
> trying to deploy my new jar files (one for catalyst and one for sql/core) on 
> ec2.
> 
> My approach was to create a new 
> spark/lib/spark-assembly-1.1.0-hadoop1.0.4.jar that merged the contents of 
> the old one with the contents of my new jar files and I propagated the 
> changes to workers. 
> 
> However when I tried the code snippet below I received the error message that 
> I paste at the end of this email. I was wondering, do you guys have any 
> suggestions on how to fix this?
> 
> thanks
> Christos
> 
> the code is:
> 
> import org.apache.spark.{SparkContext, SparkConf}
> import org.apache.spark.sql.SQLContext
> val sqlContext = new SQLContext(sc)
> 
> the error message is:
> 
> error: bad symbolic reference. A signature in package.class refers to term 
> scalalogging
> in package com.typesafe which is not available.
> It may be completely missing from the current classpath, or the version on
> the classpath might be incompatible with the version used when compiling 
> package.class.
> <console>:14: error: bad symbolic reference. A signature in package.class 
> refers to term slf4j
> in value com.typesafe.scalalogging which is not available.
> It may be completely missing from the current classpath, or the version on
> the classpath might be incompatible with the version used when compiling 
> package.class.
>        val sqlContext = new SQLContext(sc)
> 

Reply via email to