Since it’s from Scala, it might mean you’re running with a different version of 
Scala than you compiled Spark with. Spark 0.8 and earlier use Scala 2.9, while 
Spark 0.9 uses Scala 2.10.

Matei

On Mar 11, 2014, at 8:19 AM, Jeyaraj, Arockia R (Arockia) 
<arockia.r.jeya...@verizon.com> wrote:

> Hi,
>  
> Can anyone help me to resolve this issue? Why am I getting NoSuchMethod 
> exception?
>  
> 14/03/11 09:56:11 ERROR executor.Executor: Exception in task ID 0
> java.lang.NoSuchMethodError: 
> scala.Predef$.augmentString(Ljava/lang/String;)Lsca
> la/collection/immutable/StringOps;
> at kafka.utils.VerifiableProperties.getIntInRange(VerifiableProperties.s
> cala:75)
> at kafka.utils.VerifiableProperties.getInt(VerifiableProperties.scala:58
> )
> at kafka.utils.ZKConfig.<init>(ZkUtils.scala:837)
> at kafka.consumer.ConsumerConfig.<init>(ConsumerConfig.scala:73)
> at kafka.consumer.ConsumerConfig.<init>(ConsumerConfig.scala:77)
> at org.apache.spark.streaming.kafka.KafkaReceiver.onStart(KafkaInputDStr
> eam.scala:98)
> at org.apache.spark.streaming.dstream.NetworkReceiver.start(NetworkInput
> DStream.scala:126)
> at org.apache.spark.streaming.scheduler.NetworkInputTracker$ReceiverExec
> utor$$anonfun$8.apply(NetworkInputTracker.scala:173)
> at org.apache.spark.streaming.scheduler.NetworkInputTracker$ReceiverExec
> utor$$anonfun$8.apply(NetworkInputTracker.scala:169)
> at org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.sc
> ala:884)
> at org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.sc
> ala:884)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:109)
> at org.apache.spark.scheduler.Task.run(Task.scala:53)
> at org.apache.spark.executor.Executor$TaskRunner$$anonfun$run$1.apply$mc
> V$sp(Executor.scala:213)
> at org.apache.spark.deploy.SparkHadoopUtil.runAsUser(SparkHadoopUtil.sca
> la:49)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:178)
> 
> at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExec
> utor.java:886)
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor
> .java:908)
> at java.lang.Thread.run(Thread.java:619)
> 14/03/11 09:56:11 WARN scheduler.TaskSetManager: Lost TID 0 (task 0.0:0)
> 14/03/11 09:56:11 WARN scheduler.TaskSetManager: Loss was due to 
> java.lang.NoSuc
> hMethodError
> java.lang.NoSuchMethodError: 
> scala.Predef$.augmentString(Ljava/lang/String;)Lsca
> la/collection/immutable/StringOps;
> at kafka.utils.VerifiableProperties.getIntInRange(VerifiableProperties.s
> cala:75)
> at kafka.utils.VerifiableProperties.getInt(VerifiableProperties.scala:58
> )
> at kafka.utils.ZKConfig.<init>(ZkUtils.scala:837)
> at kafka.consumer.ConsumerConfig.<init>(ConsumerConfig.scala:73)
> at kafka.consumer.ConsumerConfig.<init>(ConsumerConfig.scala:77)
> at org.apache.spark.streaming.kafka.KafkaReceiver.onStart(KafkaInputDStr
> eam.scala:98)
> at org.apache.spark.streaming.dstream.NetworkReceiver.start(NetworkInput
> DStream.scala:126)
> at org.apache.spark.streaming.scheduler.NetworkInputTracker$ReceiverExec
> utor$$anonfun$8.apply(NetworkInputTracker.scala:173)
> at org.apache.spark.streaming.scheduler.NetworkInputTracker$ReceiverExec
> utor$$anonfun$8.apply(NetworkInputTracker.scala:169)
> at org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.sc
> ala:884)
> at org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.sc
> ala:884)
> at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:109)
> at org.apache.spark.scheduler.Task.run(Task.scala:53)
> at org.apache.spark.executor.Executor$TaskRunner$$anonfun$run$1.apply$mc
> V$sp(Executor.scala:213)
> at org.apache.spark.deploy.SparkHadoopUtil.runAsUser(SparkHadoopUtil.sca
> la:49)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:178)
> 
> at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExec
> utor.java:886)
> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor
> .java:908)
> at java.lang.Thread.run(Thread.java:619)
> 14/03/11 09:56:11 ERROR scheduler.TaskSetManager: Task 0.0:0 failed 1 times; 
> abo
> rting job
> 14/03/11 09:56:11 INFO scheduler.TaskSchedulerImpl: Remove TaskSet 0.0 from 
> pool
> 
> 14/03/11 09:56:11 INFO scheduler.DAGScheduler: Failed to run runJob at 
> NetworkIn
> putTracker.scala:182
> [error] (Thread-34) org.apache.spark.SparkException: Job aborted: Task 0.0:0 
> fai
> 
>  
>  
> Thanks
> Arockia Raja

Reply via email to