Thank you very much for the reply. 
I tried different ways and when I tried setting up the root pom.xml to 2.11 

                <scala.version>2.11.6</scala.version>
                <scala.binary.version>2.11</scala.binary.version>

I got the following error
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal on project flink-scala: Could not resolve depende
ncies for project org.apache.flink:flink-scala:jar:1.0-SNAPSHOT: Could not find
artifact org.scalamacros:quasiquotes_2.11:jar:2.0.1 in central (http://repo.mave
n.apache.org/maven2) -> [Help 1]

If I leave the scala.binary.verson to be at 2.10 and the scala version to be at 
2.11.6 then I get the following problem
[INFO] C:\flink\flink\flink-runtime\src\test\scala:-1: info: compiling
[INFO] Compiling 366 source files to C:\flink\flink\flink-runtime\target\test-cl
asses at 1452508064750
[ERROR] C:\flink\flink\flink-runtime\src\test\scala\org\apache\flink\runtime\job
manager\JobManagerITCase.scala:700: error: can't expand macros compiled by previ
ous versions of Scala
[ERROR]               assert(cachedGraph2.isArchived)
[ERROR]                                   ^

So am not pretty sure how to proceed with this. If I try to change the version 
of scala to 2.10 in the IDE then I get lot of compilation issues.  IS there any 
way to over come this?

Once again thanks a lot and apologies for the naïve question.

Regards
Ram
-----Original Message-----
From: ewenstep...@gmail.com [mailto:ewenstep...@gmail.com] On Behalf Of Stephan 
Ewen
Sent: Friday, January 8, 2016 5:01 PM
To: dev@flink.apache.org
Subject: Re: Naive question

Hi!

This looks like a mismatch between the Scala dependency in Flink and Scala in 
your Eclipse. Make sure you use the same for both. By default, Flink reference 
Scala 2.10

If your IDE is set up for Scala 2.11, set the Scala version variable in the 
Flink root pom.xml also to 2.11

Greetings,
Stephan




On Fri, Jan 8, 2016 at 12:06 PM, Vasudevan, Ramkrishna S < 
ramkrishna.s.vasude...@intel.com> wrote:

> I have been trying to install, learn and understand Flink. I am using
> Scala- EclipseIDE as my IDE.
>
> I have downloaded the flink source coded, compiled and created the project.
>
> My work laptop is Windows based and I don't have eclipse based 
> workstation but I do have linux boxes for running and testing things.
>
> Some of the examples given in Flink source code do run directly from 
> Eclipse but when I try to run the Wordcount example from Eclipse I get 
> this error
>
> Exception in thread "main" java.lang.NoSuchMethodError:
> scala.collection.immutable.HashSet$.empty()Lscala/collection/immutable/HashSet;
>      at akka.actor.ActorCell$.<init>(ActorCell.scala:336)
>      at akka.actor.ActorCell$.<clinit>(ActorCell.scala)
>      at akka.actor.RootActorPath.$div(ActorPath.scala:159)
>      at akka.actor.LocalActorRefProvider.<init>(ActorRefProvider.scala:464)
>      at akka.actor.LocalActorRefProvider.<init>(ActorRefProvider.scala:452)
>      at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
> Method)
>      at sun.reflect.NativeConstructorAccessorImpl.newInstance(Unknown
> Source)
>      at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(Unknown
> Source)
>      at java.lang.reflect.Constructor.newInstance(Unknown Source)
>      at
> akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$2.apply(DynamicAccess.scala:78)
>      at scala.util.Try$.apply(Try.scala:191)
>      at
> akka.actor.ReflectiveDynamicAccess.createInstanceFor(DynamicAccess.scala:73)
>      at
> akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$3.apply(DynamicAccess.scala:84)
>      at
> akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$3.apply(DynamicAccess.scala:84)
>      at scala.util.Success.flatMap(Try.scala:230)
>      at
> akka.actor.ReflectiveDynamicAccess.createInstanceFor(DynamicAccess.scala:84)
>      at akka.actor.ActorSystemImpl.liftedTree1$1(ActorSystem.scala:585)
>      at akka.actor.ActorSystemImpl.<init>(ActorSystem.scala:578)
>      at akka.actor.ActorSystem$.apply(ActorSystem.scala:142)
>      at akka.actor.ActorSystem$.apply(ActorSystem.scala:119)
>      at akka.actor.ActorSystem$.create(ActorSystem.scala:67)
>      at
> org.apache.flink.runtime.akka.AkkaUtils$.createActorSystem(AkkaUtils.scala:84)
>      at
> org.apache.flink.runtime.minicluster.FlinkMiniCluster.startJobManagerActorSystem(FlinkMiniCluster.scala:196)
>      at
> org.apache.flink.runtime.minicluster.FlinkMiniCluster.singleActorSystem$lzycompute$1(FlinkMiniCluster.scala:225)
>      at org.apache.flink.runtime.minicluster.FlinkMiniCluster.org
> $apache$flink$runtime$minicluster$FlinkMiniCluster$$singleActorSystem$1(FlinkMiniCluster.scala:225)
>      at
> org.apache.flink.runtime.minicluster.FlinkMiniCluster$$anonfun$1.apply(FlinkMiniCluster.scala:230)
>      at
> org.apache.flink.runtime.minicluster.FlinkMiniCluster$$anonfun$1.apply(FlinkMiniCluster.scala:228)
>      at
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:245)
>      at
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:245)
>      at scala.collection.immutable.Range.foreach(Range.scala:166)
>      at
> scala.collection.TraversableLike$class.map(TraversableLike.scala:245)
>      at scala.collection.AbstractTraversable.map(Traversable.scala:104)
>      at
> org.apache.flink.runtime.minicluster.FlinkMiniCluster.start(FlinkMiniCluster.scala:228)
>      at
> org.apache.flink.runtime.minicluster.FlinkMiniCluster.start(FlinkMiniCluster.scala:219)
>      at
> org.apache.flink.streaming.api.environment.LocalStreamEnvironment.execute(LocalStreamEnvironment.java:104)
>      at
> org.apache.flink.streaming.examples.wordcount.WordCount.main(WordCount
> .java:80)
>
> I know this is a naïve question but I would like to get some help in 
> order to over come this issue. I tried various options like setting 
> scala-2.10 as the compiler for the project (then it shows completely 
> different error) and many of the projects don't even compile. But with 
> 2.11 version I get the above stack trace. Any help here is welcome.
>
> Regards
> Ram
>

Reply via email to