Thanks to all. I tried with Scala Eclipse IDE with all these 
'change-scala-version.sh'. But in vain. 

So I switched over to Intellij and thing work fine over there. I am new to 
Intellij so will try using it.  

Once again thanks for helping me out.

Regards
Ram

-----Original Message-----
From: Chiwan Park [mailto:chiwanp...@apache.org] 
Sent: Monday, January 11, 2016 4:37 PM
To: dev@flink.apache.org
Subject: Re: Naive question

Hi Ram,

If you want to build Flink with Scala 2.10, just checkout Flink repository from 
github or download source code from homepage, run `mvn clean install 
-DskipTests` and import projects to your IDE. If you want to build Flink with 
Scala 2.11, you have to run `tools/change-scala-version.sh 2.11` before build 
the project. You can revert Scala version change by running 
`tools/change-scala-version.sh 2.10`.

About IDE, Flink community recommends IntelliJ IDEA because Scala IDE have some 
problems in Java/Scala mixed project like Flink. But I tested importing Flink 
project with Scala IDE 4.3.0, Scala 2.11.7 and Flink 0.10.0 source code. Note 
that you should import the project as maven project.

By the way, the community welcomes any questions. Please feel free to post 
questions. :)

> On Jan 11, 2016, at 7:30 PM, Vasudevan, Ramkrishna S 
> <ramkrishna.s.vasude...@intel.com> wrote:
> 
> Thank you very much for the reply. 
> I tried different ways and when I tried setting up the root pom.xml to 
> 2.11
> 
>               <scala.version>2.11.6</scala.version>
>               <scala.binary.version>2.11</scala.binary.version>
> 
> I got the following error
> [INFO] 
> ----------------------------------------------------------------------
> -- [ERROR] Failed to execute goal on project flink-scala: Could not 
> resolve depende ncies for project 
> org.apache.flink:flink-scala:jar:1.0-SNAPSHOT: Could not find artifact 
> org.scalamacros:quasiquotes_2.11:jar:2.0.1 in central 
> (http://repo.mave
> n.apache.org/maven2) -> [Help 1]
> 
> If I leave the scala.binary.verson to be at 2.10 and the scala version 
> to be at 2.11.6 then I get the following problem [INFO] 
> C:\flink\flink\flink-runtime\src\test\scala:-1: info: compiling [INFO] 
> Compiling 366 source files to 
> C:\flink\flink\flink-runtime\target\test-cl
> asses at 1452508064750
> [ERROR] 
> C:\flink\flink\flink-runtime\src\test\scala\org\apache\flink\runtime\j
> ob
> manager\JobManagerITCase.scala:700: error: can't expand macros 
> compiled by previ ous versions of Scala
> [ERROR]               assert(cachedGraph2.isArchived)
> [ERROR]                                   ^
> 
> So am not pretty sure how to proceed with this. If I try to change the 
> version of scala to 2.10 in the IDE then I get lot of compilation issues.  IS 
> there any way to over come this?
> 
> Once again thanks a lot and apologies for the naïve question.
> 
> Regards
> Ram
> -----Original Message-----
> From: ewenstep...@gmail.com [mailto:ewenstep...@gmail.com] On Behalf 
> Of Stephan Ewen
> Sent: Friday, January 8, 2016 5:01 PM
> To: dev@flink.apache.org
> Subject: Re: Naive question
> 
> Hi!
> 
> This looks like a mismatch between the Scala dependency in Flink and 
> Scala in your Eclipse. Make sure you use the same for both. By 
> default, Flink reference Scala 2.10
> 
> If your IDE is set up for Scala 2.11, set the Scala version variable 
> in the Flink root pom.xml also to 2.11
> 
> Greetings,
> Stephan
> 
> 
> 
> 
> On Fri, Jan 8, 2016 at 12:06 PM, Vasudevan, Ramkrishna S < 
> ramkrishna.s.vasude...@intel.com> wrote:
> 
>> I have been trying to install, learn and understand Flink. I am using
>> Scala- EclipseIDE as my IDE.
>> 
>> I have downloaded the flink source coded, compiled and created the project.
>> 
>> My work laptop is Windows based and I don't have eclipse based 
>> workstation but I do have linux boxes for running and testing things.
>> 
>> Some of the examples given in Flink source code do run directly from 
>> Eclipse but when I try to run the Wordcount example from Eclipse I 
>> get this error
>> 
>> Exception in thread "main" java.lang.NoSuchMethodError:
>> scala.collection.immutable.HashSet$.empty()Lscala/collection/immutable/HashSet;
>>     at akka.actor.ActorCell$.<init>(ActorCell.scala:336)
>>     at akka.actor.ActorCell$.<clinit>(ActorCell.scala)
>>     at akka.actor.RootActorPath.$div(ActorPath.scala:159)
>>     at akka.actor.LocalActorRefProvider.<init>(ActorRefProvider.scala:464)
>>     at akka.actor.LocalActorRefProvider.<init>(ActorRefProvider.scala:452)
>>     at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
>> Method)
>>     at sun.reflect.NativeConstructorAccessorImpl.newInstance(Unknown
>> Source)
>>     at
>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(Unknown
>> Source)
>>     at java.lang.reflect.Constructor.newInstance(Unknown Source)
>>     at
>> akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$2.apply(DynamicAccess.scala:78)
>>     at scala.util.Try$.apply(Try.scala:191)
>>     at
>> akka.actor.ReflectiveDynamicAccess.createInstanceFor(DynamicAccess.scala:73)
>>     at
>> akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$3.apply(DynamicAccess.scala:84)
>>     at
>> akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$3.apply(DynamicAccess.scala:84)
>>     at scala.util.Success.flatMap(Try.scala:230)
>>     at
>> akka.actor.ReflectiveDynamicAccess.createInstanceFor(DynamicAccess.scala:84)
>>     at akka.actor.ActorSystemImpl.liftedTree1$1(ActorSystem.scala:585)
>>     at akka.actor.ActorSystemImpl.<init>(ActorSystem.scala:578)
>>     at akka.actor.ActorSystem$.apply(ActorSystem.scala:142)
>>     at akka.actor.ActorSystem$.apply(ActorSystem.scala:119)
>>     at akka.actor.ActorSystem$.create(ActorSystem.scala:67)
>>     at
>> org.apache.flink.runtime.akka.AkkaUtils$.createActorSystem(AkkaUtils.scala:84)
>>     at
>> org.apache.flink.runtime.minicluster.FlinkMiniCluster.startJobManagerActorSystem(FlinkMiniCluster.scala:196)
>>     at
>> org.apache.flink.runtime.minicluster.FlinkMiniCluster.singleActorSystem$lzycompute$1(FlinkMiniCluster.scala:225)
>>     at org.apache.flink.runtime.minicluster.FlinkMiniCluster.org
>> $apache$flink$runtime$minicluster$FlinkMiniCluster$$singleActorSystem$1(FlinkMiniCluster.scala:225)
>>     at
>> org.apache.flink.runtime.minicluster.FlinkMiniCluster$$anonfun$1.apply(FlinkMiniCluster.scala:230)
>>     at
>> org.apache.flink.runtime.minicluster.FlinkMiniCluster$$anonfun$1.apply(FlinkMiniCluster.scala:228)
>>     at
>> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:245)
>>     at
>> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:245)
>>     at scala.collection.immutable.Range.foreach(Range.scala:166)
>>     at
>> scala.collection.TraversableLike$class.map(TraversableLike.scala:245)
>>     at scala.collection.AbstractTraversable.map(Traversable.scala:104)
>>     at
>> org.apache.flink.runtime.minicluster.FlinkMiniCluster.start(FlinkMiniCluster.scala:228)
>>     at
>> org.apache.flink.runtime.minicluster.FlinkMiniCluster.start(FlinkMiniCluster.scala:219)
>>     at
>> org.apache.flink.streaming.api.environment.LocalStreamEnvironment.execute(LocalStreamEnvironment.java:104)
>>     at
>> org.apache.flink.streaming.examples.wordcount.WordCount.main(WordCoun
>> t
>> .java:80)
>> 
>> I know this is a naïve question but I would like to get some help in 
>> order to over come this issue. I tried various options like setting
>> scala-2.10 as the compiler for the project (then it shows completely 
>> different error) and many of the projects don't even compile. But 
>> with
>> 2.11 version I get the above stack trace. Any help here is welcome.
>> 
>> Regards
>> Ram
>> 

Regards,
Chiwan Park


Reply via email to