Sorry to hear that it did not work out with Eclipse at all in the end, even
with all adjustments.

Just making sure: You imported Flink as a Maven project, not manually
adding the big Flink dependency JAR?

On Tue, Jan 12, 2016 at 5:15 AM, Vasudevan, Ramkrishna S <
ramkrishna.s.vasude...@intel.com> wrote:

> Thanks to all. I tried with Scala Eclipse IDE with all these
> 'change-scala-version.sh'. But in vain.
>
> So I switched over to Intellij and thing work fine over there. I am new to
> Intellij so will try using it.
>
> Once again thanks for helping me out.
>
> Regards
> Ram
>
> -----Original Message-----
> From: Chiwan Park [mailto:chiwanp...@apache.org]
> Sent: Monday, January 11, 2016 4:37 PM
> To: dev@flink.apache.org
> Subject: Re: Naive question
>
> Hi Ram,
>
> If you want to build Flink with Scala 2.10, just checkout Flink repository
> from github or download source code from homepage, run `mvn clean install
> -DskipTests` and import projects to your IDE. If you want to build Flink
> with Scala 2.11, you have to run `tools/change-scala-version.sh 2.11`
> before build the project. You can revert Scala version change by running
> `tools/change-scala-version.sh 2.10`.
>
> About IDE, Flink community recommends IntelliJ IDEA because Scala IDE have
> some problems in Java/Scala mixed project like Flink. But I tested
> importing Flink project with Scala IDE 4.3.0, Scala 2.11.7 and Flink 0.10.0
> source code. Note that you should import the project as maven project.
>
> By the way, the community welcomes any questions. Please feel free to post
> questions. :)
>
> > On Jan 11, 2016, at 7:30 PM, Vasudevan, Ramkrishna S <
> ramkrishna.s.vasude...@intel.com> wrote:
> >
> > Thank you very much for the reply.
> > I tried different ways and when I tried setting up the root pom.xml to
> > 2.11
> >
> >               <scala.version>2.11.6</scala.version>
> >               <scala.binary.version>2.11</scala.binary.version>
> >
> > I got the following error
> > [INFO]
> > ----------------------------------------------------------------------
> > -- [ERROR] Failed to execute goal on project flink-scala: Could not
> > resolve depende ncies for project
> > org.apache.flink:flink-scala:jar:1.0-SNAPSHOT: Could not find artifact
> > org.scalamacros:quasiquotes_2.11:jar:2.0.1 in central
> > (http://repo.mave
> > n.apache.org/maven2) -> [Help 1]
> >
> > If I leave the scala.binary.verson to be at 2.10 and the scala version
> > to be at 2.11.6 then I get the following problem [INFO]
> > C:\flink\flink\flink-runtime\src\test\scala:-1: info: compiling [INFO]
> > Compiling 366 source files to
> > C:\flink\flink\flink-runtime\target\test-cl
> > asses at 1452508064750
> > [ERROR]
> > C:\flink\flink\flink-runtime\src\test\scala\org\apache\flink\runtime\j
> > ob
> > manager\JobManagerITCase.scala:700: error: can't expand macros
> > compiled by previ ous versions of Scala
> > [ERROR]               assert(cachedGraph2.isArchived)
> > [ERROR]                                   ^
> >
> > So am not pretty sure how to proceed with this. If I try to change the
> version of scala to 2.10 in the IDE then I get lot of compilation issues.
> IS there any way to over come this?
> >
> > Once again thanks a lot and apologies for the naïve question.
> >
> > Regards
> > Ram
> > -----Original Message-----
> > From: ewenstep...@gmail.com [mailto:ewenstep...@gmail.com] On Behalf
> > Of Stephan Ewen
> > Sent: Friday, January 8, 2016 5:01 PM
> > To: dev@flink.apache.org
> > Subject: Re: Naive question
> >
> > Hi!
> >
> > This looks like a mismatch between the Scala dependency in Flink and
> > Scala in your Eclipse. Make sure you use the same for both. By
> > default, Flink reference Scala 2.10
> >
> > If your IDE is set up for Scala 2.11, set the Scala version variable
> > in the Flink root pom.xml also to 2.11
> >
> > Greetings,
> > Stephan
> >
> >
> >
> >
> > On Fri, Jan 8, 2016 at 12:06 PM, Vasudevan, Ramkrishna S <
> ramkrishna.s.vasude...@intel.com> wrote:
> >
> >> I have been trying to install, learn and understand Flink. I am using
> >> Scala- EclipseIDE as my IDE.
> >>
> >> I have downloaded the flink source coded, compiled and created the
> project.
> >>
> >> My work laptop is Windows based and I don't have eclipse based
> >> workstation but I do have linux boxes for running and testing things.
> >>
> >> Some of the examples given in Flink source code do run directly from
> >> Eclipse but when I try to run the Wordcount example from Eclipse I
> >> get this error
> >>
> >> Exception in thread "main" java.lang.NoSuchMethodError:
> >>
> scala.collection.immutable.HashSet$.empty()Lscala/collection/immutable/HashSet;
> >>     at akka.actor.ActorCell$.<init>(ActorCell.scala:336)
> >>     at akka.actor.ActorCell$.<clinit>(ActorCell.scala)
> >>     at akka.actor.RootActorPath.$div(ActorPath.scala:159)
> >>     at
> akka.actor.LocalActorRefProvider.<init>(ActorRefProvider.scala:464)
> >>     at
> akka.actor.LocalActorRefProvider.<init>(ActorRefProvider.scala:452)
> >>     at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
> >> Method)
> >>     at sun.reflect.NativeConstructorAccessorImpl.newInstance(Unknown
> >> Source)
> >>     at
> >> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(Unknown
> >> Source)
> >>     at java.lang.reflect.Constructor.newInstance(Unknown Source)
> >>     at
> >>
> akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$2.apply(DynamicAccess.scala:78)
> >>     at scala.util.Try$.apply(Try.scala:191)
> >>     at
> >>
> akka.actor.ReflectiveDynamicAccess.createInstanceFor(DynamicAccess.scala:73)
> >>     at
> >>
> akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$3.apply(DynamicAccess.scala:84)
> >>     at
> >>
> akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$3.apply(DynamicAccess.scala:84)
> >>     at scala.util.Success.flatMap(Try.scala:230)
> >>     at
> >>
> akka.actor.ReflectiveDynamicAccess.createInstanceFor(DynamicAccess.scala:84)
> >>     at akka.actor.ActorSystemImpl.liftedTree1$1(ActorSystem.scala:585)
> >>     at akka.actor.ActorSystemImpl.<init>(ActorSystem.scala:578)
> >>     at akka.actor.ActorSystem$.apply(ActorSystem.scala:142)
> >>     at akka.actor.ActorSystem$.apply(ActorSystem.scala:119)
> >>     at akka.actor.ActorSystem$.create(ActorSystem.scala:67)
> >>     at
> >>
> org.apache.flink.runtime.akka.AkkaUtils$.createActorSystem(AkkaUtils.scala:84)
> >>     at
> >>
> org.apache.flink.runtime.minicluster.FlinkMiniCluster.startJobManagerActorSystem(FlinkMiniCluster.scala:196)
> >>     at
> >>
> org.apache.flink.runtime.minicluster.FlinkMiniCluster.singleActorSystem$lzycompute$1(FlinkMiniCluster.scala:225)
> >>     at org.apache.flink.runtime.minicluster.FlinkMiniCluster.org
> >>
> $apache$flink$runtime$minicluster$FlinkMiniCluster$$singleActorSystem$1(FlinkMiniCluster.scala:225)
> >>     at
> >>
> org.apache.flink.runtime.minicluster.FlinkMiniCluster$$anonfun$1.apply(FlinkMiniCluster.scala:230)
> >>     at
> >>
> org.apache.flink.runtime.minicluster.FlinkMiniCluster$$anonfun$1.apply(FlinkMiniCluster.scala:228)
> >>     at
> >>
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:245)
> >>     at
> >>
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:245)
> >>     at scala.collection.immutable.Range.foreach(Range.scala:166)
> >>     at
> >> scala.collection.TraversableLike$class.map(TraversableLike.scala:245)
> >>     at scala.collection.AbstractTraversable.map(Traversable.scala:104)
> >>     at
> >>
> org.apache.flink.runtime.minicluster.FlinkMiniCluster.start(FlinkMiniCluster.scala:228)
> >>     at
> >>
> org.apache.flink.runtime.minicluster.FlinkMiniCluster.start(FlinkMiniCluster.scala:219)
> >>     at
> >>
> org.apache.flink.streaming.api.environment.LocalStreamEnvironment.execute(LocalStreamEnvironment.java:104)
> >>     at
> >> org.apache.flink.streaming.examples.wordcount.WordCount.main(WordCoun
> >> t
> >> .java:80)
> >>
> >> I know this is a naïve question but I would like to get some help in
> >> order to over come this issue. I tried various options like setting
> >> scala-2.10 as the compiler for the project (then it shows completely
> >> different error) and many of the projects don't even compile. But
> >> with
> >> 2.11 version I get the above stack trace. Any help here is welcome.
> >>
> >> Regards
> >> Ram
> >>
>
> Regards,
> Chiwan Park
>
>
>

Reply via email to