Hi Ram,

If you want to build Flink with Scala 2.10, just checkout Flink repository from 
github or download source code from homepage, run `mvn clean install 
-DskipTests` and import projects to your IDE. If you want to build Flink with 
Scala 2.11, you have to run `tools/change-scala-version.sh 2.11` before build 
the project. You can revert Scala version change by running 
`tools/change-scala-version.sh 2.10`.

About IDE, Flink community recommends IntelliJ IDEA because Scala IDE have some 
problems in Java/Scala mixed project like Flink. But I tested importing Flink 
project with Scala IDE 4.3.0, Scala 2.11.7 and Flink 0.10.0 source code. Note 
that you should import the project as maven project.

By the way, the community welcomes any questions. Please feel free to post 
questions. :)

> On Jan 11, 2016, at 7:30 PM, Vasudevan, Ramkrishna S 
> <ramkrishna.s.vasude...@intel.com> wrote:
> 
> Thank you very much for the reply. 
> I tried different ways and when I tried setting up the root pom.xml to 2.11 
> 
>               <scala.version>2.11.6</scala.version>
>               <scala.binary.version>2.11</scala.binary.version>
> 
> I got the following error
> [INFO] 
> ------------------------------------------------------------------------
> [ERROR] Failed to execute goal on project flink-scala: Could not resolve 
> depende
> ncies for project org.apache.flink:flink-scala:jar:1.0-SNAPSHOT: Could not 
> find
> artifact org.scalamacros:quasiquotes_2.11:jar:2.0.1 in central 
> (http://repo.mave
> n.apache.org/maven2) -> [Help 1]
> 
> If I leave the scala.binary.verson to be at 2.10 and the scala version to be 
> at 2.11.6 then I get the following problem
> [INFO] C:\flink\flink\flink-runtime\src\test\scala:-1: info: compiling
> [INFO] Compiling 366 source files to 
> C:\flink\flink\flink-runtime\target\test-cl
> asses at 1452508064750
> [ERROR] 
> C:\flink\flink\flink-runtime\src\test\scala\org\apache\flink\runtime\job
> manager\JobManagerITCase.scala:700: error: can't expand macros compiled by 
> previ
> ous versions of Scala
> [ERROR]               assert(cachedGraph2.isArchived)
> [ERROR]                                   ^
> 
> So am not pretty sure how to proceed with this. If I try to change the 
> version of scala to 2.10 in the IDE then I get lot of compilation issues.  IS 
> there any way to over come this?
> 
> Once again thanks a lot and apologies for the naïve question.
> 
> Regards
> Ram
> -----Original Message-----
> From: ewenstep...@gmail.com [mailto:ewenstep...@gmail.com] On Behalf Of 
> Stephan Ewen
> Sent: Friday, January 8, 2016 5:01 PM
> To: dev@flink.apache.org
> Subject: Re: Naive question
> 
> Hi!
> 
> This looks like a mismatch between the Scala dependency in Flink and Scala in 
> your Eclipse. Make sure you use the same for both. By default, Flink 
> reference Scala 2.10
> 
> If your IDE is set up for Scala 2.11, set the Scala version variable in the 
> Flink root pom.xml also to 2.11
> 
> Greetings,
> Stephan
> 
> 
> 
> 
> On Fri, Jan 8, 2016 at 12:06 PM, Vasudevan, Ramkrishna S < 
> ramkrishna.s.vasude...@intel.com> wrote:
> 
>> I have been trying to install, learn and understand Flink. I am using
>> Scala- EclipseIDE as my IDE.
>> 
>> I have downloaded the flink source coded, compiled and created the project.
>> 
>> My work laptop is Windows based and I don't have eclipse based 
>> workstation but I do have linux boxes for running and testing things.
>> 
>> Some of the examples given in Flink source code do run directly from 
>> Eclipse but when I try to run the Wordcount example from Eclipse I get 
>> this error
>> 
>> Exception in thread "main" java.lang.NoSuchMethodError:
>> scala.collection.immutable.HashSet$.empty()Lscala/collection/immutable/HashSet;
>>     at akka.actor.ActorCell$.<init>(ActorCell.scala:336)
>>     at akka.actor.ActorCell$.<clinit>(ActorCell.scala)
>>     at akka.actor.RootActorPath.$div(ActorPath.scala:159)
>>     at akka.actor.LocalActorRefProvider.<init>(ActorRefProvider.scala:464)
>>     at akka.actor.LocalActorRefProvider.<init>(ActorRefProvider.scala:452)
>>     at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
>> Method)
>>     at sun.reflect.NativeConstructorAccessorImpl.newInstance(Unknown
>> Source)
>>     at 
>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(Unknown
>> Source)
>>     at java.lang.reflect.Constructor.newInstance(Unknown Source)
>>     at
>> akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$2.apply(DynamicAccess.scala:78)
>>     at scala.util.Try$.apply(Try.scala:191)
>>     at
>> akka.actor.ReflectiveDynamicAccess.createInstanceFor(DynamicAccess.scala:73)
>>     at
>> akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$3.apply(DynamicAccess.scala:84)
>>     at
>> akka.actor.ReflectiveDynamicAccess$$anonfun$createInstanceFor$3.apply(DynamicAccess.scala:84)
>>     at scala.util.Success.flatMap(Try.scala:230)
>>     at
>> akka.actor.ReflectiveDynamicAccess.createInstanceFor(DynamicAccess.scala:84)
>>     at akka.actor.ActorSystemImpl.liftedTree1$1(ActorSystem.scala:585)
>>     at akka.actor.ActorSystemImpl.<init>(ActorSystem.scala:578)
>>     at akka.actor.ActorSystem$.apply(ActorSystem.scala:142)
>>     at akka.actor.ActorSystem$.apply(ActorSystem.scala:119)
>>     at akka.actor.ActorSystem$.create(ActorSystem.scala:67)
>>     at
>> org.apache.flink.runtime.akka.AkkaUtils$.createActorSystem(AkkaUtils.scala:84)
>>     at
>> org.apache.flink.runtime.minicluster.FlinkMiniCluster.startJobManagerActorSystem(FlinkMiniCluster.scala:196)
>>     at
>> org.apache.flink.runtime.minicluster.FlinkMiniCluster.singleActorSystem$lzycompute$1(FlinkMiniCluster.scala:225)
>>     at org.apache.flink.runtime.minicluster.FlinkMiniCluster.org
>> $apache$flink$runtime$minicluster$FlinkMiniCluster$$singleActorSystem$1(FlinkMiniCluster.scala:225)
>>     at
>> org.apache.flink.runtime.minicluster.FlinkMiniCluster$$anonfun$1.apply(FlinkMiniCluster.scala:230)
>>     at
>> org.apache.flink.runtime.minicluster.FlinkMiniCluster$$anonfun$1.apply(FlinkMiniCluster.scala:228)
>>     at
>> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:245)
>>     at
>> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:245)
>>     at scala.collection.immutable.Range.foreach(Range.scala:166)
>>     at
>> scala.collection.TraversableLike$class.map(TraversableLike.scala:245)
>>     at scala.collection.AbstractTraversable.map(Traversable.scala:104)
>>     at
>> org.apache.flink.runtime.minicluster.FlinkMiniCluster.start(FlinkMiniCluster.scala:228)
>>     at
>> org.apache.flink.runtime.minicluster.FlinkMiniCluster.start(FlinkMiniCluster.scala:219)
>>     at
>> org.apache.flink.streaming.api.environment.LocalStreamEnvironment.execute(LocalStreamEnvironment.java:104)
>>     at
>> org.apache.flink.streaming.examples.wordcount.WordCount.main(WordCount
>> .java:80)
>> 
>> I know this is a naïve question but I would like to get some help in 
>> order to over come this issue. I tried various options like setting 
>> scala-2.10 as the compiler for the project (then it shows completely 
>> different error) and many of the projects don't even compile. But with 
>> 2.11 version I get the above stack trace. Any help here is welcome.
>> 
>> Regards
>> Ram
>> 

Regards,
Chiwan Park


Reply via email to