Re: Contributing to Flink

2015-02-26 Thread Robert Metzger
Hi Niraj,

Welcome to the Flink community ;)
I'm really excited that you want to contribute to our project, and since
you've asked for something in the security area, I actually have something
very concrete in mind.
We recently added some support for accessing (Kerberos) secured HDFS
clusters in Flink: https://issues.apache.org/jira/browse/FLINK-1504.
However, the implementation is very simple because it assumes that every
worker of Flink (TaskManager) is authenticated with Kerberos (kinit). Its
not very practical for large setups because you have to ssh to all machines
to log into Kerberos.

What I would really like to have in Flink would be an way to transfer the
authentication tokens form the JobManager (master) to the TaskManagers.
This way, users only have to be authenticated with Kerberos at the
JobManager, and Flink is taking care of the rest.
As far as I understood it, Hadoop has already all the utilities in place
for getting and transferring the delegation tokens.
Max Michels, another committer in our project has quite a good
understanding of the details there. It would be great if you (Max) could
chime in if I forgot something.

If you are interested in working on this, you can file a JIRA (
https://issues.apache.org/jira/browse/FLINK) for tracking the progress and
discussing the details.
If not I'm sure we'll come up with more interesting ideas.


Robert







On Thu, Feb 26, 2015 at 11:07 PM, Henry Saputra henry.sapu...@gmail.com
wrote:

 Hi Niraj,

 Thanks for your interest at Apache Flink. The quickest is to just give
 Flink a spin and figure out how it works.
 This would get you start on how it works before actually doing work on
 Flink =)

 Please do visit Flink how to contribute page [1] and subscribe to dev
 mailing list [2] to start following up.

 Welcome =)

 [1] http://flink.apache.org/how-to-contribute.html
 [2] http://flink.apache.org/community.html#mailing-lists

 On Thu, Feb 26, 2015 at 1:45 PM, Rai, Niraj niraj@intel.com wrote:
  Hi Flink Dev,
  I am looking to contribute to Flink, especially in the area of security.
 In the past, I have contributed to Pig, Hive and HDFS. I would really
 appreciate, if I can get some work assigned to me. Looking forward to hear
 back from the development community of Flink.
  Thanks
  Niraj
 



Re: gelli graph algorithm

2015-02-26 Thread Andra Lungu
Hello,

Also, for guidelines on how to implement a graph algorithm in Gelly, you
can
use the provided examples:
https://github.com/apache/flink/tree/master/flink-staging/flink-gelly/src/main/java/org/apache/flink/graph/example

Have fun!
Andra

On Thu, Feb 26, 2015 at 5:31 PM, Fabian Hueske fhue...@gmail.com wrote:

 Hi Martin,

 as a start, there is a PR with Gelly documentation:
 https://github.com/vasia/flink/blob/gelly-guide/docs/gelly_guide.md

 Cheers, Fabian

 2015-02-26 17:12 GMT+01:00 Martin Neumann mneum...@spotify.com:

  Hej,
 
  I was busy with other stuff for a while but I hope I will have more time
 to
  work on Flink and Graphs again now.
 
  I need to do some basic analytic's on a large graph set (stuff like
 degree
  distribution, triangle count, component size distribution etc.)
  Is there anything implemented in Gelli already that I can use? Is there
  some basic documentation about Gelli somewhere?
 
  If I need to program those algorithm myself I could do it in a way that
 we
  can integrate them in Gelli. But I might need some help to make sure I
  design them in a way that fits, so I might need some guidance/help.
 
  cheers Martin
 



gelli graph algorithm

2015-02-26 Thread Martin Neumann
Hej,

I was busy with other stuff for a while but I hope I will have more time to
work on Flink and Graphs again now.

I need to do some basic analytic's on a large graph set (stuff like degree
distribution, triangle count, component size distribution etc.)
Is there anything implemented in Gelli already that I can use? Is there
some basic documentation about Gelli somewhere?

If I need to program those algorithm myself I could do it in a way that we
can integrate them in Gelli. But I might need some help to make sure I
design them in a way that fits, so I might need some guidance/help.

cheers Martin


gelli graph algorithm

2015-02-26 Thread Martin Neumann
Hej,

I was busy with other stuff for a while but I hope I will have more time to
work on Flink and Graphs again now.

I need to do some basic analytic's on a large graph set (stuff like degree
distribution, triangle count, component size distribution etc.)
Is there anything implemented in Gelli already that I can use? Is there
some basic documentation about Gelli somewhere?

If I need to program those algorithm myself I could do it in a way that we
can integrate them in Gelli. But I might need some help to make sure I
design them in a way that fits, so I might need some guidance/help.

cheers Martin


Re: gelli graph algorithm

2015-02-26 Thread Fabian Hueske
Hi Martin,

as a start, there is a PR with Gelly documentation:
https://github.com/vasia/flink/blob/gelly-guide/docs/gelly_guide.md

Cheers, Fabian

2015-02-26 17:12 GMT+01:00 Martin Neumann mneum...@spotify.com:

 Hej,

 I was busy with other stuff for a while but I hope I will have more time to
 work on Flink and Graphs again now.

 I need to do some basic analytic's on a large graph set (stuff like degree
 distribution, triangle count, component size distribution etc.)
 Is there anything implemented in Gelli already that I can use? Is there
 some basic documentation about Gelli somewhere?

 If I need to program those algorithm myself I could do it in a way that we
 can integrate them in Gelli. But I might need some help to make sure I
 design them in a way that fits, so I might need some guidance/help.

 cheers Martin



Re: Could not build up connection to JobManager

2015-02-26 Thread Dulaj Viduranga
Hi,
It’s great to help out. :)

Setting 127.0.0.1 instead of “localhost” in jobmanager.rpc.address, 
helped to build the connection to the jobmanager. Apparently localhost 
resolving is different in webclient and the jobmanager. I think it’s good to 
set jobmanager.rpc.address: 127.0.0.1 in future builds.
But then I get this error when I tried to run examples. I don’t know if 
I should move this issue to another thread. If so please tell me.

bin/flink run 
/Users/Vidura/Documents/Development/flink/flink-dist/target/flink-0.9-SNAPSHOT-bin/flink-0.9-SNAPSHOT/examples/flink-java-examples-0.9-SNAPSHOT-WordCount.jar
 
/Users/Vidura/Documents/Development/flink/flink-dist/target/flink-0.9-SNAPSHOT-bin/flink-0.9-SNAPSHOT/hamlet.txt
 $FLINK_DIRECTORY/count


20:46:21,998 WARN  org.apache.hadoop.util.NativeCodeLoader  
 - Unable to load native-hadoop library for your platform... using builtin-java 
classes where applicable
02/26/2015 20:46:23 Job execution switched to status RUNNING.
02/26/2015 20:46:23 CHAIN DataSource (at getTextDataSet(WordCount.java:141) 
(org.apache.flink.api.java.io.TextInputFormat)) - FlatMap (FlatMap at 
main(WordCount.java:69)) - Combine(SUM(1), at main(WordCount.java:72)(1/1) 
switched to SCHEDULED 
02/26/2015 20:46:23 CHAIN DataSource (at getTextDataSet(WordCount.java:141) 
(org.apache.flink.api.java.io.TextInputFormat)) - FlatMap (FlatMap at 
main(WordCount.java:69)) - Combine(SUM(1), at main(WordCount.java:72)(1/1) 
switched to DEPLOYING 
02/26/2015 20:48:03 CHAIN DataSource (at getTextDataSet(WordCount.java:141) 
(org.apache.flink.api.java.io.TextInputFormat)) - FlatMap (FlatMap at 
main(WordCount.java:69)) - Combine(SUM(1), at main(WordCount.java:72)(1/1) 
switched to FAILED 
akka.pattern.AskTimeoutException: Ask timed out on 
[Actor[akka://flink/user/taskmanager#-1628133761]] after [10 ms]
at 
akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
at 
scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
at 
scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
at 
akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
at 
akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
at 
akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
at 
akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
at java.lang.Thread.run(Thread.java:745)

02/26/2015 20:48:03 Job execution switched to status FAILING.
02/26/2015 20:48:03 Reduce (SUM(1), at main(WordCount.java:72)(1/1) 
switched to CANCELED 
02/26/2015 20:48:03 DataSink(CsvOutputFormat (path: 
/Users/Vidura/Documents/Development/flink/flink-dist/target/flink-0.9-SNAPSHOT-bin/flink-0.9-SNAPSHOT/count,
 delimiter:  ))(1/1) switched to CANCELED 
02/26/2015 20:48:03 Job execution switched to status FAILED.
org.apache.flink.client.program.ProgramInvocationException: The program 
execution failed.
at org.apache.flink.client.program.Client.run(Client.java:344)
at org.apache.flink.client.program.Client.run(Client.java:306)
at org.apache.flink.client.program.Client.run(Client.java:300)
at 
org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:55)
at 
org.apache.flink.examples.java.wordcount.WordCount.main(WordCount.java:82)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:483)
at 
org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:437)
at 
org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:353)
at org.apache.flink.client.program.Client.run(Client.java:250)
at 
org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:371)
at org.apache.flink.client.CliFrontend.run(CliFrontend.java:344)
at 
org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:1087)
at org.apache.flink.client.CliFrontend.main(CliFrontend.java:1114)
Caused by: org.apache.flink.runtime.client.JobExecutionException: Job execution 
failed.
at 
org.apache.flink.runtime.jobmanager.JobManager$$anonfun$receiveWithLogMessages$1.applyOrElse(JobManager.scala:284)
at 
scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33)
at 
scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33)
at 

Re: Log4j configuration JUnit vs. Scalatest in IntelliJ

2015-02-26 Thread Ufuk Celebi

On 25 Feb 2015, at 16:35, Till Rohrmann trohrm...@apache.org wrote:

 The reason for this behaviour is the following:
 
 The log4j-test.properties is not a standard log4j properties file. It is
 only used if it is explicitly given to the executing JVM by
 -Dlog4j.configuration. The parent pom defines for the surefire plugin a
 corresponding argument line with this option. Intellij uses these maven
 settings for executing Java JUnit tests. For the scala tests which are
 executed with JUnitRunner, these settings are not used. I don't know why
 this is the case for the scala tests. But in order to get consistent
 behaviour you can turn off the maven settings import by unticking Build,
 Execution, Deployment - Build Tools - Maven - Running Tests - argLine
 in the IntelliJ settings.

Thanks. Makes more sense now. :-)


[jira] [Created] (FLINK-1613) Cannost submit to remote ExecutionEnvironment from IDE

2015-02-26 Thread Alexander Alexandrov (JIRA)
Alexander Alexandrov created FLINK-1613:
---

 Summary: Cannost submit to remote ExecutionEnvironment from IDE
 Key: FLINK-1613
 URL: https://issues.apache.org/jira/browse/FLINK-1613
 Project: Flink
  Issue Type: Bug
  Components: Distributed Runtime
Affects Versions: 0.8.1
 Environment: * Ubuntu Linux 14.04
* Flink 0.9-SNAPSHOT or 0.8.1 running in standalone mode on localhost
Reporter: Alexander Alexandrov
 Fix For: 0.9, 0.8.2


I am reporting this as [~rmetzler] mentioned offline that it was working in the 
past.

At the moment it is not possible to submit jobs directly from the IDE. Both the 
Java and the Scala quickstart guides fail on both 0.8.1 and 0.9-SNAPSHOT with 
ClassNotFoundException exceptions.

To reproduce the error, run the quickstart scripts and change the 
ExecutionEnvironment initialization:

{code:java}
env = ExecutionEnvironment.createRemoteEnvironment(localhost, 6123)
{code}


This is the cause for Java:

{noformat}
Caused by: java.lang.ClassNotFoundException: 
org.myorg.quickstart.WordCount$LineSplitter
at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
at java.lang.Class.forName0(Native Method)
at java.lang.Class.forName(Class.java:274)
at 
org.apache.flink.util.InstantiationUtil$ClassLoaderObjectInputStream.resolveClass(InstantiationUtil.java:54)
at 
java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1612)
at java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517)
at 
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1771)
at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
at 
java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
at 
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370)
at 
org.apache.flink.util.InstantiationUtil.deserializeObject(InstantiationUtil.java:274)
at 
org.apache.flink.util.InstantiationUtil.readObjectFromConfig(InstantiationUtil.java:236)
at 
org.apache.flink.runtime.operators.util.TaskConfig.getStubWrapper(TaskConfig.java:281)
{noformat}

This is for Scala:

{noformat}
java.lang.ClassNotFoundException: org.myorg.quickstart.WordCount$$anon$2$$anon$1
at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
at java.lang.Class.forName0(Native Method)
at java.lang.Class.forName(Class.java:274)
at 
org.apache.flink.util.InstantiationUtil$ClassLoaderObjectInputStream.resolveClass(InstantiationUtil.java:54)
at 
java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1612)
at java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1517)
at 
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1771)
at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370)
at 
org.apache.flink.util.InstantiationUtil.deserializeObject(InstantiationUtil.java:274)
at 
org.apache.flink.util.InstantiationUtil.readObjectFromConfig(InstantiationUtil.java:236)
at 
org.apache.flink.api.java.typeutils.runtime.RuntimeSerializerFactory.readParametersFromConfig(RuntimeSerializerFactory.java:76)
at 
org.apache.flink.runtime.operators.util.TaskConfig.getTypeSerializerFactory(TaskConfig.java:1084)
{noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Could not build up connection to JobManager

2015-02-26 Thread Stephan Ewen
Hi Dulaj!

Thanks for helping to debug.

My guess is that you are seeing now the same thing between JobManager and
TaskManager as you saw before between JobManager and JobClient. I have a
patch pending that should help the issue (see
https://issues.apache.org/jira/browse/FLINK-1608), let's see if that solves
it.

What seems not right is that the JobManager initially accepted the
TaskManager and later the communication. Can you paste the TaskManager log
as well?

Also: There must be something fairly unique about your network
configuration, as it works on all other setups that we use (locally, cloud,
test servers, YARN, ...). Can you paste your ipconfig / ifconfig by any
chance?

Greetings,
Stephan



On Thu, Feb 26, 2015 at 4:33 PM, Dulaj Viduranga vidura...@icloud.com
wrote:

 Hi,
 It’s great to help out. :)

 Setting 127.0.0.1 instead of “localhost” in
 jobmanager.rpc.address, helped to build the connection to the jobmanager.
 Apparently localhost resolving is different in webclient and the
 jobmanager. I think it’s good to set jobmanager.rpc.address: 127.0.0.1 in
 future builds.
 But then I get this error when I tried to run examples. I don’t
 know if I should move this issue to another thread. If so please tell me.

 bin/flink run
 /Users/Vidura/Documents/Development/flink/flink-dist/target/flink-0.9-SNAPSHOT-bin/flink-0.9-SNAPSHOT/examples/flink-java-examples-0.9-SNAPSHOT-WordCount.jar
 /Users/Vidura/Documents/Development/flink/flink-dist/target/flink-0.9-SNAPSHOT-bin/flink-0.9-SNAPSHOT/hamlet.txt
 $FLINK_DIRECTORY/count


 20:46:21,998 WARN  org.apache.hadoop.util.NativeCodeLoader
- Unable to load native-hadoop library for your platform... using
 builtin-java classes where applicable
 02/26/2015 20:46:23 Job execution switched to status RUNNING.
 02/26/2015 20:46:23 CHAIN DataSource (at
 getTextDataSet(WordCount.java:141)
 (org.apache.flink.api.java.io.TextInputFormat)) - FlatMap (FlatMap at
 main(WordCount.java:69)) - Combine(SUM(1), at main(WordCount.java:72)(1/1)
 switched to SCHEDULED
 02/26/2015 20:46:23 CHAIN DataSource (at
 getTextDataSet(WordCount.java:141)
 (org.apache.flink.api.java.io.TextInputFormat)) - FlatMap (FlatMap at
 main(WordCount.java:69)) - Combine(SUM(1), at main(WordCount.java:72)(1/1)
 switched to DEPLOYING
 02/26/2015 20:48:03 CHAIN DataSource (at
 getTextDataSet(WordCount.java:141)
 (org.apache.flink.api.java.io.TextInputFormat)) - FlatMap (FlatMap at
 main(WordCount.java:69)) - Combine(SUM(1), at main(WordCount.java:72)(1/1)
 switched to FAILED
 akka.pattern.AskTimeoutException: Ask timed out on
 [Actor[akka://flink/user/taskmanager#-1628133761]] after [10 ms]
 at
 akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333)
 at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117)
 at
 scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694)
 at
 scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691)
 at
 akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467)
 at
 akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419)
 at
 akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423)
 at
 akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375)
 at java.lang.Thread.run(Thread.java:745)

 02/26/2015 20:48:03 Job execution switched to status FAILING.
 02/26/2015 20:48:03 Reduce (SUM(1), at main(WordCount.java:72)(1/1)
 switched to CANCELED
 02/26/2015 20:48:03 DataSink(CsvOutputFormat (path:
 /Users/Vidura/Documents/Development/flink/flink-dist/target/flink-0.9-SNAPSHOT-bin/flink-0.9-SNAPSHOT/count,
 delimiter:  ))(1/1) switched to CANCELED
 02/26/2015 20:48:03 Job execution switched to status FAILED.
 org.apache.flink.client.program.ProgramInvocationException: The program
 execution failed.
 at org.apache.flink.client.program.Client.run(Client.java:344)
 at org.apache.flink.client.program.Client.run(Client.java:306)
 at org.apache.flink.client.program.Client.run(Client.java:300)
 at
 org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:55)
 at
 org.apache.flink.examples.java.wordcount.WordCount.main(WordCount.java:82)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
 at
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:483)
 at
 org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:437)
 at
 org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:353)
 at 

Contributing to Flink

2015-02-26 Thread Rai, Niraj
Hi Flink Dev,
I am looking to contribute to Flink, especially in the area of security. In the 
past, I have contributed to Pig, Hive and HDFS. I would really appreciate, if I 
can get some work assigned to me. Looking forward to hear back from the 
development community of Flink.
Thanks
Niraj



Re: Questions about flink-streaming-examples

2015-02-26 Thread Matthias J. Sax
Thanks for clarifying Marton!

I was on the latest build already. However, my local maven repository
contained old jars. After removing all flink-jars from my local maven
repository it works!

Why does maven no automatically update the local repository?


-Matthias



On 02/26/2015 09:20 AM, Márton Balassi wrote:
 Dear Mathias,
 
 Thanks for reporting the issue. I have successfully built
 flink-streaming-examples with maven, you can depend on test classes, the
 following in the pom does the trick:
 
 dependency
   groupIdorg.apache.flink/groupId
   artifactIdflink-streaming-core/artifactId
   version${project.version}/version
   scopetest/scope
   classifiertests/classifier
 /dependency
 
 This tells maven that the test classes of the examples (scope) might depend
 on the tests of the core (classifier). Without explicitly stating the tests
 classifier indeed maven only uses the classes under src/main.
 
 I think the problem occurred because you only built
 flink-streaming-examples and you might have used an outdated version of the
 flink-streaming-core as a dependency installed to your local maven
 repository - which did not include the StreamingTestBase yet. The change
 that we are talking is about a week old, I wrote it.
 
 Kind regards,
 
 Marton
 
 On Thu, Feb 26, 2015 at 9:04 AM, Matthias J. Sax 
 mj...@informatik.hu-berlin.de wrote:
 
 Hi,

 I just build flink-streaming and avoid the problem. I guess, that the
 issue is related to the module structure and dependencies.
 flink-streaming-examples uses
 org.apache.flink.streaming.util.StreamingProgramTestBase (that is
 defined in flink-streaming-core/src/TEST

 From my understanding, maven does not support dependencies into src/test
 but only into src/main (maybe I am wrong about this).

 I would suggest, to move StreamingProgramTestBase into a new module
 flink-streaming-testUtils (into src/main) and change the dependency of
 flink-streaming-examples accordingly.

 Why the build work on the parent-project, is still a riddle to me... But
 it seems that the dependency into scr/test can be resolved for this case.


 -Matthias




 On 02/25/2015 06:12 PM, Dulaj Viduranga wrote:
 Hi,
 Were you able to solve this? It seams the examples depend on
 flink-streaming-core”. I think you have to build it before.

 On Feb 25, 2015, at 10:04 PM, Matthias J. Sax 
 mj...@informatik.hu-berlin.de wrote:

 Thanks!

 Even if I am not a building and/or maven expert, it seems to me that the
 dependencies are not configures correctly... No clue how to fix it. But
 it might be a good idea to have a look into it IMHO.

 -Matthias


 On 02/25/2015 05:07 PM, Stephan Ewen wrote:
 Hi Matthias!

 Can you try mvn clean package? It package is generally preferable to
 compile, in my opinion.

 There may be a dependency through a test jar that is not properly
 handled
 when you do not execute the package goal.

 Stephan
 Am 25.02.2015 16:54 schrieb Max Michels m...@apache.org:

 Hi Matthias,

 Did you really pull from the latest master? I just tried to compile
 flink-streaming-examples using mvn clean compile test-compile and it
 worked.

 Best regards,
 Max

 On Wed, Feb 25, 2015 at 4:13 PM, Matthias J. Sax
 mj...@informatik.hu-berlin.de wrote:
 Hi,

 I tried to build flink-streaming-examples, but got an compilation
 error.
 If I build flink-streaming (that implicitly builds
 flink-streaming-examples), it works.

 I tried it on a fresh clone of flink/master using mvn clean compile
 test-compile

 Can anybody explain this behavior?


 -Matthias




 mjsax@T420s-dbis-mjsax
 :~/workspace_flink/dummy/flink/flink-staging/flink-streaming/flink-streaming-examples$
 mvn clean compile test-compile
 [INFO] Scanning for projects...
 [WARNING]
 [WARNING] Some problems were encountered while building the
 effective
 model for org.apache.flink:flink-streaming-examples:jar:0.9-SNAPSHOT
 [WARNING] 'build.plugins.plugin.(groupId:artifactId)' must be unique
 but found duplicate declaration of plugin
 org.apache.maven.plugins:maven-jar-plugin @ line 462, column 12
 [WARNING]
 [WARNING] It is highly recommended to fix these problems because
 they
 threaten the stability of your build.
 [WARNING]
 [WARNING] For this reason, future Maven versions might no longer
 support building such malformed projects.
 [WARNING]
 [INFO]
 [INFO]

 
 [INFO] Building flink-streaming-examples 0.9-SNAPSHOT
 [INFO]

 
 [INFO]
 [INFO] --- maven-clean-plugin:2.5:clean (default-clean) @
 flink-streaming-examples ---
 [INFO] Deleting

 /home/mjsax/workspace_flink/dummy/flink/flink-staging/flink-streaming/flink-streaming-examples/target
 [INFO]
 [INFO] --- maven-checkstyle-plugin:2.12.1:check (validate) @
 flink-streaming-examples ---
 [INFO]
 [INFO]
 [INFO] --- maven-enforcer-plugin:1.3.1:enforce (enforce-maven) @
 flink-streaming-examples ---
 [INFO]
 [INFO] --- 

Re: Questions about flink-streaming-examples

2015-02-26 Thread Robert Metzger
To update the local repository, you have to do execute the install goal.
I can recommend to always do a  mvn clean install

On Thu, Feb 26, 2015 at 10:11 AM, Matthias J. Sax 
mj...@informatik.hu-berlin.de wrote:

 Thanks for clarifying Marton!

 I was on the latest build already. However, my local maven repository
 contained old jars. After removing all flink-jars from my local maven
 repository it works!

 Why does maven no automatically update the local repository?


 -Matthias



 On 02/26/2015 09:20 AM, Márton Balassi wrote:
  Dear Mathias,
 
  Thanks for reporting the issue. I have successfully built
  flink-streaming-examples with maven, you can depend on test classes, the
  following in the pom does the trick:
 
  dependency
groupIdorg.apache.flink/groupId
artifactIdflink-streaming-core/artifactId
version${project.version}/version
scopetest/scope
classifiertests/classifier
  /dependency
 
  This tells maven that the test classes of the examples (scope) might
 depend
  on the tests of the core (classifier). Without explicitly stating the
 tests
  classifier indeed maven only uses the classes under src/main.
 
  I think the problem occurred because you only built
  flink-streaming-examples and you might have used an outdated version of
 the
  flink-streaming-core as a dependency installed to your local maven
  repository - which did not include the StreamingTestBase yet. The change
  that we are talking is about a week old, I wrote it.
 
  Kind regards,
 
  Marton
 
  On Thu, Feb 26, 2015 at 9:04 AM, Matthias J. Sax 
  mj...@informatik.hu-berlin.de wrote:
 
  Hi,
 
  I just build flink-streaming and avoid the problem. I guess, that the
  issue is related to the module structure and dependencies.
  flink-streaming-examples uses
  org.apache.flink.streaming.util.StreamingProgramTestBase (that is
  defined in flink-streaming-core/src/TEST
 
  From my understanding, maven does not support dependencies into src/test
  but only into src/main (maybe I am wrong about this).
 
  I would suggest, to move StreamingProgramTestBase into a new module
  flink-streaming-testUtils (into src/main) and change the dependency of
  flink-streaming-examples accordingly.
 
  Why the build work on the parent-project, is still a riddle to me... But
  it seems that the dependency into scr/test can be resolved for this
 case.
 
 
  -Matthias
 
 
 
 
  On 02/25/2015 06:12 PM, Dulaj Viduranga wrote:
  Hi,
  Were you able to solve this? It seams the examples depend on
  flink-streaming-core”. I think you have to build it before.
 
  On Feb 25, 2015, at 10:04 PM, Matthias J. Sax 
  mj...@informatik.hu-berlin.de wrote:
 
  Thanks!
 
  Even if I am not a building and/or maven expert, it seems to me that
 the
  dependencies are not configures correctly... No clue how to fix it.
 But
  it might be a good idea to have a look into it IMHO.
 
  -Matthias
 
 
  On 02/25/2015 05:07 PM, Stephan Ewen wrote:
  Hi Matthias!
 
  Can you try mvn clean package? It package is generally preferable
 to
  compile, in my opinion.
 
  There may be a dependency through a test jar that is not properly
  handled
  when you do not execute the package goal.
 
  Stephan
  Am 25.02.2015 16:54 schrieb Max Michels m...@apache.org:
 
  Hi Matthias,
 
  Did you really pull from the latest master? I just tried to compile
  flink-streaming-examples using mvn clean compile test-compile and
 it
  worked.
 
  Best regards,
  Max
 
  On Wed, Feb 25, 2015 at 4:13 PM, Matthias J. Sax
  mj...@informatik.hu-berlin.de wrote:
  Hi,
 
  I tried to build flink-streaming-examples, but got an compilation
  error.
  If I build flink-streaming (that implicitly builds
  flink-streaming-examples), it works.
 
  I tried it on a fresh clone of flink/master using mvn clean
 compile
  test-compile
 
  Can anybody explain this behavior?
 
 
  -Matthias
 
 
 
 
  mjsax@T420s-dbis-mjsax
 
 :~/workspace_flink/dummy/flink/flink-staging/flink-streaming/flink-streaming-examples$
  mvn clean compile test-compile
  [INFO] Scanning for projects...
  [WARNING]
  [WARNING] Some problems were encountered while building the
  effective
  model for org.apache.flink:flink-streaming-examples:jar:0.9-SNAPSHOT
  [WARNING] 'build.plugins.plugin.(groupId:artifactId)' must be
 unique
  but found duplicate declaration of plugin
  org.apache.maven.plugins:maven-jar-plugin @ line 462, column 12
  [WARNING]
  [WARNING] It is highly recommended to fix these problems because
  they
  threaten the stability of your build.
  [WARNING]
  [WARNING] For this reason, future Maven versions might no longer
  support building such malformed projects.
  [WARNING]
  [INFO]
  [INFO]
 
  
  [INFO] Building flink-streaming-examples 0.9-SNAPSHOT
  [INFO]
 
  
  [INFO]
  [INFO] --- maven-clean-plugin:2.5:clean (default-clean) @
  flink-streaming-examples ---
  [INFO] 

Re: Questions about flink-streaming-examples

2015-02-26 Thread Till Rohrmann
If the streaming-examples module uses the classifier tag to add the
test-core dependency then we should change it into type tag as
recommended by maven [1]. Otherwise it might come to build failures if the
install lifecycle is not executed.

The dependency import should look like:

dependency
  groupIdorg.apache.flink/groupId
  artifactIdflink-streaming-core/artifactId
  version${project.version}/version
  scopetest/scope
  typetest-jar/type
/dependency

[1] http://maven.apache.org/guides/mini/guide-attached-tests.html

On Thu, Feb 26, 2015 at 10:31 AM, Robert Metzger rmetz...@apache.org
wrote:

 To update the local repository, you have to do execute the install goal.
 I can recommend to always do a  mvn clean install

 On Thu, Feb 26, 2015 at 10:11 AM, Matthias J. Sax 
 mj...@informatik.hu-berlin.de wrote:

  Thanks for clarifying Marton!
 
  I was on the latest build already. However, my local maven repository
  contained old jars. After removing all flink-jars from my local maven
  repository it works!
 
  Why does maven no automatically update the local repository?
 
 
  -Matthias
 
 
 
  On 02/26/2015 09:20 AM, Márton Balassi wrote:
   Dear Mathias,
  
   Thanks for reporting the issue. I have successfully built
   flink-streaming-examples with maven, you can depend on test classes,
 the
   following in the pom does the trick:
  
   dependency
 groupIdorg.apache.flink/groupId
 artifactIdflink-streaming-core/artifactId
 version${project.version}/version
 scopetest/scope
 classifiertests/classifier
   /dependency
  
   This tells maven that the test classes of the examples (scope) might
  depend
   on the tests of the core (classifier). Without explicitly stating the
  tests
   classifier indeed maven only uses the classes under src/main.
  
   I think the problem occurred because you only built
   flink-streaming-examples and you might have used an outdated version of
  the
   flink-streaming-core as a dependency installed to your local maven
   repository - which did not include the StreamingTestBase yet. The
 change
   that we are talking is about a week old, I wrote it.
  
   Kind regards,
  
   Marton
  
   On Thu, Feb 26, 2015 at 9:04 AM, Matthias J. Sax 
   mj...@informatik.hu-berlin.de wrote:
  
   Hi,
  
   I just build flink-streaming and avoid the problem. I guess, that
 the
   issue is related to the module structure and dependencies.
   flink-streaming-examples uses
   org.apache.flink.streaming.util.StreamingProgramTestBase (that is
   defined in flink-streaming-core/src/TEST
  
   From my understanding, maven does not support dependencies into
 src/test
   but only into src/main (maybe I am wrong about this).
  
   I would suggest, to move StreamingProgramTestBase into a new module
   flink-streaming-testUtils (into src/main) and change the dependency
 of
   flink-streaming-examples accordingly.
  
   Why the build work on the parent-project, is still a riddle to me...
 But
   it seems that the dependency into scr/test can be resolved for this
  case.
  
  
   -Matthias
  
  
  
  
   On 02/25/2015 06:12 PM, Dulaj Viduranga wrote:
   Hi,
   Were you able to solve this? It seams the examples depend on
   flink-streaming-core”. I think you have to build it before.
  
   On Feb 25, 2015, at 10:04 PM, Matthias J. Sax 
   mj...@informatik.hu-berlin.de wrote:
  
   Thanks!
  
   Even if I am not a building and/or maven expert, it seems to me that
  the
   dependencies are not configures correctly... No clue how to fix it.
  But
   it might be a good idea to have a look into it IMHO.
  
   -Matthias
  
  
   On 02/25/2015 05:07 PM, Stephan Ewen wrote:
   Hi Matthias!
  
   Can you try mvn clean package? It package is generally preferable
  to
   compile, in my opinion.
  
   There may be a dependency through a test jar that is not properly
   handled
   when you do not execute the package goal.
  
   Stephan
   Am 25.02.2015 16:54 schrieb Max Michels m...@apache.org:
  
   Hi Matthias,
  
   Did you really pull from the latest master? I just tried to
 compile
   flink-streaming-examples using mvn clean compile test-compile
 and
  it
   worked.
  
   Best regards,
   Max
  
   On Wed, Feb 25, 2015 at 4:13 PM, Matthias J. Sax
   mj...@informatik.hu-berlin.de wrote:
   Hi,
  
   I tried to build flink-streaming-examples, but got an compilation
   error.
   If I build flink-streaming (that implicitly builds
   flink-streaming-examples), it works.
  
   I tried it on a fresh clone of flink/master using mvn clean
  compile
   test-compile
  
   Can anybody explain this behavior?
  
  
   -Matthias
  
  
  
  
   mjsax@T420s-dbis-mjsax
  
 
 :~/workspace_flink/dummy/flink/flink-staging/flink-streaming/flink-streaming-examples$
   mvn clean compile test-compile
   [INFO] Scanning for projects...
   [WARNING]
   [WARNING] Some problems were encountered while building the
   effective
   model for
 org.apache.flink:flink-streaming-examples:jar:0.9-SNAPSHOT
   [WARNING] 

Re: Questions about flink-streaming-examples

2015-02-26 Thread Márton Balassi
Dear Mathias,

Thanks for reporting the issue. I have successfully built
flink-streaming-examples with maven, you can depend on test classes, the
following in the pom does the trick:

dependency
  groupIdorg.apache.flink/groupId
  artifactIdflink-streaming-core/artifactId
  version${project.version}/version
  scopetest/scope
  classifiertests/classifier
/dependency

This tells maven that the test classes of the examples (scope) might depend
on the tests of the core (classifier). Without explicitly stating the tests
classifier indeed maven only uses the classes under src/main.

I think the problem occurred because you only built
flink-streaming-examples and you might have used an outdated version of the
flink-streaming-core as a dependency installed to your local maven
repository - which did not include the StreamingTestBase yet. The change
that we are talking is about a week old, I wrote it.

Kind regards,

Marton

On Thu, Feb 26, 2015 at 9:04 AM, Matthias J. Sax 
mj...@informatik.hu-berlin.de wrote:

 Hi,

 I just build flink-streaming and avoid the problem. I guess, that the
 issue is related to the module structure and dependencies.
 flink-streaming-examples uses
 org.apache.flink.streaming.util.StreamingProgramTestBase (that is
 defined in flink-streaming-core/src/TEST

 From my understanding, maven does not support dependencies into src/test
 but only into src/main (maybe I am wrong about this).

 I would suggest, to move StreamingProgramTestBase into a new module
 flink-streaming-testUtils (into src/main) and change the dependency of
 flink-streaming-examples accordingly.

 Why the build work on the parent-project, is still a riddle to me... But
 it seems that the dependency into scr/test can be resolved for this case.


 -Matthias




 On 02/25/2015 06:12 PM, Dulaj Viduranga wrote:
  Hi,
  Were you able to solve this? It seams the examples depend on
 flink-streaming-core”. I think you have to build it before.
 
  On Feb 25, 2015, at 10:04 PM, Matthias J. Sax 
 mj...@informatik.hu-berlin.de wrote:
 
  Thanks!
 
  Even if I am not a building and/or maven expert, it seems to me that the
  dependencies are not configures correctly... No clue how to fix it. But
  it might be a good idea to have a look into it IMHO.
 
  -Matthias
 
 
  On 02/25/2015 05:07 PM, Stephan Ewen wrote:
  Hi Matthias!
 
  Can you try mvn clean package? It package is generally preferable to
  compile, in my opinion.
 
  There may be a dependency through a test jar that is not properly
 handled
  when you do not execute the package goal.
 
  Stephan
  Am 25.02.2015 16:54 schrieb Max Michels m...@apache.org:
 
  Hi Matthias,
 
  Did you really pull from the latest master? I just tried to compile
  flink-streaming-examples using mvn clean compile test-compile and it
  worked.
 
  Best regards,
  Max
 
  On Wed, Feb 25, 2015 at 4:13 PM, Matthias J. Sax
  mj...@informatik.hu-berlin.de wrote:
  Hi,
 
  I tried to build flink-streaming-examples, but got an compilation
 error.
  If I build flink-streaming (that implicitly builds
  flink-streaming-examples), it works.
 
  I tried it on a fresh clone of flink/master using mvn clean compile
  test-compile
 
  Can anybody explain this behavior?
 
 
  -Matthias
 
 
 
 
  mjsax@T420s-dbis-mjsax
 :~/workspace_flink/dummy/flink/flink-staging/flink-streaming/flink-streaming-examples$
  mvn clean compile test-compile
  [INFO] Scanning for projects...
  [WARNING]
  [WARNING] Some problems were encountered while building the
 effective
  model for org.apache.flink:flink-streaming-examples:jar:0.9-SNAPSHOT
  [WARNING] 'build.plugins.plugin.(groupId:artifactId)' must be unique
  but found duplicate declaration of plugin
  org.apache.maven.plugins:maven-jar-plugin @ line 462, column 12
  [WARNING]
  [WARNING] It is highly recommended to fix these problems because
 they
  threaten the stability of your build.
  [WARNING]
  [WARNING] For this reason, future Maven versions might no longer
  support building such malformed projects.
  [WARNING]
  [INFO]
  [INFO]
 
 
  [INFO] Building flink-streaming-examples 0.9-SNAPSHOT
  [INFO]
 
 
  [INFO]
  [INFO] --- maven-clean-plugin:2.5:clean (default-clean) @
  flink-streaming-examples ---
  [INFO] Deleting
 
 /home/mjsax/workspace_flink/dummy/flink/flink-staging/flink-streaming/flink-streaming-examples/target
  [INFO]
  [INFO] --- maven-checkstyle-plugin:2.12.1:check (validate) @
  flink-streaming-examples ---
  [INFO]
  [INFO]
  [INFO] --- maven-enforcer-plugin:1.3.1:enforce (enforce-maven) @
  flink-streaming-examples ---
  [INFO]
  [INFO] --- build-helper-maven-plugin:1.7:add-source (add-source) @
  flink-streaming-examples ---
  [INFO] Source directory:
 
 /home/mjsax/workspace_flink/dummy/flink/flink-staging/flink-streaming/flink-streaming-examples/src/main/scala
  added.
  [INFO]
  [INFO] --- 

Re: Contributing to Flink

2015-02-26 Thread Henry Saputra
Hi Niraj,

Thanks for your interest at Apache Flink. The quickest is to just give
Flink a spin and figure out how it works.
This would get you start on how it works before actually doing work on Flink =)

Please do visit Flink how to contribute page [1] and subscribe to dev
mailing list [2] to start following up.

Welcome =)

[1] http://flink.apache.org/how-to-contribute.html
[2] http://flink.apache.org/community.html#mailing-lists

On Thu, Feb 26, 2015 at 1:45 PM, Rai, Niraj niraj@intel.com wrote:
 Hi Flink Dev,
 I am looking to contribute to Flink, especially in the area of security. In 
 the past, I have contributed to Pig, Hive and HDFS. I would really 
 appreciate, if I can get some work assigned to me. Looking forward to hear 
 back from the development community of Flink.
 Thanks
 Niraj



Re: Drop support for CDH4 / Hadoop 2.0.0-alpha

2015-02-26 Thread Henry Saputra
If we were to drop CDH4 / Hadoop 2.0.0-alpha, would this mean we do
not even to shade the hadoop fat jars, or we do still needed to
support 1.x ?

- Henry

On Thu, Feb 26, 2015 at 8:57 AM, Robert Metzger rmetz...@apache.org wrote:
 Hi,

 I'm currently working on https://issues.apache.org/jira/browse/FLINK-1605
 and its a hell of a mess.

 I got almost everything working, except for the hadoop 2.0.0-alpha profile.
 The profile exists because google protobuf has a different version in that
 Hadoop release.
 Since maven is setting the version of protobuf for the entire project to
 the older version, we have to use an older akka version which is causing
 issues.

 The logical conclusion from that would be shading Hadoop's protobuf version
 into the Hadoop jars. That by itself is working, however its not working
 for the flink-yarn-tests.

 I think I can also solve the issue with the flink-yarn-tests, but it would
 be a very dirty hack (either injecting shaded code into the failsafe
 tests-classpath or putting test code into src/main).

 But the general question remains: Are we willing to continue spending a lot
 of time on maintaining the profile?
 Till has spend a lot of time recently to fix failing testcases for that old
 akka version, I spend almost two days now on getting the
 shading/dependencies right, and I'm sure we'll keep having troubles with
 the profile.


 Therefore, I was wondering if this is the right time to drop support for
 CDH4 / Hadoop 2.0.0-alpha.


 Best,
 Robert