Memory segment error when migrating functional code from Flink 0.9 to 0.8

2015-02-09 Thread Andra Lungu
Hello everyone,

I am implementing a graph algorithm as part of a course and I will also add
it to the Flink- Gelly examples.
My problem is that I started developing it in the Gelly repository, which
runs on flink 0.9. It works like a charm there, but in order to test in on
a cluster to see its real capabilities, I need to move it to the course
repository, which runs on flink 0.8.

Initially, I thought this migration should occur without incidents since
flink 0.8 is more stable. Instead, I got the following exception:
java.lang.IllegalArgumentException: *Too few memory segments provided. Hash
Table needs at least 33 memory segments.*
at
org.apache.flink.runtime.operators.hash.CompactingHashTable.init(CompactingHashTable.java:238)
at
org.apache.flink.runtime.operators.hash.CompactingHashTable.init(CompactingHashTable.java:227)
at
org.apache.flink.runtime.iterative.task.IterationHeadPactTask.initCompactingHashTable(IterationHeadPactTask.java:177)
at
org.apache.flink.runtime.iterative.task.IterationHeadPactTask.run(IterationHeadPactTask.java:279)
at
org.apache.flink.runtime.operators.RegularPactTask.invoke(RegularPactTask.java:360)
at
org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:257)
at java.lang.Thread.run(Thread.java:745)

This is the code for Gelly, where all tests pass:
https://github.com/andralungu/flink-graph/tree/minSpanningTree
Unfortunately, the code for the course is private, so you cannot actually
see it... maybe @aalexandrov can do something about the privacy settings of
this repo.
https://github.com/andralungu/IMPRO-3.WS14/tree/dmst_algorithm

Thank you!
Andra


Re: Eclipse JDT, Java 8, lambdas

2015-02-09 Thread Nam-Luc Tran
I did try the 4.5 M4 release and it did not go straightforward.



--
View this message in context: 
http://apache-flink-incubator-mailing-list-archive.1008284.n3.nabble.com/Eclipse-JDT-Java-8-lambdas-tp3664p3688.html
Sent from the Apache Flink (Incubator) Mailing List archive. mailing list 
archive at Nabble.com.


Re: Memory segment error when migrating functional code from Flink 0.9 to 0.8

2015-02-09 Thread Till Rohrmann
Hi Andra,

have you tried increasing the number of network buffers in your cluster?
You can control by the configuration value:

taskmanager.network.numberOfBuffers: #numberBuffers

Greets,

Till

On Mon, Feb 9, 2015 at 9:56 AM, Andra Lungu lungu.an...@gmail.com wrote:

 Hello everyone,

 I am implementing a graph algorithm as part of a course and I will also add
 it to the Flink- Gelly examples.
 My problem is that I started developing it in the Gelly repository, which
 runs on flink 0.9. It works like a charm there, but in order to test in on
 a cluster to see its real capabilities, I need to move it to the course
 repository, which runs on flink 0.8.

 Initially, I thought this migration should occur without incidents since
 flink 0.8 is more stable. Instead, I got the following exception:
 java.lang.IllegalArgumentException: *Too few memory segments provided. Hash
 Table needs at least 33 memory segments.*
 at

 org.apache.flink.runtime.operators.hash.CompactingHashTable.init(CompactingHashTable.java:238)
 at

 org.apache.flink.runtime.operators.hash.CompactingHashTable.init(CompactingHashTable.java:227)
 at

 org.apache.flink.runtime.iterative.task.IterationHeadPactTask.initCompactingHashTable(IterationHeadPactTask.java:177)
 at

 org.apache.flink.runtime.iterative.task.IterationHeadPactTask.run(IterationHeadPactTask.java:279)
 at

 org.apache.flink.runtime.operators.RegularPactTask.invoke(RegularPactTask.java:360)
 at

 org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:257)
 at java.lang.Thread.run(Thread.java:745)

 This is the code for Gelly, where all tests pass:
 https://github.com/andralungu/flink-graph/tree/minSpanningTree
 Unfortunately, the code for the course is private, so you cannot actually
 see it... maybe @aalexandrov can do something about the privacy settings of
 this repo.
 https://github.com/andralungu/IMPRO-3.WS14/tree/dmst_algorithm

 Thank you!
 Andra



Re: Memory segment error when migrating functional code from Flink 0.9 to 0.8

2015-02-09 Thread Stephan Ewen
This is actually a problem of the number of memory segments available to
the hash table for the solution set.

For complex pipelines, memory currently gets too fragmented.

There are two workarounds, until we do the dynamic memory management, or
break it into shorter pipelines: Break the job up into subparts, or move
the solution set to the user code part of the heap. There is a flag to
attach to the Delta Iterations, see
DeltaIteration#setSolutionSetUnManaged()

Greetinigs,
Stephan


On Mon, Feb 9, 2015 at 10:32 AM, Till Rohrmann trohrm...@apache.org wrote:

 Hi Andra,

 have you tried increasing the number of network buffers in your cluster?
 You can control by the configuration value:

 taskmanager.network.numberOfBuffers: #numberBuffers

 Greets,

 Till

 On Mon, Feb 9, 2015 at 9:56 AM, Andra Lungu lungu.an...@gmail.com wrote:

  Hello everyone,
 
  I am implementing a graph algorithm as part of a course and I will also
 add
  it to the Flink- Gelly examples.
  My problem is that I started developing it in the Gelly repository, which
  runs on flink 0.9. It works like a charm there, but in order to test in
 on
  a cluster to see its real capabilities, I need to move it to the course
  repository, which runs on flink 0.8.
 
  Initially, I thought this migration should occur without incidents since
  flink 0.8 is more stable. Instead, I got the following exception:
  java.lang.IllegalArgumentException: *Too few memory segments provided.
 Hash
  Table needs at least 33 memory segments.*
  at
 
 
 org.apache.flink.runtime.operators.hash.CompactingHashTable.init(CompactingHashTable.java:238)
  at
 
 
 org.apache.flink.runtime.operators.hash.CompactingHashTable.init(CompactingHashTable.java:227)
  at
 
 
 org.apache.flink.runtime.iterative.task.IterationHeadPactTask.initCompactingHashTable(IterationHeadPactTask.java:177)
  at
 
 
 org.apache.flink.runtime.iterative.task.IterationHeadPactTask.run(IterationHeadPactTask.java:279)
  at
 
 
 org.apache.flink.runtime.operators.RegularPactTask.invoke(RegularPactTask.java:360)
  at
 
 
 org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:257)
  at java.lang.Thread.run(Thread.java:745)
 
  This is the code for Gelly, where all tests pass:
  https://github.com/andralungu/flink-graph/tree/minSpanningTree
  Unfortunately, the code for the course is private, so you cannot actually
  see it... maybe @aalexandrov can do something about the privacy settings
 of
  this repo.
  https://github.com/andralungu/IMPRO-3.WS14/tree/dmst_algorithm
 
  Thank you!
  Andra
 



Re: Eclipse JDT, Java 8, lambdas

2015-02-09 Thread Timo Walther

Hey,

it seems that 4.4.2 also includes the fix 
(https://projects.eclipse.org/projects/eclipse/releases/4.4.2/bugs) and 
will be released end of february. I will try Eclipse Luna SR2 RC2 today 
and check if it is working.


Regards,
Timo


On 09.02.2015 10:05, Nam-Luc Tran wrote:

I did try the 4.5 M4 release and it did not go straightforward.



--
View this message in context: 
http://apache-flink-incubator-mailing-list-archive.1008284.n3.nabble.com/Eclipse-JDT-Java-8-lambdas-tp3664p3688.html
Sent from the Apache Flink (Incubator) Mailing List archive. mailing list 
archive at Nabble.com.




[jira] [Created] (FLINK-1494) Build fails on BlobCacheTest

2015-02-09 Thread Fabian Hueske (JIRA)
Fabian Hueske created FLINK-1494:


 Summary: Build fails on BlobCacheTest
 Key: FLINK-1494
 URL: https://issues.apache.org/jira/browse/FLINK-1494
 Project: Flink
  Issue Type: Bug
  Components: Local Runtime, TaskManager
 Environment: Apache Maven 3.0.5
Maven home: /usr/share/maven
Java version: 1.7.0_65, vendor: Oracle Corporation
Java home: /usr/lib/jvm/java-7-openjdk-amd64/jre
Default locale: en_US, platform encoding: UTF-8
OS name: linux, version: 3.16.0-4-amd64, arch: amd64, family: unix
Reporter: Fabian Hueske


Building Flink with Maven repeatedly fails with the following error:

{code}
Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 127.283
sec  FAILURE! - in org.apache.flink.runtime.blob.BlobCacheTest
testBlobCache(org.apache.flink.runtime.blob.BlobCacheTest)  Time
elapsed: 127.282 sec   FAILURE!
java.lang.AssertionError: Could not connect to BlobServer at address
0.0.0.0/0.0.0.0:56760
at org.junit.Assert.fail(Assert.java:88)
at
org.apache.flink.runtime.blob.BlobCacheTest.testBlobCache(BlobCacheTest.java:109)

java.io.IOException: Could not connect to BlobServer at address
0.0.0.0/0.0.0.0:52657
at
org.apache.flink.runtime.blob.BlobClient.init(BlobClient.java:61)
at
org.apache.flink.runtime.execution.librarycache.BlobLibraryCacheManagerTest.testLibraryCacheManagerCleanup(BlobLibraryCacheManagerTest.java:56)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:483)
at
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
at
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
at
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
at
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
at
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
at
org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:264)
at
org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:153)
at
org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:124)
at
org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:200)
at
org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:153)
at
org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:103)
Caused by: java.net.ConnectException: Connection timed out
at java.net.PlainSocketImpl.socketConnect(Native Method)
at
java.net.AbstractPlainSocketImpl.doConnect(AbstractPlainSocketImpl.java:345)
at
java.net.AbstractPlainSocketImpl.connectToAddress(AbstractPlainSocketImpl.java:204)
at
java.net.AbstractPlainSocketImpl.connect(AbstractPlainSocketImpl.java:188)
at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392)
at java.net.Socket.connect(Socket.java:589)
at java.net.Socket.connect(Socket.java:538)
at
org.apache.flink.runtime.blob.BlobClient.init(BlobClient.java:59)
... 24 more
Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 127.299
sec  FAILURE! - in
org.apache.flink.runtime.execution.librarycache.BlobLibraryCacheManagerTest
testLibraryCacheManagerCleanup(org.apache.flink.runtime.execution.librarycache.BlobLibraryCacheManagerTest)
 Time elapsed: 127.298 sec   FAILURE!
java.lang.AssertionError: Could not connect to BlobServer at address
0.0.0.0/0.0.0.0:52657
at org.junit.Assert.fail(Assert.java:88)
at
org.apache.flink.runtime.execution.librarycache.BlobLibraryCacheManagerTest.testLibraryCacheManagerCleanup(BlobLibraryCacheManagerTest.java:108)


Results :

Failed tests:
  BlobCacheTest.testBlobCache:109 Could not connect to BlobServer at
address 0.0.0.0/0.0.0.0:56760
  BlobLibraryCacheManagerTest.testLibraryCacheManagerCleanup:108 Could
not connect to BlobServer at address 

[jira] [Created] (FLINK-1497) No documentation on MiniCluster

2015-02-09 Thread Sergey Dudoladov (JIRA)
Sergey Dudoladov created FLINK-1497:
---

 Summary: No documentation on MiniCluster
 Key: FLINK-1497
 URL: https://issues.apache.org/jira/browse/FLINK-1497
 Project: Flink
  Issue Type: Improvement
Affects Versions: 0.9
Reporter: Sergey Dudoladov
Priority: Trivial



 It looks like the Flink docs do not show how to run a MiniCluster. 
 It might be worth to document this feature  and  add relevant scripts to the 
/bin folder, e.g. start_mini.sh and stop_mini.sh



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: Planning Release 0.8.1

2015-02-09 Thread Márton Balassi
Went through the streaming commits with Gyula and assembled the ones
missing. Pushing it as soon as travis passes.

On Fri, Feb 6, 2015 at 2:26 PM, Robert Metzger rmetz...@apache.org wrote:

 It seems that quite a few important fixes still need some work until they
 are ready.
 I'll extend the deadline to Monday morning (CET), since we can not vote
 during the weekends anyways.

 On Fri, Feb 6, 2015 at 1:16 PM, Stephan Ewen se...@apache.org wrote:

  I am working on fixes for local split assignment and custom input split.
  Would be good to include those as well...
 
  On Fri, Feb 6, 2015 at 12:52 PM, Aljoscha Krettek aljos...@apache.org
  wrote:
 
   @robert, yes, will do
  
   On Fri, Feb 6, 2015 at 12:28 PM, Márton Balassi
   balassi.mar...@gmail.com wrote:
Found a streaming bug, Gyula fixed it. Pushing it soon to both master
  and
branch-0.8.
   
On Fri, Feb 6, 2015 at 11:51 AM, Robert Metzger rmetz...@apache.org
 
   wrote:
   
@Aljoscha, can you merge the backported fix to the release-0.8
  branch
when its ready?
   
On Fri, Feb 6, 2015 at 11:39 AM, Aljoscha Krettek 
  aljos...@apache.org
wrote:
   
 I have a fix for this user-discovered bug:


   
  
 
 https://issues.apache.org/jira/browse/FLINK-1463?jql=project%20%3D%20FLINK%20AND%20assignee%20%3D%20currentUser()%20AND%20resolution%20%3D%20Unresolved

 in this PR: https://github.com/apache/flink/pull/353

 This should probably also be back-ported to 0.8.1

 On Thu, Feb 5, 2015 at 3:54 PM, Aljoscha Krettek 
  aljos...@apache.org
   
 wrote:
  I have the PR ready. :D
 
 
  On Thu, Feb 5, 2015 at 2:43 PM, Stephan Ewen se...@apache.org
   wrote:
  I would like to add it to 0.8.1. People have asked for it...
 
  On Thu, Feb 5, 2015 at 2:36 PM, Aljoscha Krettek 
   aljos...@apache.org

  wrote:
 
  Do we want to have HadoopInputFormat support for Scala in
 0.8.1?
 
  On Thu, Feb 5, 2015 at 2:29 PM, Stephan Ewen 
 se...@apache.org
wrote:
   I think we need to make a pass through the recent 0.9
 commits
   and
 cherry
   pick some more into 0.8.1. There were quite a few bug fixes.
  
   Also, this one is rather critical and pending:
   https://github.com/apache/flink/pull/318
  
   On Thu, Feb 5, 2015 at 2:27 PM, Robert Metzger 
rmetz...@apache.org
  wrote:
  
   Hi guys,
  
   I would like to bundle a minor bugfix release for Flink
 soon.
   Some users were complaining about incomplete Kryo support,
 in
 particular
   for Avro.
  
   Also, we fixed some other issues which are easy to to port
 to
0.8.1
  (some
   of them are already in the branch).
  
   I would like to start the vote on the 0.8.1 release in
  roughly
   26
 hours.
   So please merge bugfixes you would like to have in 0.8.1
 into
   the
   release-0.8 branch in the next 24 hours.
  
   There are currently 2 open JIRAs assigned to 0.8.1, one
 with
  a
 pending
  pull
   request:
  
  
 

   
  
 
 https://issues.apache.org/jira/browse/FLINK-1422?jql=project%20%3D%20FLINK%20AND%20status%20%3D%20Open%20AND%20fixVersion%20%3D%200.8.1%20ORDER%20BY%20status%20DESC%2C%20priority%20DESC
  
  
   Please chime in if you are suggesting a different plan for
  the
0.8.1
   release.
  
  
   - Robert
  
 

   
  
 



Re: Planning Release 0.8.1

2015-02-09 Thread Robert Metzger
Cool.

I'm currently also testing my last change (kryo serializers). I think I'll
start creating the release candidate in one hour.

On Mon, Feb 9, 2015 at 2:24 PM, Márton Balassi balassi.mar...@gmail.com
wrote:

 Went through the streaming commits with Gyula and assembled the ones
 missing. Pushing it as soon as travis passes.

 On Fri, Feb 6, 2015 at 2:26 PM, Robert Metzger rmetz...@apache.org
 wrote:

  It seems that quite a few important fixes still need some work until they
  are ready.
  I'll extend the deadline to Monday morning (CET), since we can not vote
  during the weekends anyways.
 
  On Fri, Feb 6, 2015 at 1:16 PM, Stephan Ewen se...@apache.org wrote:
 
   I am working on fixes for local split assignment and custom input
 split.
   Would be good to include those as well...
  
   On Fri, Feb 6, 2015 at 12:52 PM, Aljoscha Krettek aljos...@apache.org
 
   wrote:
  
@robert, yes, will do
   
On Fri, Feb 6, 2015 at 12:28 PM, Márton Balassi
balassi.mar...@gmail.com wrote:
 Found a streaming bug, Gyula fixed it. Pushing it soon to both
 master
   and
 branch-0.8.

 On Fri, Feb 6, 2015 at 11:51 AM, Robert Metzger 
 rmetz...@apache.org
  
wrote:

 @Aljoscha, can you merge the backported fix to the release-0.8
   branch
 when its ready?

 On Fri, Feb 6, 2015 at 11:39 AM, Aljoscha Krettek 
   aljos...@apache.org
 wrote:

  I have a fix for this user-discovered bug:
 
 

   
  
 
 https://issues.apache.org/jira/browse/FLINK-1463?jql=project%20%3D%20FLINK%20AND%20assignee%20%3D%20currentUser()%20AND%20resolution%20%3D%20Unresolved
 
  in this PR: https://github.com/apache/flink/pull/353
 
  This should probably also be back-ported to 0.8.1
 
  On Thu, Feb 5, 2015 at 3:54 PM, Aljoscha Krettek 
   aljos...@apache.org

  wrote:
   I have the PR ready. :D
  
  
   On Thu, Feb 5, 2015 at 2:43 PM, Stephan Ewen 
 se...@apache.org
wrote:
   I would like to add it to 0.8.1. People have asked for it...
  
   On Thu, Feb 5, 2015 at 2:36 PM, Aljoscha Krettek 
aljos...@apache.org
 
   wrote:
  
   Do we want to have HadoopInputFormat support for Scala in
  0.8.1?
  
   On Thu, Feb 5, 2015 at 2:29 PM, Stephan Ewen 
  se...@apache.org
 wrote:
I think we need to make a pass through the recent 0.9
  commits
and
  cherry
pick some more into 0.8.1. There were quite a few bug
 fixes.
   
Also, this one is rather critical and pending:
https://github.com/apache/flink/pull/318
   
On Thu, Feb 5, 2015 at 2:27 PM, Robert Metzger 
 rmetz...@apache.org
   wrote:
   
Hi guys,
   
I would like to bundle a minor bugfix release for Flink
  soon.
Some users were complaining about incomplete Kryo
 support,
  in
  particular
for Avro.
   
Also, we fixed some other issues which are easy to to
 port
  to
 0.8.1
   (some
of them are already in the branch).
   
I would like to start the vote on the 0.8.1 release in
   roughly
26
  hours.
So please merge bugfixes you would like to have in 0.8.1
  into
the
release-0.8 branch in the next 24 hours.
   
There are currently 2 open JIRAs assigned to 0.8.1, one
  with
   a
  pending
   pull
request:
   
   
  
 

   
  
 
 https://issues.apache.org/jira/browse/FLINK-1422?jql=project%20%3D%20FLINK%20AND%20status%20%3D%20Open%20AND%20fixVersion%20%3D%200.8.1%20ORDER%20BY%20status%20DESC%2C%20priority%20DESC
   
   
Please chime in if you are suggesting a different plan
 for
   the
 0.8.1
release.
   
   
- Robert
   
  
 

   
  
 



[DISCUSS] Distributed TPC-H DataGenerator for flink-contrib

2015-02-09 Thread Robert Metzger
Hi,

we recently added the flink-contrib module for user contributed tools etc.

On one of the last weekends, I've created a distributed tpch generator,
based on this libary: https://github.com/airlift/tpch (which is from a
PrestoDB developer and available on Maven central).

You can find my code here:
https://github.com/rmetzger/scratch/tree/distributed-tpch-generator

It contains two examples:
a) a full TPC data generator (as a flink program):
https://github.com/rmetzger/scratch/blob/distributed-tpch-generator/src/main/java/flink/generators/programs/TPCHGenerator.java

b) an example which generates two TPC-H tables on-the-fly to join them:
https://github.com/rmetzger/scratch/blob/distributed-tpch-generator/src/main/java/flink/generators/programs/TPCHGeneratorExample.java


Before I spend time on integrating it into the flink-contrib package, I
was wondering if the community is willing this contribution to Flink.


Best,
Robert