[jira] [Updated] (FLINK-1404) Add support to cache intermediate results

2015-04-07 Thread Ufuk Celebi (JIRA)

 [ 
https://issues.apache.org/jira/browse/FLINK-1404?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ufuk Celebi updated FLINK-1404:
---
Summary: Add support to cache intermediate results  (was: Trigger recycling 
of buffers held by historic intermediate result partitions)

 Add support to cache intermediate results
 -

 Key: FLINK-1404
 URL: https://issues.apache.org/jira/browse/FLINK-1404
 Project: Flink
  Issue Type: Improvement
  Components: Distributed Runtime
Reporter: Ufuk Celebi
Assignee: Ufuk Celebi

 With blocking intermediate results (FLINK-1350) and proper partition state 
 management (FLINK-1359) it is necessary to allow the network buffer pool to 
 request eviction of historic intermediate results when not enough buffers are 
 available. With the currently available pipelined intermediate partitions 
 this is not an issue, because buffer pools can be released as soon as a 
 partition is consumed.
 We need to be able to trigger the recycling of buffers held by historic 
 intermediate results when not enough buffers are available for new local 
 pools.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (FLINK-1831) runtime.taskmanager.RegistrationTests fails sporiously

2015-04-07 Thread JIRA
Márton Balassi created FLINK-1831:
-

 Summary: runtime.taskmanager.RegistrationTests fails sporiously
 Key: FLINK-1831
 URL: https://issues.apache.org/jira/browse/FLINK-1831
 Project: Flink
  Issue Type: Bug
  Components: Distributed Runtime
Affects Versions: 0.9
Reporter: Márton Balassi


Failing in a decent chunk of travis builds, I could also reproduce it locally:

akka.actor.InvalidActorNameException: actor name [FAKE_JOB_MANAGER] is not 
unique!
at 
akka.actor.dungeon.ChildrenContainer$TerminatingChildrenContainer.reserve(ChildrenContainer.scala:192)
at akka.actor.dungeon.Children$class.reserveChild(Children.scala:77)
at akka.actor.ActorCell.reserveChild(ActorCell.scala:369)
at akka.actor.dungeon.Children$class.makeChild(Children.scala:202)
at akka.actor.dungeon.Children$class.attachChild(Children.scala:42)
at akka.actor.ActorCell.attachChild(ActorCell.scala:369)
at akka.actor.ActorSystemImpl.actorOf(ActorSystem.scala:553)
at 
org.apache.flink.runtime.taskmanager.RegistrationTest$5.init(RegistrationTest.java:308)
at 
org.apache.flink.runtime.taskmanager.RegistrationTest.testTaskManagerResumesConnectAfterJobManagerFailure(RegistrationTest.java:266)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
at 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at 
org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
at org.junit.runner.JUnitCore.run(JUnitCore.java:160)
at 
com.intellij.junit4.JUnit4IdeaTestRunner.startRunnerWithArgs(JUnit4IdeaTestRunner.java:74)
at 
com.intellij.rt.execution.junit.JUnitStarter.prepareStreamsAndStart(JUnitStarter.java:211)
at 
com.intellij.rt.execution.junit.JUnitStarter.main(JUnitStarter.java:67)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at com.intellij.rt.execution.application.AppMain.main(AppMain.java:134)
09:46:39.486 [flink-akka.actor.default-dispatcher-2] 
[flink-akka.actor.default-dispatcher-3 - akka://flink/user/FAKE_JOB_MANAGER] 
ERROR akka.actor.OneForOneStrategy - Kill
akka.actor.ActorKilledException: Kill

java.lang.AssertionError: actor name [FAKE_JOB_MANAGER] is not unique!
at org.junit.Assert.fail(Assert.java:88)
at 
org.apache.flink.runtime.taskmanager.RegistrationTest$5.init(RegistrationTest.java:328)
at 
org.apache.flink.runtime.taskmanager.RegistrationTest.testTaskManagerResumesConnectAfterJobManagerFailure(RegistrationTest.java:266)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
 

[jira] [Commented] (FLINK-1829) Conflicting Jackson version in the Flink POMs

2015-04-07 Thread Stephan Ewen (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1829?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14482808#comment-14482808
 ] 

Stephan Ewen commented on FLINK-1829:
-

The `hbase-client` dependency should not be in there by default, or is it 
pulled in transitively?

For the others, I think we need to add an entry to the dependency management. 

 Conflicting Jackson version in the Flink POMs
 -

 Key: FLINK-1829
 URL: https://issues.apache.org/jira/browse/FLINK-1829
 Project: Flink
  Issue Type: Bug
  Components: Build System
Affects Versions: 0.9
Reporter: Alexander Alexandrov
 Fix For: 0.9


 The current POM setup transitively includes multiple conflicting versions of 
 the Jackson library over
 * {{com.amazonaws:aws-java-sdk}} (v. 2.1.1)
 * {{org.apache.avro:avro}} (v. 1.9.13)
 * {{org.apache.hbase:hbase-client}} (v. 1.8.8)
 When running jobs against a Flink local runtime embedded with:
 {code:xml}
 dependency
 groupIdorg.apache.flink/groupId
 artifactIdflink-scala/artifactId
 version${flink.version}/version
 scopeprovided/scope
 /dependency
 dependency
 groupIdorg.apache.flink/groupId
 artifactIdflink-java/artifactId
 version${flink.version}/version
 scopeprovided/scope
 /dependency
 dependency
 groupIdorg.apache.flink/groupId
 artifactIdflink-clients/artifactId
 version${flink.version}/version
 scopeprovided/scope
 /dependency
 {code}
 I get the following error:
 {noformat}
 15-04-04 15:52:04 ERROR exception during creation
 akka.actor.ActorInitializationException: exception during creation
   at akka.actor.ActorInitializationException$.apply(Actor.scala:164)
   at akka.actor.ActorCell.create(ActorCell.scala:596)
   at akka.actor.ActorCell.invokeAll$1(ActorCell.scala:456)
   at akka.actor.ActorCell.systemInvoke(ActorCell.scala:478)
   at akka.dispatch.Mailbox.processAllSystemMessages(Mailbox.scala:279)
   at akka.dispatch.Mailbox.run(Mailbox.scala:220)
   at akka.dispatch.Mailbox.exec(Mailbox.scala:231)
   at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
   at 
 scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
   at 
 scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
   at 
 scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
 Caused by: java.lang.reflect.InvocationTargetException
   at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
   at 
 sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
   at 
 sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
   at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
   at akka.util.Reflect$.instantiate(Reflect.scala:66)
   at akka.actor.ArgsReflectConstructor.produce(Props.scala:352)
   at akka.actor.Props.newActor(Props.scala:252)
   at akka.actor.ActorCell.newActor(ActorCell.scala:552)
   at akka.actor.ActorCell.create(ActorCell.scala:578)
   ... 9 more
 Caused by: java.lang.NoSuchMethodError: 
 com.fasterxml.jackson.core.JsonFactory.requiresPropertyOrdering()Z
   at 
 com.fasterxml.jackson.databind.ObjectMapper.init(ObjectMapper.java:445)
   at 
 com.fasterxml.jackson.databind.ObjectMapper.init(ObjectMapper.java:366)
   at 
 org.apache.flink.runtime.taskmanager.TaskManager.init(TaskManager.scala:134)
   ... 18 more
 {noformat}
 Fixing the Jackson version on the client side, e.g, with the following snippet
 {code:xml}
 dependency
 groupIdcom.fasterxml.jackson.core/groupId
 artifactIdjackson-core/artifactId
 version2.2.1/version
 scopeprovided/scope
 /dependency
 dependency
 groupIdcom.fasterxml.jackson.core/groupId
 artifactIdjackson-databind/artifactId
 version2.2.1/version
 scopeprovided/scope
 /dependency
 dependency
 groupIdcom.fasterxml.jackson.core/groupId
 artifactIdjackson-annotations/artifactId
 version2.2.1/version
 scopeprovided/scope
 /dependency
 {code}
 solves the problem, but I guess it will be better if we can stick with one 
 version in the build artifacts.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-967) Make intermediate results a first-class citizen in the JobGraph

2015-04-07 Thread Maximilian Michels (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-967?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14482816#comment-14482816
 ] 

Maximilian Michels commented on FLINK-967:
--

The intermediate results are now integrated into the JobGraph. We are currently 
working on the runtime parts.

 Make intermediate results a first-class citizen in the JobGraph
 ---

 Key: FLINK-967
 URL: https://issues.apache.org/jira/browse/FLINK-967
 Project: Flink
  Issue Type: New Feature
  Components: JobManager, TaskManager
Affects Versions: 0.6-incubating
Reporter: Stephan Ewen
Assignee: Stephan Ewen
 Fix For: 0.9


 In order to add incremental plan rollout to the system, we need to make 
 intermediate results a first-class citizen in the job graph and scheduler.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Closed] (FLINK-1831) runtime.taskmanager.RegistrationTests fails spuriously

2015-04-07 Thread Stephan Ewen (JIRA)

 [ 
https://issues.apache.org/jira/browse/FLINK-1831?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen closed FLINK-1831.
---

 runtime.taskmanager.RegistrationTests fails spuriously
 --

 Key: FLINK-1831
 URL: https://issues.apache.org/jira/browse/FLINK-1831
 Project: Flink
  Issue Type: Bug
  Components: Distributed Runtime
Affects Versions: 0.9
Reporter: Márton Balassi
Assignee: Stephan Ewen
 Fix For: 0.9


 Failing in a decent chunk of travis builds, I could also reproduce it locally:
 akka.actor.InvalidActorNameException: actor name [FAKE_JOB_MANAGER] is not 
 unique!
   at 
 akka.actor.dungeon.ChildrenContainer$TerminatingChildrenContainer.reserve(ChildrenContainer.scala:192)
   at akka.actor.dungeon.Children$class.reserveChild(Children.scala:77)
   at akka.actor.ActorCell.reserveChild(ActorCell.scala:369)
   at akka.actor.dungeon.Children$class.makeChild(Children.scala:202)
   at akka.actor.dungeon.Children$class.attachChild(Children.scala:42)
   at akka.actor.ActorCell.attachChild(ActorCell.scala:369)
   at akka.actor.ActorSystemImpl.actorOf(ActorSystem.scala:553)
   at 
 org.apache.flink.runtime.taskmanager.RegistrationTest$5.init(RegistrationTest.java:308)
   at 
 org.apache.flink.runtime.taskmanager.RegistrationTest.testTaskManagerResumesConnectAfterJobManagerFailure(RegistrationTest.java:266)
   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   at java.lang.reflect.Method.invoke(Method.java:497)
   at 
 org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
   at 
 org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
   at 
 org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
   at 
 org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
   at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
   at 
 org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
   at 
 org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
   at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
   at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
   at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
   at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
   at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
   at 
 org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
   at 
 org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
   at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
   at org.junit.runner.JUnitCore.run(JUnitCore.java:160)
   at 
 com.intellij.junit4.JUnit4IdeaTestRunner.startRunnerWithArgs(JUnit4IdeaTestRunner.java:74)
   at 
 com.intellij.rt.execution.junit.JUnitStarter.prepareStreamsAndStart(JUnitStarter.java:211)
   at 
 com.intellij.rt.execution.junit.JUnitStarter.main(JUnitStarter.java:67)
   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   at java.lang.reflect.Method.invoke(Method.java:497)
   at com.intellij.rt.execution.application.AppMain.main(AppMain.java:134)
 09:46:39.486 [flink-akka.actor.default-dispatcher-2] 
 [flink-akka.actor.default-dispatcher-3 - akka://flink/user/FAKE_JOB_MANAGER] 
 ERROR akka.actor.OneForOneStrategy - Kill
 akka.actor.ActorKilledException: Kill
 java.lang.AssertionError: actor name [FAKE_JOB_MANAGER] is not unique!
   at org.junit.Assert.fail(Assert.java:88)
   at 
 org.apache.flink.runtime.taskmanager.RegistrationTest$5.init(RegistrationTest.java:328)
   at 
 org.apache.flink.runtime.taskmanager.RegistrationTest.testTaskManagerResumesConnectAfterJobManagerFailure(RegistrationTest.java:266)
   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   at 
 org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
   at 
 org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
   at 
 

[jira] [Resolved] (FLINK-1831) runtime.taskmanager.RegistrationTests fails spuriously

2015-04-07 Thread Stephan Ewen (JIRA)

 [ 
https://issues.apache.org/jira/browse/FLINK-1831?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen resolved FLINK-1831.
-
   Resolution: Fixed
Fix Version/s: 0.9
 Assignee: Stephan Ewen

Fixed via 5cd9e9d94bb3af4a5a868789f663af262672cac8

 runtime.taskmanager.RegistrationTests fails spuriously
 --

 Key: FLINK-1831
 URL: https://issues.apache.org/jira/browse/FLINK-1831
 Project: Flink
  Issue Type: Bug
  Components: Distributed Runtime
Affects Versions: 0.9
Reporter: Márton Balassi
Assignee: Stephan Ewen
 Fix For: 0.9


 Failing in a decent chunk of travis builds, I could also reproduce it locally:
 akka.actor.InvalidActorNameException: actor name [FAKE_JOB_MANAGER] is not 
 unique!
   at 
 akka.actor.dungeon.ChildrenContainer$TerminatingChildrenContainer.reserve(ChildrenContainer.scala:192)
   at akka.actor.dungeon.Children$class.reserveChild(Children.scala:77)
   at akka.actor.ActorCell.reserveChild(ActorCell.scala:369)
   at akka.actor.dungeon.Children$class.makeChild(Children.scala:202)
   at akka.actor.dungeon.Children$class.attachChild(Children.scala:42)
   at akka.actor.ActorCell.attachChild(ActorCell.scala:369)
   at akka.actor.ActorSystemImpl.actorOf(ActorSystem.scala:553)
   at 
 org.apache.flink.runtime.taskmanager.RegistrationTest$5.init(RegistrationTest.java:308)
   at 
 org.apache.flink.runtime.taskmanager.RegistrationTest.testTaskManagerResumesConnectAfterJobManagerFailure(RegistrationTest.java:266)
   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   at java.lang.reflect.Method.invoke(Method.java:497)
   at 
 org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
   at 
 org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
   at 
 org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
   at 
 org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
   at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
   at 
 org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
   at 
 org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
   at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
   at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
   at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
   at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
   at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
   at 
 org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
   at 
 org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
   at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
   at org.junit.runner.JUnitCore.run(JUnitCore.java:160)
   at 
 com.intellij.junit4.JUnit4IdeaTestRunner.startRunnerWithArgs(JUnit4IdeaTestRunner.java:74)
   at 
 com.intellij.rt.execution.junit.JUnitStarter.prepareStreamsAndStart(JUnitStarter.java:211)
   at 
 com.intellij.rt.execution.junit.JUnitStarter.main(JUnitStarter.java:67)
   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   at java.lang.reflect.Method.invoke(Method.java:497)
   at com.intellij.rt.execution.application.AppMain.main(AppMain.java:134)
 09:46:39.486 [flink-akka.actor.default-dispatcher-2] 
 [flink-akka.actor.default-dispatcher-3 - akka://flink/user/FAKE_JOB_MANAGER] 
 ERROR akka.actor.OneForOneStrategy - Kill
 akka.actor.ActorKilledException: Kill
 java.lang.AssertionError: actor name [FAKE_JOB_MANAGER] is not unique!
   at org.junit.Assert.fail(Assert.java:88)
   at 
 org.apache.flink.runtime.taskmanager.RegistrationTest$5.init(RegistrationTest.java:328)
   at 
 org.apache.flink.runtime.taskmanager.RegistrationTest.testTaskManagerResumesConnectAfterJobManagerFailure(RegistrationTest.java:266)
   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   at 
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   at 
 org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
   at 
 

[jira] [Created] (FLINK-1832) start-local.bat/start-local.sh does not work if there is a white space in the file path (windows)

2015-04-07 Thread Nikolaas Steenbergen (JIRA)
Nikolaas Steenbergen created FLINK-1832:
---

 Summary: start-local.bat/start-local.sh does not work if there is 
a white space in the file path (windows)
 Key: FLINK-1832
 URL: https://issues.apache.org/jira/browse/FLINK-1832
 Project: Flink
  Issue Type: Bug
  Components: Local Runtime
Affects Versions: 0.8.1
 Environment: Windows 8.1
Reporter: Nikolaas Steenbergen
Priority: Minor


start-local.bat can be run, but exits immediately.

start-local.sh with cygwin exits with error:



owner_@owner /cygdrive/e/Program 
Files/flink-0.8.1-bin-hadoop2/flink-0.8.1-bin-hadoop2/flink-0.8.1/bin
$ ./start-local.sh
cygpath: error converting 
Files/flink-0.8.1-bin-hadoop2/flink-0.8.1-bin-hadoop2/flink-0.8.1/bin/.. - No 
such file or directory
sed: can't read E:\Program/conf/flink-conf.yaml: No such file or directory
sed: can't read E:\Program/conf/flink-conf.yaml: No such file or directory
sed: can't read E:\Program/conf/flink-conf.yaml: No such file or directory
sed: can't read E:\Program/conf/flink-conf.yaml: No such file or directory
sed: can't read E:\Program/conf/flink-conf.yaml: No such file or directory
sed: can't read E:\Program/conf/flink-conf.yaml: No such file or directory
sed: can't read E:\Program/conf/flink-conf.yaml: No such file or directory
cygpath: error converting 
/cygdrive/e/Program:Files/flink-0.8.1-bin-hadoop2/flink-0.8.1-bin-hadoop2/flink-0.8.1/bin/../lib/*.jar:.
 - Unknown error -1
./start-local.sh: line 27: E:\Program/bin/jobmanager.sh: No such file or 
directory




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (FLINK-1577) Misleading error messages when cancelling tasks

2015-04-07 Thread Ufuk Celebi (JIRA)

 [ 
https://issues.apache.org/jira/browse/FLINK-1577?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ufuk Celebi resolved FLINK-1577.

Resolution: Fixed

 Misleading error messages when cancelling tasks
 ---

 Key: FLINK-1577
 URL: https://issues.apache.org/jira/browse/FLINK-1577
 Project: Flink
  Issue Type: Improvement
  Components: Distributed Runtime
Affects Versions: master
Reporter: Ufuk Celebi
Assignee: Ufuk Celebi

 A user running a Flink version before bec9c4d ran into a job manager failure 
 (fixed in bec9c4d), which lead to restarting the JM and cancelling/clearing 
 all tasks on the TMs.
 The logs of the TMs were inconclusive. I think part of that has been fixed by 
 now, e.g. there is a log message when cancelAndClearEverything is called, but 
 the task thread (RuntimeEnvironment) always logs an error when interrupted 
 during the run method -- even if the task gets cancelled.
 I think these error messages are misleading and only the root cause is 
 important (i.e. non-failed tasks should be silently cancelled).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (FLINK-1469) Initialize network environment at task manager startup

2015-04-07 Thread Ufuk Celebi (JIRA)

 [ 
https://issues.apache.org/jira/browse/FLINK-1469?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ufuk Celebi resolved FLINK-1469.

   Resolution: Fixed
Fix Version/s: 0.9

Fixed in e74521c1fe5f5842185be5c454d27adc3bd254d5.

 Initialize network environment at task manager startup
 --

 Key: FLINK-1469
 URL: https://issues.apache.org/jira/browse/FLINK-1469
 Project: Flink
  Issue Type: Improvement
  Components: Distributed Runtime
Affects Versions: master
Reporter: Ufuk Celebi
Priority: Minor
 Fix For: 0.9


 Currently, the network environment is initialized only after the registration 
 of a task manager with the job manager finishes.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (FLINK-1834) Is mapred.output.dir conf parameter really required?

2015-04-07 Thread Flavio Pompermaier (JIRA)
Flavio Pompermaier created FLINK-1834:
-

 Summary: Is mapred.output.dir conf parameter really required?
 Key: FLINK-1834
 URL: https://issues.apache.org/jira/browse/FLINK-1834
 Project: Flink
  Issue Type: Improvement
  Components: Hadoop Compatibility
Affects Versions: 0.9
Reporter: Flavio Pompermaier
Priority: Minor


At the moment if you don't set mapred.output.dir parameter when output to an 
hadoop sink Flink throws an error. Probably this parameter could be defaulted 
to some tmp dir + jobId (for each output sink).
For example, when output to an HBase sink, the only things you'll find in the 
mapred.output.dir will be the SUCCESS files



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1832) start-local.bat/start-local.sh does not work if there is a white space in the file path (windows)

2015-04-07 Thread Maximilian Michels (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1832?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14482901#comment-14482901
 ] 

Maximilian Michels commented on FLINK-1832:
---

Thanks for reporting this. Looks like an easy fix.

 start-local.bat/start-local.sh does not work if there is a white space in the 
 file path (windows)
 -

 Key: FLINK-1832
 URL: https://issues.apache.org/jira/browse/FLINK-1832
 Project: Flink
  Issue Type: Bug
  Components: Local Runtime
Affects Versions: 0.8.1
 Environment: Windows 8.1
Reporter: Nikolaas Steenbergen
Priority: Minor

 start-local.bat can be run, but exits immediately.
 start-local.sh with cygwin exits with error:
 owner_@owner /cygdrive/e/Program 
 Files/flink-0.8.1-bin-hadoop2/flink-0.8.1-bin-hadoop2/flink-0.8.1/bin
 $ ./start-local.sh
 cygpath: error converting 
 Files/flink-0.8.1-bin-hadoop2/flink-0.8.1-bin-hadoop2/flink-0.8.1/bin/.. - 
 No such file or directory
 sed: can't read E:\Program/conf/flink-conf.yaml: No such file or directory
 sed: can't read E:\Program/conf/flink-conf.yaml: No such file or directory
 sed: can't read E:\Program/conf/flink-conf.yaml: No such file or directory
 sed: can't read E:\Program/conf/flink-conf.yaml: No such file or directory
 sed: can't read E:\Program/conf/flink-conf.yaml: No such file or directory
 sed: can't read E:\Program/conf/flink-conf.yaml: No such file or directory
 sed: can't read E:\Program/conf/flink-conf.yaml: No such file or directory
 cygpath: error converting 
 /cygdrive/e/Program:Files/flink-0.8.1-bin-hadoop2/flink-0.8.1-bin-hadoop2/flink-0.8.1/bin/../lib/*.jar:.
  - Unknown error -1
 ./start-local.sh: line 27: E:\Program/bin/jobmanager.sh: No such file or 
 directory



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1828) Impossible to output data to an HBase table

2015-04-07 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1828?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14482898#comment-14482898
 ] 

ASF GitHub Bot commented on FLINK-1828:
---

Github user fpompermaier commented on the pull request:

https://github.com/apache/flink/pull/571#issuecomment-90481118
  
Ok, I created this issue (https://issues.apache.org/jira/browse/FLINK-1834) 
about the mapred.output.dir 


 Impossible to output data to an HBase table
 ---

 Key: FLINK-1828
 URL: https://issues.apache.org/jira/browse/FLINK-1828
 Project: Flink
  Issue Type: Bug
  Components: Hadoop Compatibility
Affects Versions: 0.9
Reporter: Flavio Pompermaier
  Labels: hadoop, hbase
 Fix For: 0.9


 Right now it is not possible to use HBase TableOutputFormat as output format 
 because Configurable.setConf  is not called in the configure() method of the 
 HadoopOutputFormatBase



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: Fixed Configurable HadoopOutputFormat (FLINK-1...

2015-04-07 Thread fpompermaier
Github user fpompermaier commented on the pull request:

https://github.com/apache/flink/pull/571#issuecomment-90481118
  
Ok, I created this issue (https://issues.apache.org/jira/browse/FLINK-1834) 
about the mapred.output.dir 


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-1822) Update Community page on website with a Getting Help section

2015-04-07 Thread Maximilian Michels (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1822?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14482818#comment-14482818
 ] 

Maximilian Michels commented on FLINK-1822:
---

+1 very nice idea.

The path of the linked how_to_contribute page is wrong.

 Update Community page on website with a Getting Help section
 

 Key: FLINK-1822
 URL: https://issues.apache.org/jira/browse/FLINK-1822
 Project: Flink
  Issue Type: Task
  Components: Project Website
Reporter: Robert Metzger
Assignee: Robert Metzger
 Attachments: update_community.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: Fixed Configurable HadoopOutputFormat (FLINK-1...

2015-04-07 Thread mxm
Github user mxm commented on the pull request:

https://github.com/apache/flink/pull/571#issuecomment-90475934
  
Is the Hadoop configuration specified in the flink-conf.yaml loaded? If we 
set `mapred.output.dir` then we should check for an existing config entry 
beforehand. Otherwise, we overwrite Hadoop configuration values. Like @fhueske 
suggested, please open a JIRA for investigation. 


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-1828) Impossible to output data to an HBase table

2015-04-07 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1828?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14482834#comment-14482834
 ] 

ASF GitHub Bot commented on FLINK-1828:
---

Github user mxm commented on the pull request:

https://github.com/apache/flink/pull/571#issuecomment-90475934
  
Is the Hadoop configuration specified in the flink-conf.yaml loaded? If we 
set `mapred.output.dir` then we should check for an existing config entry 
beforehand. Otherwise, we overwrite Hadoop configuration values. Like @fhueske 
suggested, please open a JIRA for investigation. 


 Impossible to output data to an HBase table
 ---

 Key: FLINK-1828
 URL: https://issues.apache.org/jira/browse/FLINK-1828
 Project: Flink
  Issue Type: Bug
  Components: Hadoop Compatibility
Affects Versions: 0.9
Reporter: Flavio Pompermaier
  Labels: hadoop, hbase
 Fix For: 0.9


 Right now it is not possible to use HBase TableOutputFormat as output format 
 because Configurable.setConf  is not called in the configure() method of the 
 HadoopOutputFormatBase



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (FLINK-1833) Refactor partition availability notification in ExecutionGraph

2015-04-07 Thread Ufuk Celebi (JIRA)
Ufuk Celebi created FLINK-1833:
--

 Summary: Refactor partition availability notification in 
ExecutionGraph
 Key: FLINK-1833
 URL: https://issues.apache.org/jira/browse/FLINK-1833
 Project: Flink
  Issue Type: Improvement
  Components: Distributed Runtime, JobManager
Affects Versions: master
Reporter: Ufuk Celebi


The mechanism to notify the JobManager about available result partitions is 
hard to understand. The are two parts to this:

1) JobManager

- The deployment of receivers happens in the Execution class although it is by 
now totally unrelated to the state of a specific execution. I propose to move 
this to the respective IntermediateResultPartition.
- The deployment information for a receiver is spread across different 
components: when creating the TaskDeploymentDescriptor and the caching of 
partition infos at the consuming vertex. This is very hard to follow and 
results in unnecessary messages being sent (which are discarded at the TM).

2) TaskManager

- Pipelined results notify where you would expect it in the ResultPartition, 
but blocking results don't have an extra message and are implicitly 
piggy-backed to the final state transition, after which the job manager deploys 
receivers if all blocking partitions of a result have been produced.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (FLINK-1835) Spurious failure of YARN tests

2015-04-07 Thread Stephan Ewen (JIRA)
Stephan Ewen created FLINK-1835:
---

 Summary: Spurious failure of YARN tests
 Key: FLINK-1835
 URL: https://issues.apache.org/jira/browse/FLINK-1835
 Project: Flink
  Issue Type: Bug
  Components: YARN Client
Affects Versions: 0.9
Reporter: Stephan Ewen
Assignee: Robert Metzger
 Fix For: 0.9


THe failure was caused by detecting an exception in the log.

Stack trace of the exception (extracted from the log) below

{code}
21:18:29,555 WARN  org.apache.hadoop.util.NativeCodeLoader  
 - Unable to load native-hadoop library for your platform... using builtin-java 
classes where applicable
21:18:29,806 INFO  org.apache.flink.yarn.ApplicationMaster$ 
 - YARN daemon runs as travis setting user to execute Flink 
ApplicationMaster/JobManager to travis
21:18:29,808 INFO  org.apache.flink.yarn.ApplicationMaster$ 
 - 

21:18:29,809 INFO  org.apache.flink.yarn.ApplicationMaster$ 
 -  Starting YARN ApplicationMaster/JobManager (Version: 0.9-SNAPSHOT, 
Rev:d2020b5, Date:06.04.2015 @ 18:00:21 UTC)
21:18:29,809 INFO  org.apache.flink.yarn.ApplicationMaster$ 
 -  Current user: travis
21:18:29,809 INFO  org.apache.flink.yarn.ApplicationMaster$ 
 -  JVM: Java HotSpot(TM) 64-Bit Server VM - Oracle Corporation - 1.8/25.31-b07
21:18:29,809 INFO  org.apache.flink.yarn.ApplicationMaster$ 
 -  Maximum heap size: 393 MiBytes
21:18:29,826 INFO  org.apache.flink.yarn.ApplicationMaster$ 
 -  JAVA_HOME: /usr/lib/jvm/java-8-oracle
21:18:29,827 INFO  org.apache.flink.yarn.ApplicationMaster$ 
 -  JVM Options:
21:18:29,827 INFO  org.apache.flink.yarn.ApplicationMaster$ 
 - -Xmx409M
21:18:29,827 INFO  org.apache.flink.yarn.ApplicationMaster$ 
 - 
-Dlog.file=/home/travis/build/StephanEwen/incubator-flink/flink-yarn-tests/target/flink-yarn-tests-fifo/flink-yarn-tests-fifo-logDir-nm-1_0/application_1428355034517_0004/container_1428355034517_0004_01_01/jobmanager-main.log
21:18:29,827 INFO  org.apache.flink.yarn.ApplicationMaster$ 
 - -Dlogback.configurationFile=file:logback.xml
21:18:29,827 INFO  org.apache.flink.yarn.ApplicationMaster$ 
 - -Dlog4j.configuration=file:log4j.properties
21:18:29,827 INFO  org.apache.flink.yarn.ApplicationMaster$ 
 -  Program Arguments: (none)
21:18:29,827 INFO  org.apache.flink.yarn.ApplicationMaster$ 
 - 

21:18:29,828 INFO  org.apache.flink.yarn.ApplicationMaster$ 
 - registered UNIX signal handlers for [TERM, HUP, INT]
21:18:29,843 INFO  org.apache.flink.yarn.ApplicationMaster$ 
 - Starting JobManager for YARN
21:18:29,845 INFO  org.apache.flink.yarn.ApplicationMaster$ 
 - Loading config from: 
/home/travis/build/StephanEwen/incubator-flink/flink-yarn-tests/target/flink-yarn-tests-fifo/flink-yarn-tests-fifo-localDir-nm-1_0/usercache/travis/appcache/application_1428355034517_0004/container_1428355034517_0004_01_01
21:18:30,388 INFO  akka.event.slf4j.Slf4jLogger 
 - Slf4jLogger started
21:18:30,450 INFO  Remoting 
 - Starting remoting
21:18:30,637 INFO  Remoting 
 - Remoting started; listening on addresses 
:[akka.tcp://flink@172.17.0.176:34023]
21:18:30,651 INFO  org.apache.flink.runtime.blob.BlobServer 
 - Created BLOB server storage directory 
/tmp/blobStore-e34b86da-094c-4a4e-aa02-7b0556e8af93
21:18:30,655 INFO  org.apache.flink.runtime.blob.BlobServer 
 - Started BLOB server at 0.0.0.0:33717 - max concurrent requests: 50 - max 
backlog: 1000
21:18:30,670 INFO  org.apache.flink.yarn.ApplicationMaster$ 
 - Starting Job Manger web frontend.
21:18:30,673 INFO  org.apache.flink.runtime.jobmanager.web.WebInfoServer
 - Setting up web info server, using web-root directory 
jar:file:/home/travis/build/StephanEwen/incubator-flink/flink-yarn-tests/target/flink-yarn-tests-fifo/flink-yarn-tests-fifo-localDir-nm-1_0/usercache/travis/appcache/application_1428355034517_0004/filecache/12/flink-dist-0.9-SNAPSHOT.jar!/web-docs-infoserver.
21:18:30,705 INFO  org.apache.flink.runtime.jobmanager.JobManager   
 - Starting JobManager at akka://flink/user/jobmanager#395299512.
21:18:31,184 INFO  org.eclipse.jetty.util.log   
 - jetty-0.9-SNAPSHOT
21:18:31,269 INFO  org.eclipse.jetty.util.log  

[GitHub] flink pull request: Fix issue where Windows paths were not recogni...

2015-04-07 Thread fhueske
Github user fhueske commented on the pull request:

https://github.com/apache/flink/pull/491#issuecomment-90507970
  
Btw, the path `file:/c:` that motivated this PR is not a valid absolute 
path according to Microsoft's documentation. 


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Created] (FLINK-1836) just test

2015-04-07 Thread zhu (JIRA)
zhu created FLINK-1836:
--

 Summary: just test
 Key: FLINK-1836
 URL: https://issues.apache.org/jira/browse/FLINK-1836
 Project: Flink
  Issue Type: Bug
Reporter: zhu


just test



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1817) ClassLoaderObjectInputStream fails with ClassNotFoundException for primitive classes

2015-04-07 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1817?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14483114#comment-14483114
 ] 

ASF GitHub Bot commented on FLINK-1817:
---

Github user asfgit closed the pull request at:

https://github.com/apache/flink/pull/565


 ClassLoaderObjectInputStream fails with ClassNotFoundException for primitive 
 classes
 

 Key: FLINK-1817
 URL: https://issues.apache.org/jira/browse/FLINK-1817
 Project: Flink
  Issue Type: Bug
  Components: core
Affects Versions: 0.9, 0.8.1
Reporter: Fabian Hueske
Assignee: Fabian Hueske
 Fix For: 0.9, 0.8.2


 User function objects that contain primitive type classes (such as 
 {{int.class}} or {{long.class}}) are not supported because the 
 {{ClassLoaderObjectInputStream}} fails with a {{ClassNotFoundException}} when 
 reading the object from the {{TaskConfig}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (FLINK-1838) Update streaming programming guide

2015-04-07 Thread Gyula Fora (JIRA)
Gyula Fora created FLINK-1838:
-

 Summary: Update streaming programming guide
 Key: FLINK-1838
 URL: https://issues.apache.org/jira/browse/FLINK-1838
 Project: Flink
  Issue Type: Task
  Components: Documentation, Streaming
Reporter: Gyula Fora
Assignee: Gyula Fora


The current streaming programming guide has gotten slightly outdated since the 
last release. We need to update it to reflect the changes in the API and 
runtime.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1834) Is mapred.output.dir conf parameter really required?

2015-04-07 Thread Fabian Hueske (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1834?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14483249#comment-14483249
 ] 

Fabian Hueske commented on FLINK-1834:
--

I totally agree. We should get rid of this mandatory parameter.

We try to only set it if it is set by the user and let the HadoopFormat figure 
out whether it needs it or not. If it does and it is not set, the job should 
fail.

 Is mapred.output.dir conf parameter really required?
 

 Key: FLINK-1834
 URL: https://issues.apache.org/jira/browse/FLINK-1834
 Project: Flink
  Issue Type: Improvement
  Components: Hadoop Compatibility
Affects Versions: 0.9
Reporter: Flavio Pompermaier
Priority: Minor
  Labels: hadoop

 At the moment if you don't set mapred.output.dir parameter when output to an 
 hadoop sink Flink throws an error. Probably this parameter could be defaulted 
 to some tmp dir + jobId (for each output sink).
 For example, when output to an HBase sink, the only things you'll find in the 
 mapred.output.dir will be the SUCCESS files



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Closed] (FLINK-1560) Add ITCases for streaming examples

2015-04-07 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/FLINK-1560?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Márton Balassi closed FLINK-1560.
-
   Resolution: Implemented
Fix Version/s: 0.9

Implemented via ed7d165, 464e782 and 954beca.

 Add ITCases for streaming examples
 --

 Key: FLINK-1560
 URL: https://issues.apache.org/jira/browse/FLINK-1560
 Project: Flink
  Issue Type: Test
  Components: Streaming
Affects Versions: 0.9
Reporter: Márton Balassi
Assignee: Péter Szabó
 Fix For: 0.9


 Currently there are no tests for consistency of the streaming example 
 programs. This might be a real show stopper for users who encounter an issue 
 there.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1560) Add ITCases for streaming examples

2015-04-07 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1560?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14483216#comment-14483216
 ] 

ASF GitHub Bot commented on FLINK-1560:
---

Github user asfgit closed the pull request at:

https://github.com/apache/flink/pull/519


 Add ITCases for streaming examples
 --

 Key: FLINK-1560
 URL: https://issues.apache.org/jira/browse/FLINK-1560
 Project: Flink
  Issue Type: Test
  Components: Streaming
Affects Versions: 0.9
Reporter: Márton Balassi
Assignee: Péter Szabó

 Currently there are no tests for consistency of the streaming example 
 programs. This might be a real show stopper for users who encounter an issue 
 there.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1834) Is mapred.output.dir conf parameter really required?

2015-04-07 Thread Maximilian Michels (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1834?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14483258#comment-14483258
 ] 

Maximilian Michels commented on FLINK-1834:
---

+1

 Is mapred.output.dir conf parameter really required?
 

 Key: FLINK-1834
 URL: https://issues.apache.org/jira/browse/FLINK-1834
 Project: Flink
  Issue Type: Improvement
  Components: Hadoop Compatibility
Affects Versions: 0.9
Reporter: Flavio Pompermaier
Priority: Minor
  Labels: hadoop

 At the moment if you don't set mapred.output.dir parameter when output to an 
 hadoop sink Flink throws an error. Probably this parameter could be defaulted 
 to some tmp dir + jobId (for each output sink).
 For example, when output to an HBase sink, the only things you'll find in the 
 mapred.output.dir will be the SUCCESS files



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (FLINK-1837) Throw an exceptions for iterative streaming programs with checkpointing enabled

2015-04-07 Thread JIRA
Márton Balassi created FLINK-1837:
-

 Summary: Throw an exceptions for iterative streaming programs with 
checkpointing enabled
 Key: FLINK-1837
 URL: https://issues.apache.org/jira/browse/FLINK-1837
 Project: Flink
  Issue Type: Improvement
  Components: Streaming
Affects Versions: 0.9
Reporter: Márton Balassi
Assignee: Márton Balassi


Currently streaming state checkpointing does not properly support iterative 
programs. The user should be clearly notified of this behaviour.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1834) Is mapred.output.dir conf parameter really required?

2015-04-07 Thread Fabian Hueske (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1834?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14483142#comment-14483142
 ] 

Fabian Hueske commented on FLINK-1834:
--

I would be careful with this. Do we know where we should default to?
If an OutputFormat does not require an output path we do not need to set it at 
all and if it does, it should be set by the user, IMO.

 Is mapred.output.dir conf parameter really required?
 

 Key: FLINK-1834
 URL: https://issues.apache.org/jira/browse/FLINK-1834
 Project: Flink
  Issue Type: Improvement
  Components: Hadoop Compatibility
Affects Versions: 0.9
Reporter: Flavio Pompermaier
Priority: Minor
  Labels: hadoop

 At the moment if you don't set mapred.output.dir parameter when output to an 
 hadoop sink Flink throws an error. Probably this parameter could be defaulted 
 to some tmp dir + jobId (for each output sink).
 For example, when output to an HBase sink, the only things you'll find in the 
 mapred.output.dir will be the SUCCESS files



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: FLINK-1560 - Add ITCases for streaming example...

2015-04-07 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/flink/pull/519


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-1820) Bug in DoubleParser and FloatParser - empty String is not casted to 0

2015-04-07 Thread Stephan Ewen (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1820?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14483215#comment-14483215
 ] 

Stephan Ewen commented on FLINK-1820:
-

Given the discussion here, I agree with Fabian. An exception is easier to 
understand than strange results because of a sneaky zero inserted for empty 
fields...

 Bug in DoubleParser and FloatParser - empty String is not casted to 0
 -

 Key: FLINK-1820
 URL: https://issues.apache.org/jira/browse/FLINK-1820
 Project: Flink
  Issue Type: Bug
  Components: core
Affects Versions: 0.8.0, 0.9, 0.8.1
Reporter: Felix Neutatz
Assignee: Felix Neutatz
Priority: Critical
 Fix For: 0.9


 Hi,
 I found the bug, when I wanted to read a csv file, which had a line like:
 ||\n
 If I treat it as a Tuple2Long,Long, I get as expected a tuple (0L,0L).
 But if I want to read it into a Double-Tuple or a Float-Tuple, I get the 
 following error:
 java.lang.AssertionError: Test failed due to a 
 org.apache.flink.api.common.io.ParseException: Line could not be parsed: '||'
 ParserError NUMERIC_VALUE_FORMAT_ERROR 
 This error can be solved by adding an additional condition for empty strings 
 in the FloatParser / DoubleParser.
 We definitely need the CSVReader to be able to read empty values.
 I can fix it like described if there are no better ideas :)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1834) Is mapred.output.dir conf parameter really required?

2015-04-07 Thread Maximilian Michels (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1834?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14483125#comment-14483125
 ] 

Maximilian Michels commented on FLINK-1834:
---

Let's check if the user's Hadoop configuration contains {{mapred.output.dir}} 
via {{HadoopUtils.getHadoopConfiguration()}}. Otherwise, set a default with a 
warning which also contains the output path to which we default.

Do you want to fix that?

 Is mapred.output.dir conf parameter really required?
 

 Key: FLINK-1834
 URL: https://issues.apache.org/jira/browse/FLINK-1834
 Project: Flink
  Issue Type: Improvement
  Components: Hadoop Compatibility
Affects Versions: 0.9
Reporter: Flavio Pompermaier
Priority: Minor
  Labels: hadoop

 At the moment if you don't set mapred.output.dir parameter when output to an 
 hadoop sink Flink throws an error. Probably this parameter could be defaulted 
 to some tmp dir + jobId (for each output sink).
 For example, when output to an HBase sink, the only things you'll find in the 
 mapred.output.dir will be the SUCCESS files



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (FLINK-1817) ClassLoaderObjectInputStream fails with ClassNotFoundException for primitive classes

2015-04-07 Thread Fabian Hueske (JIRA)

 [ 
https://issues.apache.org/jira/browse/FLINK-1817?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Fabian Hueske resolved FLINK-1817.
--
Resolution: Fixed

Fixed with 
- 0.9: d33b44549d416c30f01ead9aacf5cf7ed30674ca
- 0.8: a8dee2f0facae5fac8fe1ddf320ff7d377021313

 ClassLoaderObjectInputStream fails with ClassNotFoundException for primitive 
 classes
 

 Key: FLINK-1817
 URL: https://issues.apache.org/jira/browse/FLINK-1817
 Project: Flink
  Issue Type: Bug
  Components: core
Affects Versions: 0.9, 0.8.1
Reporter: Fabian Hueske
Assignee: Fabian Hueske
 Fix For: 0.9, 0.8.2


 User function objects that contain primitive type classes (such as 
 {{int.class}} or {{long.class}}) are not supported because the 
 {{ClassLoaderObjectInputStream}} fails with a {{ClassNotFoundException}} when 
 reading the object from the {{TaskConfig}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1820) Bug in DoubleParser and FloatParser - empty String is not casted to 0

2015-04-07 Thread Maximilian Michels (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1820?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14483253#comment-14483253
 ] 

Maximilian Michels commented on FLINK-1820:
---

+1 for a exception message for an empty String

 Bug in DoubleParser and FloatParser - empty String is not casted to 0
 -

 Key: FLINK-1820
 URL: https://issues.apache.org/jira/browse/FLINK-1820
 Project: Flink
  Issue Type: Bug
  Components: core
Affects Versions: 0.8.0, 0.9, 0.8.1
Reporter: Felix Neutatz
Assignee: Felix Neutatz
Priority: Critical
 Fix For: 0.9


 Hi,
 I found the bug, when I wanted to read a csv file, which had a line like:
 ||\n
 If I treat it as a Tuple2Long,Long, I get as expected a tuple (0L,0L).
 But if I want to read it into a Double-Tuple or a Float-Tuple, I get the 
 following error:
 java.lang.AssertionError: Test failed due to a 
 org.apache.flink.api.common.io.ParseException: Line could not be parsed: '||'
 ParserError NUMERIC_VALUE_FORMAT_ERROR 
 This error can be solved by adding an additional condition for empty strings 
 in the FloatParser / DoubleParser.
 We definitely need the CSVReader to be able to read empty values.
 I can fix it like described if there are no better ideas :)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1839) Failures in TwitterStreamITCase

2015-04-07 Thread Stephan Ewen (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1839?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14483441#comment-14483441
 ] 

Stephan Ewen commented on FLINK-1839:
-

Okay, from my latest Travis logs, this seems to happen in one out of three runs 
on average...

 Failures in TwitterStreamITCase
 ---

 Key: FLINK-1839
 URL: https://issues.apache.org/jira/browse/FLINK-1839
 Project: Flink
  Issue Type: Bug
  Components: Streaming
Affects Versions: 0.9
Reporter: Stephan Ewen
Priority: Critical

 The test seems unstable and fails occasionally with the error below
 {code}
 Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 4.508 sec  
 FAILURE! - in 
 org.apache.flink.streaming.examples.test.twitter.TwitterStreamITCase
 testJobWithoutObjectReuse(org.apache.flink.streaming.examples.test.twitter.TwitterStreamITCase)
   Time elapsed: 3.739 sec   FAILURE!
 java.lang.AssertionError: Different number of lines in expected and obtained 
 result. expected:146 but was:144
   at org.junit.Assert.fail(Assert.java:88)
   at org.junit.Assert.failNotEquals(Assert.java:743)
   at org.junit.Assert.assertEquals(Assert.java:118)
   at org.junit.Assert.assertEquals(Assert.java:555)
   at 
 org.apache.flink.test.util.TestBaseUtils.compareResultsByLinesInMemory(TestBaseUtils.java:256)
   at 
 org.apache.flink.test.util.TestBaseUtils.compareResultsByLinesInMemory(TestBaseUtils.java:242)
   at 
 org.apache.flink.streaming.examples.test.twitter.TwitterStreamITCase.postSubmit(TwitterStreamITCase.java:34)
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (FLINK-1839) Failures in TwitterStreamITCase

2015-04-07 Thread Stephan Ewen (JIRA)
Stephan Ewen created FLINK-1839:
---

 Summary: Failures in TwitterStreamITCase
 Key: FLINK-1839
 URL: https://issues.apache.org/jira/browse/FLINK-1839
 Project: Flink
  Issue Type: Bug
  Components: Streaming
Affects Versions: 0.9
Reporter: Stephan Ewen


The test seems unstable and fails occasionally with the error below

{code}
Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 4.508 sec  
FAILURE! - in 
org.apache.flink.streaming.examples.test.twitter.TwitterStreamITCase
testJobWithoutObjectReuse(org.apache.flink.streaming.examples.test.twitter.TwitterStreamITCase)
  Time elapsed: 3.739 sec   FAILURE!
java.lang.AssertionError: Different number of lines in expected and obtained 
result. expected:146 but was:144
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:743)
at org.junit.Assert.assertEquals(Assert.java:118)
at org.junit.Assert.assertEquals(Assert.java:555)
at 
org.apache.flink.test.util.TestBaseUtils.compareResultsByLinesInMemory(TestBaseUtils.java:256)
at 
org.apache.flink.test.util.TestBaseUtils.compareResultsByLinesInMemory(TestBaseUtils.java:242)
at 
org.apache.flink.streaming.examples.test.twitter.TwitterStreamITCase.postSubmit(TwitterStreamITCase.java:34)
{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1839) Failures in TwitterStreamITCase

2015-04-07 Thread Stephan Ewen (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1839?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14483440#comment-14483440
 ] 

Stephan Ewen commented on FLINK-1839:
-

Another occurance

{code}
Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 5.448 sec  
FAILURE! - in 
org.apache.flink.streaming.examples.test.twitter.TwitterStreamITCase
testJobWithoutObjectReuse(org.apache.flink.streaming.examples.test.twitter.TwitterStreamITCase)
  Time elapsed: 4.745 sec   FAILURE!
java.lang.AssertionError: Different number of lines in expected and obtained 
result. expected:146 but was:160
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:743)
at org.junit.Assert.assertEquals(Assert.java:118)
at org.junit.Assert.assertEquals(Assert.java:555)
at 
org.apache.flink.test.util.TestBaseUtils.compareResultsByLinesInMemory(TestBaseUtils.java:256)
at 
org.apache.flink.test.util.TestBaseUtils.compareResultsByLinesInMemory(TestBaseUtils.java:242)
at 
org.apache.flink.streaming.examples.test.twitter.TwitterStreamITCase.postSubmit(TwitterStreamITCase.java:34)
{code}


 Failures in TwitterStreamITCase
 ---

 Key: FLINK-1839
 URL: https://issues.apache.org/jira/browse/FLINK-1839
 Project: Flink
  Issue Type: Bug
  Components: Streaming
Affects Versions: 0.9
Reporter: Stephan Ewen

 The test seems unstable and fails occasionally with the error below
 {code}
 Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 4.508 sec  
 FAILURE! - in 
 org.apache.flink.streaming.examples.test.twitter.TwitterStreamITCase
 testJobWithoutObjectReuse(org.apache.flink.streaming.examples.test.twitter.TwitterStreamITCase)
   Time elapsed: 3.739 sec   FAILURE!
 java.lang.AssertionError: Different number of lines in expected and obtained 
 result. expected:146 but was:144
   at org.junit.Assert.fail(Assert.java:88)
   at org.junit.Assert.failNotEquals(Assert.java:743)
   at org.junit.Assert.assertEquals(Assert.java:118)
   at org.junit.Assert.assertEquals(Assert.java:555)
   at 
 org.apache.flink.test.util.TestBaseUtils.compareResultsByLinesInMemory(TestBaseUtils.java:256)
   at 
 org.apache.flink.test.util.TestBaseUtils.compareResultsByLinesInMemory(TestBaseUtils.java:242)
   at 
 org.apache.flink.streaming.examples.test.twitter.TwitterStreamITCase.postSubmit(TwitterStreamITCase.java:34)
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (FLINK-1839) Failures in TwitterStreamITCase

2015-04-07 Thread Stephan Ewen (JIRA)

 [ 
https://issues.apache.org/jira/browse/FLINK-1839?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen updated FLINK-1839:

Priority: Critical  (was: Major)

 Failures in TwitterStreamITCase
 ---

 Key: FLINK-1839
 URL: https://issues.apache.org/jira/browse/FLINK-1839
 Project: Flink
  Issue Type: Bug
  Components: Streaming
Affects Versions: 0.9
Reporter: Stephan Ewen
Priority: Critical

 The test seems unstable and fails occasionally with the error below
 {code}
 Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 4.508 sec  
 FAILURE! - in 
 org.apache.flink.streaming.examples.test.twitter.TwitterStreamITCase
 testJobWithoutObjectReuse(org.apache.flink.streaming.examples.test.twitter.TwitterStreamITCase)
   Time elapsed: 3.739 sec   FAILURE!
 java.lang.AssertionError: Different number of lines in expected and obtained 
 result. expected:146 but was:144
   at org.junit.Assert.fail(Assert.java:88)
   at org.junit.Assert.failNotEquals(Assert.java:743)
   at org.junit.Assert.assertEquals(Assert.java:118)
   at org.junit.Assert.assertEquals(Assert.java:555)
   at 
 org.apache.flink.test.util.TestBaseUtils.compareResultsByLinesInMemory(TestBaseUtils.java:256)
   at 
 org.apache.flink.test.util.TestBaseUtils.compareResultsByLinesInMemory(TestBaseUtils.java:242)
   at 
 org.apache.flink.streaming.examples.test.twitter.TwitterStreamITCase.postSubmit(TwitterStreamITCase.java:34)
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1319) Add static code analysis for UDFs

2015-04-07 Thread Timo Walther (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1319?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14483480#comment-14483480
 ] 

Timo Walther commented on FLINK-1319:
-

I have integrated the analyzer in the FlatMapOperator overriding the 
getSemanticProperties() method:

https://github.com/twalthr/flink/blob/a01b8334795d530b28d25f6bb20b09fca5c3cf27/flink-java/src/main/java/org/apache/flink/api/java/operators/FlatMapOperator.java

Do you think that's the right way to go?
(before I implement that in all operators ;) )

 Add static code analysis for UDFs
 -

 Key: FLINK-1319
 URL: https://issues.apache.org/jira/browse/FLINK-1319
 Project: Flink
  Issue Type: New Feature
  Components: Java API, Scala API
Reporter: Stephan Ewen
Assignee: Timo Walther
Priority: Minor

 Flink's Optimizer takes information that tells it for UDFs which fields of 
 the input elements are accessed, modified, or frwarded/copied. This 
 information frequently helps to reuse partitionings, sorts, etc. It may speed 
 up programs significantly, as it can frequently eliminate sorts and shuffles, 
 which are costly.
 Right now, users can add lightweight annotations to UDFs to provide this 
 information (such as adding {{@ConstandFields(0-3, 1, 2-1)}}.
 We worked with static code analysis of UDFs before, to determine this 
 information automatically. This is an incredible feature, as it magically 
 makes programs faster.
 For record-at-a-time operations (Map, Reduce, FlatMap, Join, Cross), this 
 works surprisingly well in many cases. We used the Soot toolkit for the 
 static code analysis. Unfortunately, Soot is LGPL licensed and thus we did 
 not include any of the code so far.
 I propose to add this functionality to Flink, in the form of a drop-in 
 addition, to work around the LGPL incompatibility with ALS 2.0. Users could 
 simply download a special flink-code-analysis.jar and drop it into the 
 lib folder to enable this functionality. We may even add a script to 
 tools that downloads that library automatically into the lib folder. This 
 should be legally fine, since we do not redistribute LGPL code and only 
 dynamically link it (the incompatibility with ASL 2.0 is mainly in the 
 patentability, if I remember correctly).
 Prior work on this has been done by [~aljoscha] and [~skunert], which could 
 provide a code base to start with.
 *Appendix*
 Hompage to Soot static analysis toolkit: http://www.sable.mcgill.ca/soot/
 Papers on static analysis and for optimization: 
 http://stratosphere.eu/assets/papers/EnablingOperatorReorderingSCA_12.pdf and 
 http://stratosphere.eu/assets/papers/openingTheBlackBoxes_12.pdf
 Quick introduction to the Optimizer: 
 http://stratosphere.eu/assets/papers/2014-VLDBJ_Stratosphere_Overview.pdf 
 (Section 6)
 Optimizer for Iterations: 
 http://stratosphere.eu/assets/papers/spinningFastIterativeDataFlows_12.pdf 
 (Sections 4.3 and 5.3)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Closed] (FLINK-1837) Throw an exceptions for iterative streaming programs with checkpointing enabled

2015-04-07 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/FLINK-1837?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Márton Balassi closed FLINK-1837.
-
   Resolution: Implemented
Fix Version/s: 0.9

Implemented via 52ebb29

 Throw an exceptions for iterative streaming programs with checkpointing 
 enabled
 ---

 Key: FLINK-1837
 URL: https://issues.apache.org/jira/browse/FLINK-1837
 Project: Flink
  Issue Type: Improvement
  Components: Streaming
Affects Versions: 0.9
Reporter: Márton Balassi
Assignee: Márton Balassi
 Fix For: 0.9


 Currently streaming state checkpointing does not properly support iterative 
 programs. The user should be clearly notified of this behaviour.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: Fix permissions on various non-executable sour...

2015-04-07 Thread hsaputra
Github user hsaputra commented on the pull request:

https://github.com/apache/flink/pull/574#issuecomment-90660629
  
Ship it :+1: 


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] flink pull request: [FLINK-951] Reworking of Iteration Synchroniza...

2015-04-07 Thread markus-h
Github user markus-h commented on the pull request:

https://github.com/apache/flink/pull/570#issuecomment-90694363
  
There seems to be a race condition somewhere in my code but I have trouble 
finding it since I can not reproduce it locally. I thought my last change would 
fix it but it didn't.
So if somebody has some free time and knows a bit about race conditions 
feel free to help me :-)


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Assigned] (FLINK-1839) Failures in TwitterStreamITCase

2015-04-07 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/FLINK-1839?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Márton Balassi reassigned FLINK-1839:
-

Assignee: Márton Balassi

 Failures in TwitterStreamITCase
 ---

 Key: FLINK-1839
 URL: https://issues.apache.org/jira/browse/FLINK-1839
 Project: Flink
  Issue Type: Bug
  Components: Streaming
Affects Versions: 0.9
Reporter: Stephan Ewen
Assignee: Márton Balassi
Priority: Critical

 The test seems unstable and fails occasionally with the error below
 {code}
 Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 4.508 sec  
 FAILURE! - in 
 org.apache.flink.streaming.examples.test.twitter.TwitterStreamITCase
 testJobWithoutObjectReuse(org.apache.flink.streaming.examples.test.twitter.TwitterStreamITCase)
   Time elapsed: 3.739 sec   FAILURE!
 java.lang.AssertionError: Different number of lines in expected and obtained 
 result. expected:146 but was:144
   at org.junit.Assert.fail(Assert.java:88)
   at org.junit.Assert.failNotEquals(Assert.java:743)
   at org.junit.Assert.assertEquals(Assert.java:118)
   at org.junit.Assert.assertEquals(Assert.java:555)
   at 
 org.apache.flink.test.util.TestBaseUtils.compareResultsByLinesInMemory(TestBaseUtils.java:256)
   at 
 org.apache.flink.test.util.TestBaseUtils.compareResultsByLinesInMemory(TestBaseUtils.java:242)
   at 
 org.apache.flink.streaming.examples.test.twitter.TwitterStreamITCase.postSubmit(TwitterStreamITCase.java:34)
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1839) Failures in TwitterStreamITCase

2015-04-07 Thread JIRA

[ 
https://issues.apache.org/jira/browse/FLINK-1839?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14483537#comment-14483537
 ] 

Márton Balassi commented on FLINK-1839:
---

My bad, I thought I have fixed it already. Thanks for the report.

 Failures in TwitterStreamITCase
 ---

 Key: FLINK-1839
 URL: https://issues.apache.org/jira/browse/FLINK-1839
 Project: Flink
  Issue Type: Bug
  Components: Streaming
Affects Versions: 0.9
Reporter: Stephan Ewen
Priority: Critical

 The test seems unstable and fails occasionally with the error below
 {code}
 Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 4.508 sec  
 FAILURE! - in 
 org.apache.flink.streaming.examples.test.twitter.TwitterStreamITCase
 testJobWithoutObjectReuse(org.apache.flink.streaming.examples.test.twitter.TwitterStreamITCase)
   Time elapsed: 3.739 sec   FAILURE!
 java.lang.AssertionError: Different number of lines in expected and obtained 
 result. expected:146 but was:144
   at org.junit.Assert.fail(Assert.java:88)
   at org.junit.Assert.failNotEquals(Assert.java:743)
   at org.junit.Assert.assertEquals(Assert.java:118)
   at org.junit.Assert.assertEquals(Assert.java:555)
   at 
 org.apache.flink.test.util.TestBaseUtils.compareResultsByLinesInMemory(TestBaseUtils.java:256)
   at 
 org.apache.flink.test.util.TestBaseUtils.compareResultsByLinesInMemory(TestBaseUtils.java:242)
   at 
 org.apache.flink.streaming.examples.test.twitter.TwitterStreamITCase.postSubmit(TwitterStreamITCase.java:34)
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-951) Reworking of Iteration Synchronization, Accumulators and Aggregators

2015-04-07 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-951?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14483752#comment-14483752
 ] 

ASF GitHub Bot commented on FLINK-951:
--

Github user markus-h commented on the pull request:

https://github.com/apache/flink/pull/570#issuecomment-90694363
  
There seems to be a race condition somewhere in my code but I have trouble 
finding it since I can not reproduce it locally. I thought my last change would 
fix it but it didn't.
So if somebody has some free time and knows a bit about race conditions 
feel free to help me :-)


 Reworking of Iteration Synchronization, Accumulators and Aggregators
 

 Key: FLINK-951
 URL: https://issues.apache.org/jira/browse/FLINK-951
 Project: Flink
  Issue Type: Improvement
  Components: Iterations, Optimizer
Affects Versions: 0.9
Reporter: Markus Holzemer
Assignee: Markus Holzemer
  Labels: refactoring
   Original Estimate: 168h
  Remaining Estimate: 168h

 I just realized that there is no real Jira issue for the task I am currently 
 working on. 
 I am currently reworking a few things regarding Iteration Synchronization, 
 Accumulators and Aggregators. Currently the synchronization at the end of one 
 superstep is done through channel events. That makes it hard to track the 
 current status of iterations. That is why I am changing this synchronization 
 to use RPC calls with the JobManager, so that the JobManager manages the 
 current status of all iterations.
 Currently we use Accumulators outside of iterations and Aggregators inside of 
 iterations. Both have a similiar function, but a bit different interfaces and 
 handling. I want to unify these two concepts. I propose that we stick in the 
 future to Accumulators only. Aggregators therefore are removed and 
 Accumulators are extended to cover the usecases Aggregators were used fore 
 before. The switch to RPC for iterations makes it possible to also send the 
 current Accumulator values at the end of each superstep, so that the 
 JobManager (and thereby the webinterface) will be able to print intermediate 
 accumulation results.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: Fix permissions on various non-executable sour...

2015-04-07 Thread StephanEwen
GitHub user StephanEwen opened a pull request:

https://github.com/apache/flink/pull/574

Fix permissions on various non-executable source files

Not a critical fix, but threw off the search for 3rd party executables as 
part of my pre-release code checking...

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/StephanEwen/incubator-flink permissions

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/574.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #574


commit 469256458c51ccf9035c4d0b6083c072d9858452
Author: Stephan Ewen se...@apache.org
Date:   2015-04-07T17:09:02Z

Remove 'executable' permission from various non-executable source files




---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Created] (FLINK-1840) Job execution fails on Windows (native and Cygwin)

2015-04-07 Thread Fabian Hueske (JIRA)
Fabian Hueske created FLINK-1840:


 Summary: Job execution fails on Windows (native and Cygwin)
 Key: FLINK-1840
 URL: https://issues.apache.org/jira/browse/FLINK-1840
 Project: Flink
  Issue Type: Bug
  Components: JobManager
Affects Versions: 0.9
 Environment: Windows 8.1 (Cygwin and native)
Reporter: Fabian Hueske
Priority: Critical
 Fix For: 0.9


Job execution fails on Windows using 
- {{./bin/flink.bat}} (native)
- {{./bin/flink.sh}} (Cygwin)
- Web submission client

The following exception is thrown:

{code}
org.apache.flink.client.program.ProgramInvocationException: Could not upload 
the program's JAR files tothe JobManager.
at org.apache.flink.client.program.Client.run(Client.java:341)
at org.apache.flink.client.program.Client.run(Client.java:313)
at org.apache.flink.client.program.Client.run(Client.java:306)
at 
org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:63)
at 
org.apache.flink.examples.java.wordcount.WordCount.main(WordCount.java:82)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
at java.lang.reflect.Method.invoke(Unknown Source)
at 
org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:437)
at 
org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedPro
   gram.java:353)
at org.apache.flink.client.program.Client.run(Client.java:273)
at 
org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:584)
at org.apache.flink.client.CliFrontend.run(CliFrontend.java:290)
at 
org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:886)
at org.apache.flink.client.CliFrontend.main(CliFrontend.java:928)
Caused by: java.io.IOException: PUT operation failed: Server side error: Cannot 
move staging file C:\cyg   
win64\tmp\blobStore-17f49f61-bcb7-45ba-909e-56beca159fe4\incoming\temp- 
to BLOB file C:\cygwin64   
\tmp\blobStore-17f49f61-bcb7-45ba-909e-56beca159fe4\cache\blob_9d5dc35b7c40b2ddda2df12d7ee8702f5683b131
at 
org.apache.flink.runtime.blob.BlobClient.putInputStream(BlobClient.java:465)
at org.apache.flink.runtime.blob.BlobClient.put(BlobClient.java:327)
at 
org.apache.flink.runtime.jobgraph.JobGraph.uploadRequiredJarFiles(JobGraph.java:457)
at 
org.apache.flink.runtime.client.JobClient$.uploadJarFiles(JobClient.scala:288)
at 
org.apache.flink.runtime.client.JobClient.uploadJarFiles(JobClient.scala)
at org.apache.flink.client.program.Client.run(Client.java:338)
... 15 more
Caused by: java.io.IOException: Server side error: Cannot move staging file 
C:\cygwin64\tmp\blobStore-17   
f49f61-bcb7-45ba-909e-56beca159fe4\incoming\temp- to BLOB file 
C:\cygwin64\tmp\blobStore-17f49f6   
1-bcb7-45ba-909e-56beca159fe4\cache\blob_9d5dc35b7c40b2ddda2df12d7ee8702f5683b131
at 
org.apache.flink.runtime.blob.BlobClient.receivePutResponseAndCompare(BlobClient.java:491)
at 
org.apache.flink.runtime.blob.BlobClient.putInputStream(BlobClient.java:461)
... 20 more
Caused by: java.io.IOException: Cannot move staging file 
C:\cygwin64\tmp\blobStore-17f49f61-bcb7-45ba-90   
9e-56beca159fe4\incoming\temp- to BLOB file 
C:\cygwin64\tmp\blobStore-17f49f61-bcb7-45ba-909e-56   
beca159fe4\cache\blob_9d5dc35b7c40b2ddda2df12d7ee8702f5683b131
at 
org.apache.flink.runtime.blob.BlobServerConnection.put(BlobServerConnection.java:325)
at 
org.apache.flink.runtime.blob.BlobServerConnection.run(BlobServerConnection.java:109)
{code} 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1758) Extend Gelly's neighborhood methods

2015-04-07 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1758?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14484000#comment-14484000
 ] 

ASF GitHub Bot commented on FLINK-1758:
---

GitHub user andralungu opened a pull request:

https://github.com/apache/flink/pull/576

[FLINK-1758][gelly] Neighborhood Methods Extensions



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/andralungu/flink extendNeighborhood

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/576.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #576


commit e5f6571c7cb05d96aac84530beb52cef968243d7
Author: andralungu lungu.an...@gmail.com
Date:   2015-04-07T20:29:18Z

[FLINK-1758][gelly] Neighborhood Methods Extensions




 Extend Gelly's neighborhood methods
 ---

 Key: FLINK-1758
 URL: https://issues.apache.org/jira/browse/FLINK-1758
 Project: Flink
  Issue Type: Improvement
  Components: Gelly
Affects Versions: 0.9
Reporter: Vasia Kalavri
Assignee: Andra Lungu

 Currently, the neighborhood methods only allow returning a single value per 
 vertex. In many cases, it is desirable to return several or no value per 
 vertex. This is the case in clustering coefficient computation, 
 vertex-centric jaccard, algorithms where a vertex computes a value per edge 
 or when a vertex computes a value only for some of its neighbors.
 This issue proposes to 
 - change the current reduceOnEdges/reduceOnNeighbors methods to use 
 combinable reduce operations where possible
 - provide groupReduce-versions, which will use a Collector and allow 
 returning none or more values per vertex.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: [FLINK-1758][gelly] Neighborhood Methods Exten...

2015-04-07 Thread andralungu
GitHub user andralungu opened a pull request:

https://github.com/apache/flink/pull/576

[FLINK-1758][gelly] Neighborhood Methods Extensions



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/andralungu/flink extendNeighborhood

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/576.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #576


commit e5f6571c7cb05d96aac84530beb52cef968243d7
Author: andralungu lungu.an...@gmail.com
Date:   2015-04-07T20:29:18Z

[FLINK-1758][gelly] Neighborhood Methods Extensions




---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-1832) start-local.bat/start-local.sh does not work if there is a white space in the file path (windows)

2015-04-07 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1832?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14483878#comment-14483878
 ] 

ASF GitHub Bot commented on FLINK-1832:
---

GitHub user fhueske opened a pull request:

https://github.com/apache/flink/pull/575

[FLINK-1832] Fix start scripts for Cygwin and paths with spaces

A few fixes for the start and submission scripts. 
Right now they fail for:
- all scripts fail on Cygwin
- `./bin/flink` fails on Unix for Flink root paths that contain spaces

Also, some duplicated code is removed.

Fixes should be backported to 0.8 as well.

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/fhueske/flink cygwin

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/575.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #575


commit dc9af9249c8fc1c8817ed73c90c0fa048130c7ba
Author: Maximilian Michels m...@apache.org
Date:   2015-04-07T10:56:12Z

[FLINK-1832] [scripts] fix conversion to Windows path with spaces

commit f20b4827c0d1fb49fd5c6773eeb05ac5540eab4b
Author: Fabian Hueske fhue...@apache.org
Date:   2015-04-07T18:59:44Z

[FLINK-1832] [scripts] fix start scripts for Cygwin and paths with spaces.




 start-local.bat/start-local.sh does not work if there is a white space in the 
 file path (windows)
 -

 Key: FLINK-1832
 URL: https://issues.apache.org/jira/browse/FLINK-1832
 Project: Flink
  Issue Type: Bug
  Components: Local Runtime
Affects Versions: 0.8.1
 Environment: Windows 8.1
Reporter: Nikolaas Steenbergen
Priority: Minor

 start-local.bat can be run, but exits immediately.
 start-local.sh with cygwin exits with error:
 owner_@owner /cygdrive/e/Program 
 Files/flink-0.8.1-bin-hadoop2/flink-0.8.1-bin-hadoop2/flink-0.8.1/bin
 $ ./start-local.sh
 cygpath: error converting 
 Files/flink-0.8.1-bin-hadoop2/flink-0.8.1-bin-hadoop2/flink-0.8.1/bin/.. - 
 No such file or directory
 sed: can't read E:\Program/conf/flink-conf.yaml: No such file or directory
 sed: can't read E:\Program/conf/flink-conf.yaml: No such file or directory
 sed: can't read E:\Program/conf/flink-conf.yaml: No such file or directory
 sed: can't read E:\Program/conf/flink-conf.yaml: No such file or directory
 sed: can't read E:\Program/conf/flink-conf.yaml: No such file or directory
 sed: can't read E:\Program/conf/flink-conf.yaml: No such file or directory
 sed: can't read E:\Program/conf/flink-conf.yaml: No such file or directory
 cygpath: error converting 
 /cygdrive/e/Program:Files/flink-0.8.1-bin-hadoop2/flink-0.8.1-bin-hadoop2/flink-0.8.1/bin/../lib/*.jar:.
  - Unknown error -1
 ./start-local.sh: line 27: E:\Program/bin/jobmanager.sh: No such file or 
 directory



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1741) Add Jaccard Similarity Metric Example

2015-04-07 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1741?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14483936#comment-14483936
 ] 

ASF GitHub Bot commented on FLINK-1741:
---

Github user andralungu commented on the pull request:

https://github.com/apache/flink/pull/544#issuecomment-90715751
  
@vasia,

I updated the PR with your latest suggestions. Travis seems to fail because 
of flink-steaming :-s; my quick refactoring couldn't have triggered that :)


 Add Jaccard Similarity Metric Example
 -

 Key: FLINK-1741
 URL: https://issues.apache.org/jira/browse/FLINK-1741
 Project: Flink
  Issue Type: Task
  Components: Gelly
Affects Versions: 0.9
Reporter: Andra Lungu
Assignee: Andra Lungu

 http://www.inside-r.org/packages/cran/igraph/docs/similarity



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: [FLINK-1741][gelly] Adds Jaccard Similarity Me...

2015-04-07 Thread andralungu
Github user andralungu commented on the pull request:

https://github.com/apache/flink/pull/544#issuecomment-90715751
  
@vasia,

I updated the PR with your latest suggestions. Travis seems to fail because 
of flink-steaming :-s; my quick refactoring couldn't have triggered that :)


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] flink pull request: [FLINK-1832] Fix start scripts for Cygwin and ...

2015-04-07 Thread fhueske
GitHub user fhueske opened a pull request:

https://github.com/apache/flink/pull/575

[FLINK-1832] Fix start scripts for Cygwin and paths with spaces

A few fixes for the start and submission scripts. 
Right now they fail for:
- all scripts fail on Cygwin
- `./bin/flink` fails on Unix for Flink root paths that contain spaces

Also, some duplicated code is removed.

Fixes should be backported to 0.8 as well.

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/fhueske/flink cygwin

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/575.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #575


commit dc9af9249c8fc1c8817ed73c90c0fa048130c7ba
Author: Maximilian Michels m...@apache.org
Date:   2015-04-07T10:56:12Z

[FLINK-1832] [scripts] fix conversion to Windows path with spaces

commit f20b4827c0d1fb49fd5c6773eeb05ac5540eab4b
Author: Fabian Hueske fhue...@apache.org
Date:   2015-04-07T18:59:44Z

[FLINK-1832] [scripts] fix start scripts for Cygwin and paths with spaces.




---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] flink pull request: [FLINK-1523][gelly] Vertex centric iteration e...

2015-04-07 Thread andralungu
Github user andralungu commented on the pull request:

https://github.com/apache/flink/pull/537#issuecomment-90728415
  
Hi @vasia ,

I simply assumed that if the previous examples work, the iteration also 
works as before. Would you like a specific test suite for that?

For the degree option not being set : the getInDegree(); getOutDegree() 
methods are located inside the Vertex class. Not sure that their availability 
can be changed by setting the option on the iteration. They are not computed 
when the degree option is not set, but you can still see the method. Do you 
have any idea how to make the degrees inaccessible?

Once this becomes clear, I can add the tests :)


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-1523) Vertex-centric iteration extensions

2015-04-07 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1523?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14484019#comment-14484019
 ] 

ASF GitHub Bot commented on FLINK-1523:
---

Github user andralungu commented on the pull request:

https://github.com/apache/flink/pull/537#issuecomment-90728415
  
Hi @vasia ,

I simply assumed that if the previous examples work, the iteration also 
works as before. Would you like a specific test suite for that?

For the degree option not being set : the getInDegree(); getOutDegree() 
methods are located inside the Vertex class. Not sure that their availability 
can be changed by setting the option on the iteration. They are not computed 
when the degree option is not set, but you can still see the method. Do you 
have any idea how to make the degrees inaccessible?

Once this becomes clear, I can add the tests :)


 Vertex-centric iteration extensions
 ---

 Key: FLINK-1523
 URL: https://issues.apache.org/jira/browse/FLINK-1523
 Project: Flink
  Issue Type: Improvement
  Components: Gelly
Reporter: Vasia Kalavri
Assignee: Andra Lungu

 We would like to make the following extensions to the vertex-centric 
 iterations of Gelly:
 - allow vertices to access their in/out degrees and the total number of 
 vertices of the graph, inside the iteration.
 - allow choosing the neighborhood type (in/out/all) over which to run the 
 vertex-centric iteration. Now, the model uses the updates of the in-neighbors 
 to calculate state and send messages to out-neighbors. We could add a 
 parameter with value in/out/all to the {{VertexUpdateFunction}} and 
 {{MessagingFunction}}, that would indicate the type of neighborhood.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1820) Bug in DoubleParser and FloatParser - empty String is not casted to 0

2015-04-07 Thread Felix Neutatz (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1820?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14483973#comment-14483973
 ] 

Felix Neutatz commented on FLINK-1820:
--

ok, I agree with you. I will add the exceptions instead

 Bug in DoubleParser and FloatParser - empty String is not casted to 0
 -

 Key: FLINK-1820
 URL: https://issues.apache.org/jira/browse/FLINK-1820
 Project: Flink
  Issue Type: Bug
  Components: core
Affects Versions: 0.8.0, 0.9, 0.8.1
Reporter: Felix Neutatz
Assignee: Felix Neutatz
Priority: Critical
 Fix For: 0.9


 Hi,
 I found the bug, when I wanted to read a csv file, which had a line like:
 ||\n
 If I treat it as a Tuple2Long,Long, I get as expected a tuple (0L,0L).
 But if I want to read it into a Double-Tuple or a Float-Tuple, I get the 
 following error:
 java.lang.AssertionError: Test failed due to a 
 org.apache.flink.api.common.io.ParseException: Line could not be parsed: '||'
 ParserError NUMERIC_VALUE_FORMAT_ERROR 
 This error can be solved by adding an additional condition for empty strings 
 in the FloatParser / DoubleParser.
 We definitely need the CSVReader to be able to read empty values.
 I can fix it like described if there are no better ideas :)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1840) Job execution fails on Windows (native and Cygwin)

2015-04-07 Thread Fabian Hueske (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1840?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14484001#comment-14484001
 ] 

Fabian Hueske commented on FLINK-1840:
--

The exception is cause by the {{java.io.File.renameTo()}} method call. The 
JavaDocs say that the behavior is ??inherently platform-depended??.
Therefore, I tried to substitute the call by an alternative but neither 
{{com.google.common.io.Files.move()}} or {{java.nio.file.Files.move()}} worked. 
It seems like the source file is still used/accessed by a process and can 
therefore not be deleted. 

 Job execution fails on Windows (native and Cygwin)
 --

 Key: FLINK-1840
 URL: https://issues.apache.org/jira/browse/FLINK-1840
 Project: Flink
  Issue Type: Bug
  Components: JobManager
Affects Versions: 0.9
 Environment: Windows 8.1 (Cygwin and native)
Reporter: Fabian Hueske
Priority: Critical
 Fix For: 0.9


 Job execution fails on Windows using 
 - {{./bin/flink.bat}} (native)
 - {{./bin/flink.sh}} (Cygwin)
 - Web submission client
 The following exception is thrown:
 {code}
 org.apache.flink.client.program.ProgramInvocationException: Could not upload 
 the program's JAR files tothe JobManager.
 at org.apache.flink.client.program.Client.run(Client.java:341)
 at org.apache.flink.client.program.Client.run(Client.java:313)
 at org.apache.flink.client.program.Client.run(Client.java:306)
 at 
 org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:63)
 at 
 org.apache.flink.examples.java.wordcount.WordCount.main(WordCount.java:82)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
 at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
 at java.lang.reflect.Method.invoke(Unknown Source)
 at 
 org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:437)
 at 
 org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedPro
gram.java:353)
 at org.apache.flink.client.program.Client.run(Client.java:273)
 at 
 org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:584)
 at org.apache.flink.client.CliFrontend.run(CliFrontend.java:290)
 at 
 org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:886)
 at org.apache.flink.client.CliFrontend.main(CliFrontend.java:928)
 Caused by: java.io.IOException: PUT operation failed: Server side error: 
 Cannot move staging file C:\cyg   
 win64\tmp\blobStore-17f49f61-bcb7-45ba-909e-56beca159fe4\incoming\temp-
  to BLOB file C:\cygwin64   
 \tmp\blobStore-17f49f61-bcb7-45ba-909e-56beca159fe4\cache\blob_9d5dc35b7c40b2ddda2df12d7ee8702f5683b131
 at 
 org.apache.flink.runtime.blob.BlobClient.putInputStream(BlobClient.java:465)
 at org.apache.flink.runtime.blob.BlobClient.put(BlobClient.java:327)
 at 
 org.apache.flink.runtime.jobgraph.JobGraph.uploadRequiredJarFiles(JobGraph.java:457)
 at 
 org.apache.flink.runtime.client.JobClient$.uploadJarFiles(JobClient.scala:288)
 at 
 org.apache.flink.runtime.client.JobClient.uploadJarFiles(JobClient.scala)
 at org.apache.flink.client.program.Client.run(Client.java:338)
 ... 15 more
 Caused by: java.io.IOException: Server side error: Cannot move staging file 
 C:\cygwin64\tmp\blobStore-17   
 f49f61-bcb7-45ba-909e-56beca159fe4\incoming\temp- to BLOB file 
 C:\cygwin64\tmp\blobStore-17f49f6   
 1-bcb7-45ba-909e-56beca159fe4\cache\blob_9d5dc35b7c40b2ddda2df12d7ee8702f5683b131
 at 
 org.apache.flink.runtime.blob.BlobClient.receivePutResponseAndCompare(BlobClient.java:491)
 at 
 org.apache.flink.runtime.blob.BlobClient.putInputStream(BlobClient.java:461)
 ... 20 more
 Caused by: java.io.IOException: Cannot move staging file 
 C:\cygwin64\tmp\blobStore-17f49f61-bcb7-45ba-90   
 9e-56beca159fe4\incoming\temp- to BLOB file 
 C:\cygwin64\tmp\blobStore-17f49f61-bcb7-45ba-909e-56   
 beca159fe4\cache\blob_9d5dc35b7c40b2ddda2df12d7ee8702f5683b131
 at 
 org.apache.flink.runtime.blob.BlobServerConnection.put(BlobServerConnection.java:325)
 at 
 org.apache.flink.runtime.blob.BlobServerConnection.run(BlobServerConnection.java:109)
 {code} 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1758) Extend Gelly's neighborhood methods

2015-04-07 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1758?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14484226#comment-14484226
 ] 

ASF GitHub Bot commented on FLINK-1758:
---

Github user vasia commented on the pull request:

https://github.com/apache/flink/pull/576#issuecomment-90749788
  
Thanks for the PR @andralungu!

The idea for this issue is to (a) use a combinable reduce for the simple 
(without vertex value) reduceOn* methods, because it will be more efficient and 
(b) offer groupReduceOn* versions with Collectors to allow returning more than 
one value per vertex.
I see you have implemented (b) here, but not (a).

To be more specific, the groupReduceOn* methods would be the ones with the 
Collectors. Basically, you'll just have to rename what you have already 
implemented. The reduceOn* methods will only return one value per vertex, like 
they do currently, but internally use the reduce operator (instead of 
reduceGroup).

For example, if you want to compute the sum of edge values, you can use the 
`reduceOnEdges` method, same as you currently do. And if you want to compute 
the Jaccard coefficient, you need to output multiple values per vertex and for 
that you can use `groupReduceOnEdges`.

Please let me know if you have any questions! Thanks again :-)


 Extend Gelly's neighborhood methods
 ---

 Key: FLINK-1758
 URL: https://issues.apache.org/jira/browse/FLINK-1758
 Project: Flink
  Issue Type: Improvement
  Components: Gelly
Affects Versions: 0.9
Reporter: Vasia Kalavri
Assignee: Andra Lungu

 Currently, the neighborhood methods only allow returning a single value per 
 vertex. In many cases, it is desirable to return several or no value per 
 vertex. This is the case in clustering coefficient computation, 
 vertex-centric jaccard, algorithms where a vertex computes a value per edge 
 or when a vertex computes a value only for some of its neighbors.
 This issue proposes to 
 - change the current reduceOnEdges/reduceOnNeighbors methods to use 
 combinable reduce operations where possible
 - provide groupReduce-versions, which will use a Collector and allow 
 returning none or more values per vertex.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (FLINK-1841) WindowJoinITCase fails

2015-04-07 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/FLINK-1841?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Márton Balassi reassigned FLINK-1841:
-

Assignee: Márton Balassi

 WindowJoinITCase fails
 --

 Key: FLINK-1841
 URL: https://issues.apache.org/jira/browse/FLINK-1841
 Project: Flink
  Issue Type: Bug
  Components: Streaming
Affects Versions: 0.9
Reporter: Fabian Hueske
Assignee: Márton Balassi

 The WindowJoinITCase failed in a Travis build: 
 https://travis-ci.org/fhueske/flink/jobs/57547875
 {code}
 Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 5.54 sec  
 FAILURE! - in org.apache.flink.streaming.examples.test.join.WindowJoinITCase
 testJobWithoutObjectReuse(org.apache.flink.streaming.examples.test.join.WindowJoinITCase)
  Time elapsed: 4.638 sec  FAILURE!
 java.lang.AssertionError: Different number of lines in expected and obtained 
 result. expected:17 but was:24
 at org.junit.Assert.fail(Assert.java:88)
 at org.junit.Assert.failNotEquals(Assert.java:743)
 at org.junit.Assert.assertEquals(Assert.java:118)
 at org.junit.Assert.assertEquals(Assert.java:555)
 at 
 org.apache.flink.test.util.TestBaseUtils.compareResultsByLinesInMemory(TestBaseUtils.java:256)
 at 
 org.apache.flink.test.util.TestBaseUtils.compareResultsByLinesInMemory(TestBaseUtils.java:242)
 at 
 org.apache.flink.streaming.examples.test.join.WindowJoinITCase.postSubmit(WindowJoinITCase.java:41)
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1840) Job execution fails on Windows (native and Cygwin)

2015-04-07 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1840?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14484129#comment-14484129
 ] 

ASF GitHub Bot commented on FLINK-1840:
---

GitHub user fhueske opened a pull request:

https://github.com/apache/flink/pull/578

[FLINK-1840] Fix BLOB manager on Windows

- closes open `FileOutputStream` before the field is moved
- replaces `java.io.File.renameTo()` by `com.google.common.io.Files.move()` 
to avoid platform-dependent behavior

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/fhueske/flink blobWindows

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/578.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #578


commit b83ab465ba727751bc13db26989c40a94ffe378b
Author: Fabian Hueske fhue...@apache.org
Date:   2015-04-07T21:04:59Z

[FLINK-1840] Fix BLOB manager on Windows




 Job execution fails on Windows (native and Cygwin)
 --

 Key: FLINK-1840
 URL: https://issues.apache.org/jira/browse/FLINK-1840
 Project: Flink
  Issue Type: Bug
  Components: JobManager
Affects Versions: 0.9
 Environment: Windows 8.1 (Cygwin and native)
Reporter: Fabian Hueske
Assignee: Fabian Hueske
Priority: Critical
 Fix For: 0.9


 Job execution fails on Windows using 
 - {{./bin/flink.bat}} (native)
 - {{./bin/flink.sh}} (Cygwin)
 - Web submission client
 The following exception is thrown:
 {code}
 org.apache.flink.client.program.ProgramInvocationException: Could not upload 
 the program's JAR files tothe JobManager.
 at org.apache.flink.client.program.Client.run(Client.java:341)
 at org.apache.flink.client.program.Client.run(Client.java:313)
 at org.apache.flink.client.program.Client.run(Client.java:306)
 at 
 org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:63)
 at 
 org.apache.flink.examples.java.wordcount.WordCount.main(WordCount.java:82)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at sun.reflect.NativeMethodAccessorImpl.invoke(Unknown Source)
 at sun.reflect.DelegatingMethodAccessorImpl.invoke(Unknown Source)
 at java.lang.reflect.Method.invoke(Unknown Source)
 at 
 org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:437)
 at 
 org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedPro
gram.java:353)
 at org.apache.flink.client.program.Client.run(Client.java:273)
 at 
 org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:584)
 at org.apache.flink.client.CliFrontend.run(CliFrontend.java:290)
 at 
 org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:886)
 at org.apache.flink.client.CliFrontend.main(CliFrontend.java:928)
 Caused by: java.io.IOException: PUT operation failed: Server side error: 
 Cannot move staging file C:\cyg   
 win64\tmp\blobStore-17f49f61-bcb7-45ba-909e-56beca159fe4\incoming\temp-
  to BLOB file C:\cygwin64   
 \tmp\blobStore-17f49f61-bcb7-45ba-909e-56beca159fe4\cache\blob_9d5dc35b7c40b2ddda2df12d7ee8702f5683b131
 at 
 org.apache.flink.runtime.blob.BlobClient.putInputStream(BlobClient.java:465)
 at org.apache.flink.runtime.blob.BlobClient.put(BlobClient.java:327)
 at 
 org.apache.flink.runtime.jobgraph.JobGraph.uploadRequiredJarFiles(JobGraph.java:457)
 at 
 org.apache.flink.runtime.client.JobClient$.uploadJarFiles(JobClient.scala:288)
 at 
 org.apache.flink.runtime.client.JobClient.uploadJarFiles(JobClient.scala)
 at org.apache.flink.client.program.Client.run(Client.java:338)
 ... 15 more
 Caused by: java.io.IOException: Server side error: Cannot move staging file 
 C:\cygwin64\tmp\blobStore-17   
 f49f61-bcb7-45ba-909e-56beca159fe4\incoming\temp- to BLOB file 
 C:\cygwin64\tmp\blobStore-17f49f6   
 1-bcb7-45ba-909e-56beca159fe4\cache\blob_9d5dc35b7c40b2ddda2df12d7ee8702f5683b131
 at 
 org.apache.flink.runtime.blob.BlobClient.receivePutResponseAndCompare(BlobClient.java:491)
 at 
 org.apache.flink.runtime.blob.BlobClient.putInputStream(BlobClient.java:461)
 ... 20 more
 Caused by: java.io.IOException: Cannot move staging file 
 C:\cygwin64\tmp\blobStore-17f49f61-bcb7-45ba-90   
 9e-56beca159fe4\incoming\temp- to BLOB file 
 

[GitHub] flink pull request: [FLINK-1840] Fix BLOB manager on Windows

2015-04-07 Thread fhueske
GitHub user fhueske opened a pull request:

https://github.com/apache/flink/pull/578

[FLINK-1840] Fix BLOB manager on Windows

- closes open `FileOutputStream` before the field is moved
- replaces `java.io.File.renameTo()` by `com.google.common.io.Files.move()` 
to avoid platform-dependent behavior

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/fhueske/flink blobWindows

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/578.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #578


commit b83ab465ba727751bc13db26989c40a94ffe378b
Author: Fabian Hueske fhue...@apache.org
Date:   2015-04-07T21:04:59Z

[FLINK-1840] Fix BLOB manager on Windows




---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] flink pull request: Fix permissions on various non-executable sour...

2015-04-07 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/flink/pull/574


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Created] (FLINK-1841) WindowJoinITCase fails

2015-04-07 Thread Fabian Hueske (JIRA)
Fabian Hueske created FLINK-1841:


 Summary: WindowJoinITCase fails
 Key: FLINK-1841
 URL: https://issues.apache.org/jira/browse/FLINK-1841
 Project: Flink
  Issue Type: Bug
  Components: Streaming
Affects Versions: 0.9
Reporter: Fabian Hueske


The WindowJoinITCase failed in a Travis build: 
https://travis-ci.org/fhueske/flink/jobs/57547875

{code}
Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 5.54 sec  
FAILURE! - in org.apache.flink.streaming.examples.test.join.WindowJoinITCase

testJobWithoutObjectReuse(org.apache.flink.streaming.examples.test.join.WindowJoinITCase)
 Time elapsed: 4.638 sec  FAILURE!

java.lang.AssertionError: Different number of lines in expected and obtained 
result. expected:17 but was:24
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:743)
at org.junit.Assert.assertEquals(Assert.java:118)
at org.junit.Assert.assertEquals(Assert.java:555)
at 
org.apache.flink.test.util.TestBaseUtils.compareResultsByLinesInMemory(TestBaseUtils.java:256)
at 
org.apache.flink.test.util.TestBaseUtils.compareResultsByLinesInMemory(TestBaseUtils.java:242)
at 
org.apache.flink.streaming.examples.test.join.WindowJoinITCase.postSubmit(WindowJoinITCase.java:41)
{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: [FLINK-1813] Avoid illegalstate exception when...

2015-04-07 Thread gyfora
GitHub user gyfora opened a pull request:

https://github.com/apache/flink/pull/577

[FLINK-1813] Avoid illegalstate exception when trying to broadcast on 
finished operators



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/mbalassi/flink FLINK-1813

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/577.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #577


commit e3810484c6fa71f9c1dac69707c81e44380a2d1c
Author: Gyula Fora gyf...@apache.org
Date:   2015-04-07T20:17:18Z

[FLINK-1813] Avoid illegalstate exception when trying to broadcast on 
finished operators




---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-1813) Barrier broadcast sometimes causes illegal state exception in source RecordWriter

2015-04-07 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1813?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14484082#comment-14484082
 ] 

ASF GitHub Bot commented on FLINK-1813:
---

GitHub user gyfora opened a pull request:

https://github.com/apache/flink/pull/577

[FLINK-1813] Avoid illegalstate exception when trying to broadcast on 
finished operators



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/mbalassi/flink FLINK-1813

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/577.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #577


commit e3810484c6fa71f9c1dac69707c81e44380a2d1c
Author: Gyula Fora gyf...@apache.org
Date:   2015-04-07T20:17:18Z

[FLINK-1813] Avoid illegalstate exception when trying to broadcast on 
finished operators




 Barrier broadcast sometimes causes illegal state exception in source 
 RecordWriter
 -

 Key: FLINK-1813
 URL: https://issues.apache.org/jira/browse/FLINK-1813
 Project: Flink
  Issue Type: Bug
  Components: Streaming
Reporter: Gyula Fora
Assignee: Gyula Fora

 The StreamCheckPointingITCase sometimes fails on an illegal state exception 
 thrown when trying to broadcast barrier from the sources. I think this is 
 caused by sending a barrier when the source is not running any more.
 The exception:
 Exception in thread Thread-341 java.lang.RuntimeException: 
 java.lang.IllegalStateException: Partition already finished.
   at 
 org.apache.flink.streaming.api.streamvertex.StreamVertex.actOnBarrier(StreamVertex.java:300)
   at 
 org.apache.flink.streaming.api.streamvertex.StreamVertex.broadcastBarrierFromSource(StreamVertex.java:105)
   at 
 org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$receiveWithLogMessages$1$$anon$4.run(TaskManager.scala:386)
   at java.lang.Thread.run(Thread.java:745)
 Caused by: java.lang.IllegalStateException: Partition already finished.
   at 
 org.apache.flink.shaded.com.google.common.base.Preconditions.checkState(Preconditions.java:173)
   at 
 org.apache.flink.runtime.io.network.partition.ResultPartition.checkInProduceState(ResultPartition.java:388)
   at 
 org.apache.flink.runtime.io.network.partition.ResultPartition.add(ResultPartition.java:221)
   at 
 org.apache.flink.runtime.io.network.api.writer.ResultPartitionWriter.writeEvent(ResultPartitionWriter.java:78)
   at 
 org.apache.flink.runtime.io.network.api.writer.RecordWriter.broadcastEvent(RecordWriter.java:117)
   at 
 org.apache.flink.streaming.api.collector.StreamOutput.broadcastEvent(StreamOutput.java:92)
   at 
 org.apache.flink.streaming.api.streamvertex.OutputHandler.broadcastBarrier(OutputHandler.java:94)
   at 
 org.apache.flink.streaming.api.streamvertex.StreamVertex.actOnBarrier(StreamVertex.java:294)
   ... 3 more



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: [FLINK-1694] [gelly] added IterationConfigurat...

2015-04-07 Thread vasia
Github user vasia commented on the pull request:

https://github.com/apache/flink/pull/547#issuecomment-90745094
  
Hi,
any opinions on this? 
I would like to merge it soon because it changes the api and vertex-centric 
related issues depend on it.
Thanks!


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] flink pull request: [FLINK-1741][gelly] Adds Jaccard Similarity Me...

2015-04-07 Thread vasia
Github user vasia commented on the pull request:

https://github.com/apache/flink/pull/544#issuecomment-90740868
  
Thank you @andralungu! This looks great now ^^


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-1694) Change the split between create/run of a vertex-centric iteration

2015-04-07 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1694?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14484173#comment-14484173
 ] 

ASF GitHub Bot commented on FLINK-1694:
---

Github user vasia commented on the pull request:

https://github.com/apache/flink/pull/547#issuecomment-90745094
  
Hi,
any opinions on this? 
I would like to merge it soon because it changes the api and vertex-centric 
related issues depend on it.
Thanks!


 Change the split between create/run of a vertex-centric iteration
 -

 Key: FLINK-1694
 URL: https://issues.apache.org/jira/browse/FLINK-1694
 Project: Flink
  Issue Type: Improvement
  Components: Gelly
Reporter: Vasia Kalavri
Assignee: Vasia Kalavri

 Currently, the vertex-centric API in Gelly looks like this:
 {code:java}
 Graph inputGaph = ... //create graph
 VertexCentricIteration iteration = inputGraph.createVertexCentricIteration();
 ... // configure the iteration
 Graph newGraph = inputGaph.runVertexCentricIteration(iteration);
 {code}
 We have this create/run split, in order to expose the iteration object and be 
 able to call the public methods of VertexCentricIteration.
 However, this is not very nice and might lead to errors, if create and run 
 are  mistakenly called on different graph objects.
 One suggestion is to change this to the following:
 {code:java}
 VertexCentricIteration iteration = inputGraph.createVertexCentricIteration();
 ... // configure the iteration
 Graph newGraph = iteration.result();
 {code}
 or to go with a single run call, where we add an IterationConfiguration 
 object as a parameter and we don't expose the iteration object to the user at 
 all:
 {code:java}
 IterationConfiguration parameters  = ...
 Graph newGraph = inputGraph.runVertexCentricIteration(parameters);
 {code}
 and we can also have a simplified method where no configuration is passed.
 What do you think?
 Personally, I like the second option a bit more.
 -Vasia.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: [FLINK-1758][gelly] Neighborhood Methods Exten...

2015-04-07 Thread vasia
Github user vasia commented on the pull request:

https://github.com/apache/flink/pull/576#issuecomment-90749788
  
Thanks for the PR @andralungu!

The idea for this issue is to (a) use a combinable reduce for the simple 
(without vertex value) reduceOn* methods, because it will be more efficient and 
(b) offer groupReduceOn* versions with Collectors to allow returning more than 
one value per vertex.
I see you have implemented (b) here, but not (a).

To be more specific, the groupReduceOn* methods would be the ones with the 
Collectors. Basically, you'll just have to rename what you have already 
implemented. The reduceOn* methods will only return one value per vertex, like 
they do currently, but internally use the reduce operator (instead of 
reduceGroup).

For example, if you want to compute the sum of edge values, you can use the 
`reduceOnEdges` method, same as you currently do. And if you want to compute 
the Jaccard coefficient, you need to output multiple values per vertex and for 
that you can use `groupReduceOnEdges`.

Please let me know if you have any questions! Thanks again :-)


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-1744) Change the reference of slaves to workers to match the description of the system

2015-04-07 Thread Henry Saputra (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1744?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14484648#comment-14484648
 ] 

Henry Saputra commented on FLINK-1744:
--

No worries, we could always re-open it when needed.

 Change the reference of slaves to workers to match the description of the 
 system
 

 Key: FLINK-1744
 URL: https://issues.apache.org/jira/browse/FLINK-1744
 Project: Flink
  Issue Type: Improvement
  Components: core, Documentation
Reporter: Henry Saputra
Priority: Trivial

 There are some references to slaves which actually mean workers.
 Need to change it to use workers whenever possible, unless it is needed when 
 communicating with external system like Apache Hadoop



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)