[jira] [Created] (FLINK-1560) Add ITCases for streaming examples

2015-02-16 Thread JIRA
Márton Balassi created FLINK-1560:
-

 Summary: Add ITCases for streaming examples
 Key: FLINK-1560
 URL: https://issues.apache.org/jira/browse/FLINK-1560
 Project: Flink
  Issue Type: Test
  Components: Streaming
Affects Versions: 0.9
Reporter: Márton Balassi
Assignee: Márton Balassi


Currently there are no tests for consistency of the streaming example programs. 
This might be a real show stopper for users who encounter an issue there.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-947) Add support for "Named Datasets"

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-947?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14323372#comment-14323372
 ] 

ASF GitHub Bot commented on FLINK-947:
--

Github user mbalassi commented on the pull request:

https://github.com/apache/flink/pull/405#issuecomment-74579541
  
Great additions! Really looking to trying it out. :)


> Add support for "Named Datasets"
> 
>
> Key: FLINK-947
> URL: https://issues.apache.org/jira/browse/FLINK-947
> Project: Flink
>  Issue Type: New Feature
>  Components: Java API
>Reporter: Aljoscha Krettek
>Assignee: Aljoscha Krettek
>Priority: Minor
>
> This would create an API that is a mix between SQL like declarativity and the 
> power of user defined functions. Example user code could look like this:
> {code:Java}
> NamedDataSet one = ...
> NamedDataSet two = ...
> NamedDataSet result = one.join(two).where("key").equalTo("otherKey")
>   .project("a", "b", "c")
>   .map( (UserTypeIn in) -> return new UserTypeOut(...) )
>   .print();
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-947) Add support for "Named Datasets"

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-947?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14323370#comment-14323370
 ] 

ASF GitHub Bot commented on FLINK-947:
--

Github user mbalassi commented on a diff in the pull request:

https://github.com/apache/flink/pull/405#discussion_r24778932
  
--- Diff: flink-staging/flink-linq/pom.xml ---
@@ -0,0 +1,227 @@
+
+
+http://maven.apache.org/POM/4.0.0"; 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance";
+   xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 
http://maven.apache.org/maven-v4_0_0.xsd";>
+
+   4.0.0
+
+   
+   org.apache.flink
+   flink-staging
+   0.9-SNAPSHOT
+   ..
+   
+
+   flink-linq
+   flink-linq
+
+   jar
+
+   
+
+   
+   org.apache.flink
+   flink-scala
+   ${project.version}
+   
+
+   
+   org.apache.flink
+   flink-streaming-scala
+   ${project.version}
+   
+
+   
+   org.apache.flink
+   flink-scala-examples
+   ${project.version}
+   
+
+   
+   org.scala-lang
+   scala-reflect
+   
+
+   
+   org.scala-lang
+   scala-library
+   
+
+   
+   org.scala-lang
+   scala-compiler
+   
+
--- End diff --

You transitively depend on the scala stuff through flink-scala, so you 
could omit these.


> Add support for "Named Datasets"
> 
>
> Key: FLINK-947
> URL: https://issues.apache.org/jira/browse/FLINK-947
> Project: Flink
>  Issue Type: New Feature
>  Components: Java API
>Reporter: Aljoscha Krettek
>Assignee: Aljoscha Krettek
>Priority: Minor
>
> This would create an API that is a mix between SQL like declarativity and the 
> power of user defined functions. Example user code could look like this:
> {code:Java}
> NamedDataSet one = ...
> NamedDataSet two = ...
> NamedDataSet result = one.join(two).where("key").equalTo("otherKey")
>   .project("a", "b", "c")
>   .map( (UserTypeIn in) -> return new UserTypeOut(...) )
>   .print();
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: [FLINK-947] Add a declarative expression API

2015-02-16 Thread mbalassi
Github user mbalassi commented on a diff in the pull request:

https://github.com/apache/flink/pull/405#discussion_r24778932
  
--- Diff: flink-staging/flink-linq/pom.xml ---
@@ -0,0 +1,227 @@
+
+
+http://maven.apache.org/POM/4.0.0"; 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance";
+   xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 
http://maven.apache.org/maven-v4_0_0.xsd";>
+
+   4.0.0
+
+   
+   org.apache.flink
+   flink-staging
+   0.9-SNAPSHOT
+   ..
+   
+
+   flink-linq
+   flink-linq
+
+   jar
+
+   
+
+   
+   org.apache.flink
+   flink-scala
+   ${project.version}
+   
+
+   
+   org.apache.flink
+   flink-streaming-scala
+   ${project.version}
+   
+
+   
+   org.apache.flink
+   flink-scala-examples
+   ${project.version}
+   
+
+   
+   org.scala-lang
+   scala-reflect
+   
+
+   
+   org.scala-lang
+   scala-library
+   
+
+   
+   org.scala-lang
+   scala-compiler
+   
+
--- End diff --

You transitively depend on the scala stuff through flink-scala, so you 
could omit these.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] flink pull request: [FLINK-947] Add a declarative expression API

2015-02-16 Thread mbalassi
Github user mbalassi commented on the pull request:

https://github.com/apache/flink/pull/405#issuecomment-74579541
  
Great additions! Really looking to trying it out. :)


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] flink pull request: [FLINK-1466] Add HCatInputFormats to read from...

2015-02-16 Thread fhueske
GitHub user fhueske opened a pull request:

https://github.com/apache/flink/pull/411

[FLINK-1466] Add HCatInputFormats to read from HCatalog tables.

Right now the Flink-tuple mode is restricted to primitive types (no ARRAY, 
STRUCT, and MAP types) and the max Java/Scala tuple width.
The default HCatRecord mode works for all types and any tuple width.

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/fhueske/flink hcatalog

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/411.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #411


commit 61d8c3e2c7948768f60bbb20755176aa8ef14fcd
Author: Fabian Hueske 
Date:   2015-01-29T09:34:29Z

[FLINK-1466] Adds HCatInputFormats to read from HCatalog tables.




---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-1466) Add InputFormat to read HCatalog tables

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1466?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14323344#comment-14323344
 ] 

ASF GitHub Bot commented on FLINK-1466:
---

GitHub user fhueske opened a pull request:

https://github.com/apache/flink/pull/411

[FLINK-1466] Add HCatInputFormats to read from HCatalog tables.

Right now the Flink-tuple mode is restricted to primitive types (no ARRAY, 
STRUCT, and MAP types) and the max Java/Scala tuple width.
The default HCatRecord mode works for all types and any tuple width.

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/fhueske/flink hcatalog

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/411.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #411


commit 61d8c3e2c7948768f60bbb20755176aa8ef14fcd
Author: Fabian Hueske 
Date:   2015-01-29T09:34:29Z

[FLINK-1466] Adds HCatInputFormats to read from HCatalog tables.




> Add InputFormat to read HCatalog tables
> ---
>
> Key: FLINK-1466
> URL: https://issues.apache.org/jira/browse/FLINK-1466
> Project: Flink
>  Issue Type: New Feature
>  Components: Java API, Scala API
>Reporter: Fabian Hueske
>Assignee: Fabian Hueske
>Priority: Minor
>
> HCatalog is a metadata repository and InputFormat to make Hive tables 
> accessible to other frameworks such as Pig.
> Adding support for HCatalog would give access to Hive managed data.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: Add auto-parallelism to Jobs (0.8 branch)

2015-02-16 Thread StephanEwen
GitHub user StephanEwen opened a pull request:

https://github.com/apache/flink/pull/410

Add auto-parallelism to Jobs (0.8 branch)



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/StephanEwen/incubator-flink autopar

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/410.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #410


commit 923d1b4309c10a86cfa8ea3c385ff751c59e29a4
Author: Stephan Ewen 
Date:   2015-02-16T20:40:06Z

Add autoparallelism to jobs




---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-785) Add Chained operators for AllReduce and AllGroupReduce

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-785?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14323225#comment-14323225
 ] 

ASF GitHub Bot commented on FLINK-785:
--

Github user StephanEwen commented on the pull request:

https://github.com/apache/flink/pull/370#issuecomment-74565386
  
I think we should use the chained reducer from this pull request and drop 
the chained combiner change.


> Add Chained operators for AllReduce and AllGroupReduce
> --
>
> Key: FLINK-785
> URL: https://issues.apache.org/jira/browse/FLINK-785
> Project: Flink
>  Issue Type: Improvement
>Reporter: GitHub Import
>Assignee: Chesnay Schepler
>  Labels: github-import
> Fix For: pre-apache
>
>
> Because the operators `AllReduce` and `AllGroupReduce` are used both for the 
> pre-reduce (combiner side) and the final reduce, they would greatly benefit 
> from a chained version.
>  Imported from GitHub 
> Url: https://github.com/stratosphere/stratosphere/issues/785
> Created by: [StephanEwen|https://github.com/StephanEwen]
> Labels: runtime, 
> Milestone: Release 0.6 (unplanned)
> Created at: Sun May 11 17:41:12 CEST 2014
> State: open



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: [FLINK-1521] Chained operators respect reuse

2015-02-16 Thread StephanEwen
Github user StephanEwen commented on the pull request:

https://github.com/apache/flink/pull/392#issuecomment-74565272
  
At this point, we also have to worry about efficiency and performance.
This is also something that the user can easily mitigate in the UDF, at a 
fraction of the cost in most cases.

I vote postpone the fix until we get some tangible user feedback concerning 
this issue.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] flink pull request: [FLINK-785] Chained AllReduce / AllGroupReduce...

2015-02-16 Thread StephanEwen
Github user StephanEwen commented on the pull request:

https://github.com/apache/flink/pull/370#issuecomment-74565386
  
I think we should use the chained reducer from this pull request and drop 
the chained combiner change.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-1521) Some Chained Drivers do not respect object-reuse/non-reuse flag

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1521?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14323223#comment-14323223
 ] 

ASF GitHub Bot commented on FLINK-1521:
---

Github user StephanEwen commented on the pull request:

https://github.com/apache/flink/pull/392#issuecomment-74565272
  
At this point, we also have to worry about efficiency and performance.
This is also something that the user can easily mitigate in the UDF, at a 
fraction of the cost in most cases.

I vote postpone the fix until we get some tangible user feedback concerning 
this issue.


> Some Chained Drivers do not respect object-reuse/non-reuse flag
> ---
>
> Key: FLINK-1521
> URL: https://issues.apache.org/jira/browse/FLINK-1521
> Project: Flink
>  Issue Type: Bug
>Reporter: Aljoscha Krettek
>Assignee: Chesnay Schepler
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: [FLINK-1556] Corrects faulty JobClient behavio...

2015-02-16 Thread StephanEwen
Github user StephanEwen commented on the pull request:

https://github.com/apache/flink/pull/406#issuecomment-74565029
  
Looks good to me!


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] flink pull request: [FLINK-1557] Move JobManager web frontend serv...

2015-02-16 Thread StephanEwen
GitHub user StephanEwen opened a pull request:

https://github.com/apache/flink/pull/409

[FLINK-1557] Move JobManager web frontend server out of JobManager actor



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/StephanEwen/incubator-flink webserver

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/409.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #409


commit a24cbec30aa3ae5c715b3d41fc11a120537b802b
Author: Stephan Ewen 
Date:   2015-02-16T15:34:13Z

[FLINK-1557] Move JobManager web frontend server out of JobManager actor




---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-1556) JobClient does not wait until a job failed completely if submission exception

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1556?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14323219#comment-14323219
 ] 

ASF GitHub Bot commented on FLINK-1556:
---

Github user StephanEwen commented on the pull request:

https://github.com/apache/flink/pull/406#issuecomment-74565029
  
Looks good to me!


> JobClient does not wait until a job failed completely if submission exception
> -
>
> Key: FLINK-1556
> URL: https://issues.apache.org/jira/browse/FLINK-1556
> Project: Flink
>  Issue Type: Bug
>Reporter: Till Rohrmann
>
> If an exception occurs during job submission the {{JobClient}} received a 
> {{SubmissionFailure}}. Upon receiving this message, the {{JobClient}} 
> terminates itself and returns the error to the {{Client}}. This indicates to 
> the user that the job has been completely failed which is not necessarily 
> true. 
> If the user directly after such a failure submits another job, then it might 
> be the case that not all slots of the formerly failed job are returned. This 
> can lead to a {{NoRessourceAvailableException}}.
> We can solve this problem by waiting for the completion of the job failure in 
> the {{JobClient}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1557) Move WebFrontend out of JobManager Actor

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1557?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14323218#comment-14323218
 ] 

ASF GitHub Bot commented on FLINK-1557:
---

GitHub user StephanEwen opened a pull request:

https://github.com/apache/flink/pull/409

[FLINK-1557] Move JobManager web frontend server out of JobManager actor



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/StephanEwen/incubator-flink webserver

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/409.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #409


commit a24cbec30aa3ae5c715b3d41fc11a120537b802b
Author: Stephan Ewen 
Date:   2015-02-16T15:34:13Z

[FLINK-1557] Move JobManager web frontend server out of JobManager actor




> Move WebFrontend out of JobManager Actor
> 
>
> Key: FLINK-1557
> URL: https://issues.apache.org/jira/browse/FLINK-1557
> Project: Flink
>  Issue Type: Improvement
>  Components: JobManager
>Affects Versions: 0.9
>Reporter: Stephan Ewen
>Assignee: Stephan Ewen
> Fix For: 0.9
>
>
> Currently, the Webfrontent is started inside the JobManager actor.
> I propose to move it out of the actor, since it is a heavy component that is 
> independent of the actor and whose failure cannot be controlled by the job 
> manager.
> The change is fairly straightforward, since the web frontend only 
> communicates with the jobmanager through actor messages.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (FLINK-1542) Test case at BlobUtilsTest should not assume user could not create new item in root directory

2015-02-16 Thread Stephan Ewen (JIRA)

 [ 
https://issues.apache.org/jira/browse/FLINK-1542?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen resolved FLINK-1542.
-
   Resolution: Fixed
Fix Version/s: 0.9

Fixed via 5ab1f1f197b4ab86382e7e20b4964e991663b62c

> Test case at BlobUtilsTest should not assume user could not create new item 
> in root directory
> -
>
> Key: FLINK-1542
> URL: https://issues.apache.org/jira/browse/FLINK-1542
> Project: Flink
>  Issue Type: Bug
>Reporter: Henry Saputra
>Assignee: Henry Saputra
> Fix For: 0.9
>
>
> The test at BlobUtilsTest #testExceptionOnCreateStorageDirectoryFailure 
> assume that current user cannot create new directory at root.
> In my MacOSX, my user has admin group so the test will always fail.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1542) Test case at BlobUtilsTest should not assume user could not create new item in root directory

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1542?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14323208#comment-14323208
 ] 

ASF GitHub Bot commented on FLINK-1542:
---

Github user asfgit closed the pull request at:

https://github.com/apache/flink/pull/396


> Test case at BlobUtilsTest should not assume user could not create new item 
> in root directory
> -
>
> Key: FLINK-1542
> URL: https://issues.apache.org/jira/browse/FLINK-1542
> Project: Flink
>  Issue Type: Bug
>Reporter: Henry Saputra
>Assignee: Henry Saputra
> Fix For: 0.9
>
>
> The test at BlobUtilsTest #testExceptionOnCreateStorageDirectoryFailure 
> assume that current user cannot create new directory at root.
> In my MacOSX, my user has admin group so the test will always fail.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: [FLINK-1542] Test case at BlobUtilsTest should...

2015-02-16 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/flink/pull/396


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-1454) CliFrontend blocks for 100 seconds when submitting to a non-existent JobManager

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1454?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14323205#comment-14323205
 ] 

ASF GitHub Bot commented on FLINK-1454:
---

Github user StephanEwen closed the pull request at:

https://github.com/apache/flink/pull/404


> CliFrontend blocks for 100 seconds when submitting to a non-existent 
> JobManager
> ---
>
> Key: FLINK-1454
> URL: https://issues.apache.org/jira/browse/FLINK-1454
> Project: Flink
>  Issue Type: Bug
>  Components: JobManager
>Affects Versions: 0.9
>Reporter: Robert Metzger
>Assignee: Stephan Ewen
> Fix For: 0.9
>
>
> When a user tries to submit a job to a job manager which doesn't exist at 
> all, the CliFrontend blocks for 100 seconds.
> Ideally, Akka would fail because it can not connect to the given 
> hostname:port.
>  
> {code}
> ./bin/flink run ./examples/flink-java-examples-0.9-SNAPSHOT-WordCount.jar -c 
> foo.Baz
> org.apache.flink.client.program.ProgramInvocationException: The main method 
> caused an error.
>   at 
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:449)
>   at 
> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:350)
>   at org.apache.flink.client.program.Client.run(Client.java:242)
>   at 
> org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:389)
>   at org.apache.flink.client.CliFrontend.run(CliFrontend.java:362)
>   at 
> org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:1078)
>   at org.apache.flink.client.CliFrontend.main(CliFrontend.java:1102)
> Caused by: java.util.concurrent.TimeoutException: Futures timed out after 
> [100 seconds]
>   at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>   at 
> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
>   at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
>   at 
> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
>   at scala.concurrent.Await$.result(package.scala:107)
>   at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
>   at 
> org.apache.flink.runtime.client.JobClient$.uploadJarFiles(JobClient.scala:169)
>   at 
> org.apache.flink.runtime.client.JobClient.uploadJarFiles(JobClient.scala)
>   at org.apache.flink.client.program.Client.run(Client.java:314)
>   at org.apache.flink.client.program.Client.run(Client.java:296)
>   at org.apache.flink.client.program.Client.run(Client.java:290)
>   at 
> org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:55)
>   at 
> org.apache.flink.examples.java.wordcount.WordCount.main(WordCount.java:82)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:606)
>   at 
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:434)
>   ... 6 more
> The exception above occurred while trying to run your command.
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1454) CliFrontend blocks for 100 seconds when submitting to a non-existent JobManager

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1454?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14323204#comment-14323204
 ] 

ASF GitHub Bot commented on FLINK-1454:
---

Github user StephanEwen commented on the pull request:

https://github.com/apache/flink/pull/404#issuecomment-74563335
  
Manually merged in b941cf2d091bb38b9a45e1f2412136acae2b0f3f


> CliFrontend blocks for 100 seconds when submitting to a non-existent 
> JobManager
> ---
>
> Key: FLINK-1454
> URL: https://issues.apache.org/jira/browse/FLINK-1454
> Project: Flink
>  Issue Type: Bug
>  Components: JobManager
>Affects Versions: 0.9
>Reporter: Robert Metzger
>Assignee: Stephan Ewen
> Fix For: 0.9
>
>
> When a user tries to submit a job to a job manager which doesn't exist at 
> all, the CliFrontend blocks for 100 seconds.
> Ideally, Akka would fail because it can not connect to the given 
> hostname:port.
>  
> {code}
> ./bin/flink run ./examples/flink-java-examples-0.9-SNAPSHOT-WordCount.jar -c 
> foo.Baz
> org.apache.flink.client.program.ProgramInvocationException: The main method 
> caused an error.
>   at 
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:449)
>   at 
> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:350)
>   at org.apache.flink.client.program.Client.run(Client.java:242)
>   at 
> org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:389)
>   at org.apache.flink.client.CliFrontend.run(CliFrontend.java:362)
>   at 
> org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:1078)
>   at org.apache.flink.client.CliFrontend.main(CliFrontend.java:1102)
> Caused by: java.util.concurrent.TimeoutException: Futures timed out after 
> [100 seconds]
>   at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>   at 
> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
>   at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
>   at 
> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
>   at scala.concurrent.Await$.result(package.scala:107)
>   at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
>   at 
> org.apache.flink.runtime.client.JobClient$.uploadJarFiles(JobClient.scala:169)
>   at 
> org.apache.flink.runtime.client.JobClient.uploadJarFiles(JobClient.scala)
>   at org.apache.flink.client.program.Client.run(Client.java:314)
>   at org.apache.flink.client.program.Client.run(Client.java:296)
>   at org.apache.flink.client.program.Client.run(Client.java:290)
>   at 
> org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:55)
>   at 
> org.apache.flink.examples.java.wordcount.WordCount.main(WordCount.java:82)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:606)
>   at 
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:434)
>   ... 6 more
> The exception above occurred while trying to run your command.
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: [FLINK-1454] [job client] Improve error handli...

2015-02-16 Thread StephanEwen
Github user StephanEwen commented on the pull request:

https://github.com/apache/flink/pull/404#issuecomment-74563335
  
Manually merged in b941cf2d091bb38b9a45e1f2412136acae2b0f3f


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] flink pull request: [FLINK-1454] [job client] Improve error handli...

2015-02-16 Thread StephanEwen
Github user StephanEwen closed the pull request at:

https://github.com/apache/flink/pull/404


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Resolved] (FLINK-1559) Normalize all Akka URLs to use IP addresses, not hostnames

2015-02-16 Thread Stephan Ewen (JIRA)

 [ 
https://issues.apache.org/jira/browse/FLINK-1559?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen resolved FLINK-1559.
-
Resolution: Fixed

Fixed via 2dcff4c122f37ca58a485e886fec9295fef0c832

> Normalize all Akka URLs to use IP addresses, not hostnames
> --
>
> Key: FLINK-1559
> URL: https://issues.apache.org/jira/browse/FLINK-1559
> Project: Flink
>  Issue Type: Bug
>  Components: Distributed Runtime
>Affects Versions: 0.9
>Reporter: Stephan Ewen
>Assignee: Stephan Ewen
> Fix For: 0.9
>
>
> Akka has the nasty property that it does not use IP addresses and hostnames 
> interchangeably. 
> Example: If the JobManager is started at 
> {{akka:tcp//flink@localhost:1234/user/JOBMANAGER}} and a job client actor 
> tries to send messages to {{akka:tcp//flink@127.0.0.1:1234/user/JOBMANAGER}}, 
> the job manager actor system will receive the messages, but throw them away, 
> because they are not destined for the actor system.
> I am not sure whether this is a bug or a feature in akka, but we need to deal 
> with it. To be on the safe side, I suggest to build all akka URLs with ip 
> addresses. We can still use host names in our logic (connection infos, local 
> assignments, etc)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (FLINK-1454) CliFrontend blocks for 100 seconds when submitting to a non-existent JobManager

2015-02-16 Thread Stephan Ewen (JIRA)

 [ 
https://issues.apache.org/jira/browse/FLINK-1454?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen resolved FLINK-1454.
-
   Resolution: Fixed
Fix Version/s: 0.9
 Assignee: Stephan Ewen

Fixed via b941cf2d091bb38b9a45e1f2412136acae2b0f3f

> CliFrontend blocks for 100 seconds when submitting to a non-existent 
> JobManager
> ---
>
> Key: FLINK-1454
> URL: https://issues.apache.org/jira/browse/FLINK-1454
> Project: Flink
>  Issue Type: Bug
>  Components: JobManager
>Affects Versions: 0.9
>Reporter: Robert Metzger
>Assignee: Stephan Ewen
> Fix For: 0.9
>
>
> When a user tries to submit a job to a job manager which doesn't exist at 
> all, the CliFrontend blocks for 100 seconds.
> Ideally, Akka would fail because it can not connect to the given 
> hostname:port.
>  
> {code}
> ./bin/flink run ./examples/flink-java-examples-0.9-SNAPSHOT-WordCount.jar -c 
> foo.Baz
> org.apache.flink.client.program.ProgramInvocationException: The main method 
> caused an error.
>   at 
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:449)
>   at 
> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:350)
>   at org.apache.flink.client.program.Client.run(Client.java:242)
>   at 
> org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:389)
>   at org.apache.flink.client.CliFrontend.run(CliFrontend.java:362)
>   at 
> org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:1078)
>   at org.apache.flink.client.CliFrontend.main(CliFrontend.java:1102)
> Caused by: java.util.concurrent.TimeoutException: Futures timed out after 
> [100 seconds]
>   at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>   at 
> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
>   at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
>   at 
> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
>   at scala.concurrent.Await$.result(package.scala:107)
>   at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
>   at 
> org.apache.flink.runtime.client.JobClient$.uploadJarFiles(JobClient.scala:169)
>   at 
> org.apache.flink.runtime.client.JobClient.uploadJarFiles(JobClient.scala)
>   at org.apache.flink.client.program.Client.run(Client.java:314)
>   at org.apache.flink.client.program.Client.run(Client.java:296)
>   at org.apache.flink.client.program.Client.run(Client.java:290)
>   at 
> org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:55)
>   at 
> org.apache.flink.examples.java.wordcount.WordCount.main(WordCount.java:82)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:606)
>   at 
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:434)
>   ... 6 more
> The exception above occurred while trying to run your command.
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: [FLINK-1542] Test case at BlobUtilsTest should...

2015-02-16 Thread hsaputra
Github user hsaputra commented on the pull request:

https://github.com/apache/flink/pull/396#issuecomment-74558015
  
Thanks for the review Ufuk and Stephan.

Also thx for Stephan for merging it.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-1542) Test case at BlobUtilsTest should not assume user could not create new item in root directory

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1542?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14323150#comment-14323150
 ] 

ASF GitHub Bot commented on FLINK-1542:
---

Github user hsaputra commented on the pull request:

https://github.com/apache/flink/pull/396#issuecomment-74558015
  
Thanks for the review Ufuk and Stephan.

Also thx for Stephan for merging it.


> Test case at BlobUtilsTest should not assume user could not create new item 
> in root directory
> -
>
> Key: FLINK-1542
> URL: https://issues.apache.org/jira/browse/FLINK-1542
> Project: Flink
>  Issue Type: Bug
>Reporter: Henry Saputra
>Assignee: Henry Saputra
>
> The test at BlobUtilsTest #testExceptionOnCreateStorageDirectoryFailure 
> assume that current user cannot create new directory at root.
> In my MacOSX, my user has admin group so the test will always fail.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1542) Test case at BlobUtilsTest should not assume user could not create new item in root directory

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1542?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14323147#comment-14323147
 ] 

ASF GitHub Bot commented on FLINK-1542:
---

Github user StephanEwen commented on the pull request:

https://github.com/apache/flink/pull/396#issuecomment-74556700
  
Looks good to me as well. Will merge it with the next batch...


> Test case at BlobUtilsTest should not assume user could not create new item 
> in root directory
> -
>
> Key: FLINK-1542
> URL: https://issues.apache.org/jira/browse/FLINK-1542
> Project: Flink
>  Issue Type: Bug
>Reporter: Henry Saputra
>Assignee: Henry Saputra
>
> The test at BlobUtilsTest #testExceptionOnCreateStorageDirectoryFailure 
> assume that current user cannot create new directory at root.
> In my MacOSX, my user has admin group so the test will always fail.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: [FLINK-1542] Test case at BlobUtilsTest should...

2015-02-16 Thread StephanEwen
Github user StephanEwen commented on the pull request:

https://github.com/apache/flink/pull/396#issuecomment-74556700
  
Looks good to me as well. Will merge it with the next batch...


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Created] (FLINK-1559) Normalize all Akka URLs to use IP addresses, not hostnames

2015-02-16 Thread Stephan Ewen (JIRA)
Stephan Ewen created FLINK-1559:
---

 Summary: Normalize all Akka URLs to use IP addresses, not hostnames
 Key: FLINK-1559
 URL: https://issues.apache.org/jira/browse/FLINK-1559
 Project: Flink
  Issue Type: Bug
  Components: Distributed Runtime
Affects Versions: 0.9
Reporter: Stephan Ewen
Assignee: Stephan Ewen
 Fix For: 0.9


Akka has the nasty property that it does not use IP addresses and hostnames 
interchangeably. 

Example: If the JobManager is started at 
{{akka:tcp//flink@localhost:1234/user/JOBMANAGER}} and a job client actor tries 
to send messages to {{akka:tcp//flink@127.0.0.1:1234/user/JOBMANAGER}}, the job 
manager actor system will receive the messages, but throw them away, because 
they are not destined for the actor system.

I am not sure whether this is a bug or a feature in akka, but we need to deal 
with it. To be on the safe side, I suggest to build all akka URLs with ip 
addresses. We can still use host names in our logic (connection infos, local 
assignments, etc)




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1514) [Gelly] Add a Gather-Sum-Apply iteration method

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1514?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14323032#comment-14323032
 ] 

ASF GitHub Bot commented on FLINK-1514:
---

Github user balidani commented on a diff in the pull request:

https://github.com/apache/flink/pull/408#discussion_r24763748
  
--- Diff: 
flink-staging/flink-gelly/src/test/java/org/apache/flink/graph/test/GatherSumApplyITCase.java
 ---
@@ -0,0 +1,212 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.graph.test;
+
+import org.apache.flink.api.common.functions.MapFunction;
+import org.apache.flink.api.java.ExecutionEnvironment;
+import org.apache.flink.api.java.tuple.Tuple2;
+import org.apache.flink.api.java.tuple.Tuple3;
+import org.apache.flink.graph.Edge;
+import org.apache.flink.graph.Graph;
+import org.apache.flink.graph.Vertex;
+import org.apache.flink.graph.gsa.ApplyFunction;
+import org.apache.flink.graph.gsa.GatherFunction;
+import org.apache.flink.graph.gsa.SumFunction;
+import org.apache.flink.test.util.MultipleProgramsTestBase;
+import org.apache.flink.util.Collector;
+import org.junit.After;
+import org.junit.Before;
+import org.junit.Rule;
+import org.junit.Test;
+import org.junit.rules.TemporaryFolder;
+import org.junit.runner.RunWith;
+import org.junit.runners.Parameterized;
+
+import java.util.ArrayList;
+import java.util.HashSet;
+import java.util.List;
+
+@RunWith(Parameterized.class)
+public class GatherSumApplyITCase extends MultipleProgramsTestBase {
+
+   public GatherSumApplyITCase(MultipleProgramsTestBase.ExecutionMode 
mode){
+   super(mode);
+   }
+
+   private String resultPath;
+   private String expectedResult;
+
+   @Rule
+   public TemporaryFolder tempFolder = new TemporaryFolder();
+
+   @Before
+   public void before() throws Exception{
+   resultPath = tempFolder.newFile().toURI().toString();
+   }
+
+   @After
+   public void after() throws Exception{
+   compareResultsByLinesInMemory(expectedResult, resultPath);
+   }
+
+   @Test
+   public void testGreedyGraphColoring() throws Exception {
+   /*
+* This test runs the Greedy Graph Coloring algorithm using a 
GSA iteration
+* The expected result is that the lowest vertex value (1) gets 
propagated to all vertices
+*/
+   final ExecutionEnvironment env = 
ExecutionEnvironment.getExecutionEnvironment();
+
+   Graph graph = 
Graph.fromDataSet(TestGraphUtils.getLongLongVertexData(env),
+   TestGraphUtils.getLongLongEdgeData(env), env);
+
+   // Gather the target vertices into a one-element set
+   GatherFunction>> 
gather =
+   new GatherFunction>>() {
+
+   @Override
+   public Tuple2>> gather(Tuple3,
+   Edge, 
Vertex> triplet) {
+
+   HashSet> 
result = new HashSet>();
+   result.add(triplet.f2);
+
+   return new Tuple2>>(triplet.f0.getId(), result);
+   }
+   };
+
+   // Merge the sets between neighbors
+   SumFunction>> sum =
+   new SumFunction>>() {
+
+   @Override
+   public Tuple2>> sum(
+   Tuple2>> arg0,
+   Tuple2>> arg1) {
+
+   HashSet> 
result = new HashSet>();
+  

[GitHub] flink pull request: [FLINK-1514][Gelly] Add a Gather-Sum-Apply ite...

2015-02-16 Thread balidani
Github user balidani commented on a diff in the pull request:

https://github.com/apache/flink/pull/408#discussion_r24763748
  
--- Diff: 
flink-staging/flink-gelly/src/test/java/org/apache/flink/graph/test/GatherSumApplyITCase.java
 ---
@@ -0,0 +1,212 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.graph.test;
+
+import org.apache.flink.api.common.functions.MapFunction;
+import org.apache.flink.api.java.ExecutionEnvironment;
+import org.apache.flink.api.java.tuple.Tuple2;
+import org.apache.flink.api.java.tuple.Tuple3;
+import org.apache.flink.graph.Edge;
+import org.apache.flink.graph.Graph;
+import org.apache.flink.graph.Vertex;
+import org.apache.flink.graph.gsa.ApplyFunction;
+import org.apache.flink.graph.gsa.GatherFunction;
+import org.apache.flink.graph.gsa.SumFunction;
+import org.apache.flink.test.util.MultipleProgramsTestBase;
+import org.apache.flink.util.Collector;
+import org.junit.After;
+import org.junit.Before;
+import org.junit.Rule;
+import org.junit.Test;
+import org.junit.rules.TemporaryFolder;
+import org.junit.runner.RunWith;
+import org.junit.runners.Parameterized;
+
+import java.util.ArrayList;
+import java.util.HashSet;
+import java.util.List;
+
+@RunWith(Parameterized.class)
+public class GatherSumApplyITCase extends MultipleProgramsTestBase {
+
+   public GatherSumApplyITCase(MultipleProgramsTestBase.ExecutionMode 
mode){
+   super(mode);
+   }
+
+   private String resultPath;
+   private String expectedResult;
+
+   @Rule
+   public TemporaryFolder tempFolder = new TemporaryFolder();
+
+   @Before
+   public void before() throws Exception{
+   resultPath = tempFolder.newFile().toURI().toString();
+   }
+
+   @After
+   public void after() throws Exception{
+   compareResultsByLinesInMemory(expectedResult, resultPath);
+   }
+
+   @Test
+   public void testGreedyGraphColoring() throws Exception {
+   /*
+* This test runs the Greedy Graph Coloring algorithm using a 
GSA iteration
+* The expected result is that the lowest vertex value (1) gets 
propagated to all vertices
+*/
+   final ExecutionEnvironment env = 
ExecutionEnvironment.getExecutionEnvironment();
+
+   Graph graph = 
Graph.fromDataSet(TestGraphUtils.getLongLongVertexData(env),
+   TestGraphUtils.getLongLongEdgeData(env), env);
+
+   // Gather the target vertices into a one-element set
+   GatherFunction>> 
gather =
+   new GatherFunction>>() {
+
+   @Override
+   public Tuple2>> gather(Tuple3,
+   Edge, 
Vertex> triplet) {
+
+   HashSet> 
result = new HashSet>();
+   result.add(triplet.f2);
+
+   return new Tuple2>>(triplet.f0.getId(), result);
+   }
+   };
+
+   // Merge the sets between neighbors
+   SumFunction>> sum =
+   new SumFunction>>() {
+
+   @Override
+   public Tuple2>> sum(
+   Tuple2>> arg0,
+   Tuple2>> arg1) {
+
+   HashSet> 
result = new HashSet>();
+   result.addAll(arg0.f1);
+   result.addAll(arg1.f1);
+
+   return new Tuple2>>(arg0.f0, result);
+   }
+

[jira] [Commented] (FLINK-1514) [Gelly] Add a Gather-Sum-Apply iteration method

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1514?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14323029#comment-14323029
 ] 

ASF GitHub Bot commented on FLINK-1514:
---

GitHub user balidani opened a pull request:

https://github.com/apache/flink/pull/408

[FLINK-1514][Gelly] Add a Gather-Sum-Apply iteration method



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/balidani/flink gelly-gsa

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/408.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #408


commit ae96bfefdd09b2d26fae644f071ea882f69a2375
Author: Dániel Bali 
Date:   2015-02-16T14:06:05Z

[FLINK-1514][Gelly] Add a Gather-Sum-Apply iteration method




> [Gelly] Add a Gather-Sum-Apply iteration method
> ---
>
> Key: FLINK-1514
> URL: https://issues.apache.org/jira/browse/FLINK-1514
> Project: Flink
>  Issue Type: New Feature
>  Components: Gelly
>Affects Versions: 0.9
>Reporter: Vasia Kalavri
>Assignee: Daniel Bali
>
> This will be a method that implements the GAS computation model, but without 
> the "scatter" step. The phases can be mapped into the following steps inside 
> a delta iteration:
> gather: a map on each < srcVertex, edge, trgVertex > that produces a partial 
> value
> sum: a reduce that combines the partial values
> apply: join with vertex set to update the vertex values using the results of 
> sum and the previous state.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: [FLINK-1514][Gelly] Add a Gather-Sum-Apply ite...

2015-02-16 Thread balidani
GitHub user balidani opened a pull request:

https://github.com/apache/flink/pull/408

[FLINK-1514][Gelly] Add a Gather-Sum-Apply iteration method



You can merge this pull request into a Git repository by running:

$ git pull https://github.com/balidani/flink gelly-gsa

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/408.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #408


commit ae96bfefdd09b2d26fae644f071ea882f69a2375
Author: Dániel Bali 
Date:   2015-02-16T14:06:05Z

[FLINK-1514][Gelly] Add a Gather-Sum-Apply iteration method




---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] flink pull request: [builds] Print Java process stack traces of st...

2015-02-16 Thread uce
Github user uce commented on the pull request:

https://github.com/apache/flink/pull/389#issuecomment-74535643
  
Subsumed by #407.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] flink pull request: [builds] Print Java process stack traces of st...

2015-02-16 Thread uce
Github user uce closed the pull request at:

https://github.com/apache/flink/pull/389


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] flink pull request: [builds] Allow to upload build artifacts to S3

2015-02-16 Thread uce
GitHub user uce opened a pull request:

https://github.com/apache/flink/pull/407

[builds] Allow to upload build artifacts to S3

This PR includes #389.

Currently, logging of Travis builds is disabled as the log files are 
discarded after a build and stdout logging is limited in size. With this PR, we 
can log to a file and have them uploaded automatically.

1. **Console logging**: For console logging, we can use 
`MavenForkNumberPrefixLayout` to prefix the log output with the fork number, 
which makes it easier to group log output together.

  We might consider moving that layout to a different package, though.

```
7 > 17:12:23,220 INFO  
org.apache.flink.runtime.executiongraph.ExecutionGraph- Deploying test 
vertex (5/7) (attempt #0) to host3
7 > 17:12:23,220 DEBUG 
org.apache.flink.runtime.jobmanager.scheduler.Scheduler   - Scheduling task 
{task=test vertex (6/7) - execution #0, sharingUnit=null, 
locationConstraint=null}
7 > 17:12:23,220 DEBUG 
org.apache.flink.runtime.jobmanager.scheduler.Scheduler   - Unconstrained 
assignment: test vertex (6/7) --> dbc442af2610e30d3309d96e14fa9d8a @ 
akka://TestingActorSystem/user/$$a - 5 slots - 1067599825
7 > 17:12:23,221 INFO  
org.apache.flink.runtime.executiongraph.ExecutionGraph- Deploying test 
vertex (6/7) (attempt #0) to host1
7 > 17:12:23,221 DEBUG 
org.apache.flink.runtime.jobmanager.scheduler.Scheduler   - Scheduling task 
{task=test vertex (7/7) - execution #0, sharingUnit=null, 
locationConstraint=null}
7 > 17:12:23,222 DEBUG 
org.apache.flink.runtime.jobmanager.scheduler.Scheduler   - Unconstrained 
assignment: test vertex (7/7) --> 8ce93c214951df5a8ad32c912ba0e456 @ 
akka://TestingActorSystem/user/$$a - 5 slots - 749927456
7 > 17:12:23,222 INFO  
org.apache.flink.runtime.executiongraph.ExecutionGraph- Deploying test 
vertex (7/7) (attempt #0) to host2
7 > 17:12:23,223 DEBUG 
org.apache.flink.runtime.executiongraph.ExecutionGraph- Attaching 1 
topologically sorted vertices to existing job graph with 0 vertices and 0 
intermediate results.
7 > 17:12:23,223 DEBUG 
org.apache.flink.runtime.executiongraph.ExecutionGraph- Connecting 
ExecutionJobVertex 126e16c348d4e5aa129104702700bf9e (test vertex) to 0 
predecessors.
10 > 17:12:23,225 INFO  
org.apache.flink.runtime.executiongraph.ExecutionGraph- Deploying v1 
(2/2) (attempt #0) to host3
6 > 17:12:23,236 DEBUG akka.event.EventStream   
 - shutting down: StandardOutLogger started
1 > 17:12:23,239 DEBUG 
org.apache.flink.runtime.executiongraph.ExecutionGraph- Sending out 
cancel request, to remove task execution from TaskManager.
1 > 17:12:23,244 DEBUG 
org.apache.flink.runtime.executiongraph.ExecutionGraph- Concurrent 
canceling/failing of TestVertex (1/1) - execution #0 while deployment was in 
progress.
10 > 17:12:23,266 DEBUG 
org.apache.flink.runtime.executiongraph.ExecutionGraph- Attaching 1 
topologically sorted vertices to existing job graph with 0 vertices and 0 
intermediate results.
Tests run: 11, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.61 sec - 
in org.apache.flink.runtime.executiongraph.ExecutionVertexCancelTest
Running org.apache.flink.runtime.io.network.api.reader.UnionBufferReaderTest
10 > 17:12:23,269 DEBUG 
org.apache.flink.runtime.executiongraph.ExecutionGraph- Connecting 
ExecutionJobVertex e8869426e7765e498ad219e3af731729 (test vertex) to 0 
predecessors.
10 > 17:12:23,269 DEBUG 
org.apache.flink.runtime.jobmanager.scheduler.Scheduler   - Scheduling task 
{task=test vertex (1/2) - execution #0, sharingUnit=null, 
locationConstraint=null}
10 > 17:12:23,270 DEBUG 
org.apache.flink.runtime.jobmanager.scheduler.Scheduler   - Local 
assignment: test vertex (1/2) --> ee9357ffc16d51c4f7522bc479a757ac @ 
akka://TestingActorSystem/user/$$a - 1 slots - 2073640037
10 > 17:12:23,270 INFO  
org.apache.flink.runtime.executiongraph.ExecutionGraph- Deploying test 
vertex (1/2) (attempt #0) to host1
10 > 17:12:23,270 DEBUG 
org.apache.flink.runtime.jobmanager.scheduler.Scheduler   - Scheduling task 
{task=test vertex (2/2) - execution #0, sharingUnit=null, 
locationConstraint=null}
10 > 17:12:23,271 DEBUG 
org.apache.flink.runtime.jobmanager.scheduler.Scheduler   - Local 
assignment: test vertex (2/2) --> 6620d9f45b7567e2b147f5761d8060e8 @ 
akka://TestingActorSystem/user/$$a - 1 slots - 1864116663
10 > 17:12:23,271 INFO  
org.apache.flink.runtime.executiongraph.ExecutionGraph- Deploying test 
vertex (2/2) (attempt #0) to host3
1 > 17:12:23,277 INFO  
org.apache.flink.runtime.executiongraph.ExecutionGraph- Deploying 
TestVertex (1/1) (attempt #0) to localhost
10 > 17:12:23,277 DEBUG 
org.apache.flink.runtime.executiongraph.ExecutionGraph- Attaching 1 
topologically sorted

[jira] [Created] (FLINK-1558) Spurious failure in PipelinedPartitionQueueTest

2015-02-16 Thread Stephan Ewen (JIRA)
Stephan Ewen created FLINK-1558:
---

 Summary: Spurious failure in PipelinedPartitionQueueTest
 Key: FLINK-1558
 URL: https://issues.apache.org/jira/browse/FLINK-1558
 Project: Flink
  Issue Type: Improvement
  Components: Distributed Runtime
Affects Versions: 0.9
Reporter: Stephan Ewen
 Fix For: 0.9


The failure is reported as
{code}
java.lang.AssertionError: Unexpected failure during test: null. Producer error: 
null, consumer error: null
at org.junit.Assert.fail(Assert.java:88)
at 
org.apache.flink.runtime.io.network.partition.queue.PipelinedPartitionQueueTest.doTestConcurrentProduceConsume(PipelinedPartitionQueueTest.java:214)
at 
org.apache.flink.runtime.io.network.partition.queue.PipelinedPartitionQueueTest.doTestConcurrentProduceConsume(PipelinedPartitionQueueTest.java:171)
at 
org.apache.flink.runtime.io.network.partition.queue.PipelinedPartitionQueueTest.testConcurrentProduceConsume(PipelinedPartitionQueueTest.java:142)
{code}

The test shows the following stack trace
{code}
java.util.concurrent.TimeoutException
at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:258)
at java.util.concurrent.FutureTask.get(FutureTask.java:119)
at 
org.apache.flink.runtime.io.network.partition.queue.PipelinedPartitionQueueTest.doTestConcurrentProduceConsume(PipelinedPartitionQueueTest.java:198)
at 
org.apache.flink.runtime.io.network.partition.queue.PipelinedPartitionQueueTest.doTestConcurrentProduceConsume(PipelinedPartitionQueueTest.java:171)
at 
org.apache.flink.runtime.io.network.partition.queue.PipelinedPartitionQueueTest.testConcurrentProduceConsume(PipelinedPartitionQueueTest.java:142)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:622)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:264)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:153)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:124)
at 
org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:200)
at 
org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:153)
at 
org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:103)
{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1528) Add local clustering coefficient library method and example

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1528?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14322906#comment-14322906
 ] 

ASF GitHub Bot commented on FLINK-1528:
---

Github user vasia commented on the pull request:

https://github.com/apache/flink/pull/400#issuecomment-74527767
  
Right, so let me copy:

The problem is here:
```
DataSet>> neighborhoods = input.reduceOnEdges(new 
NeighborhoodEdgesFunction(), EdgeDirection.OUT);
```

and we try to get the return type `Tuple2>` like this:

```
public TypeInformation getProducedType() {
return TypeExtractor.createTypeInfo(EdgesFunction.class, 
function.getClass(), 2, null, null);
}
```

Anyone have an idea? Is it because of the nested type parameter in 
`Tuple2>` ?


> Add local clustering coefficient library method and example
> ---
>
> Key: FLINK-1528
> URL: https://issues.apache.org/jira/browse/FLINK-1528
> Project: Flink
>  Issue Type: Task
>  Components: Gelly
>Reporter: Vasia Kalavri
>Assignee: Daniel Bali
>
> Add a gelly library method and example to compute the local clustering 
> coefficient.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (FLINK-1557) Move WebFrontend out of JobManager Actor

2015-02-16 Thread Stephan Ewen (JIRA)
Stephan Ewen created FLINK-1557:
---

 Summary: Move WebFrontend out of JobManager Actor
 Key: FLINK-1557
 URL: https://issues.apache.org/jira/browse/FLINK-1557
 Project: Flink
  Issue Type: Improvement
  Components: JobManager
Affects Versions: 0.9
Reporter: Stephan Ewen
Assignee: Stephan Ewen
 Fix For: 0.9


Currently, the Webfrontent is started inside the JobManager actor.

I propose to move it out of the actor, since it is a heavy component that is 
independent of the actor and whose failure cannot be controlled by the job 
manager.

The change is fairly straightforward, since the web frontend only communicates 
with the jobmanager through actor messages.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: [FLINK-1528] Added local clustering coefficien...

2015-02-16 Thread vasia
Github user vasia commented on the pull request:

https://github.com/apache/flink/pull/400#issuecomment-74527767
  
Right, so let me copy:

The problem is here:
```
DataSet>> neighborhoods = input.reduceOnEdges(new 
NeighborhoodEdgesFunction(), EdgeDirection.OUT);
```

and we try to get the return type `Tuple2>` like this:

```
public TypeInformation getProducedType() {
return TypeExtractor.createTypeInfo(EdgesFunction.class, 
function.getClass(), 2, null, null);
}
```

Anyone have an idea? Is it because of the nested type parameter in 
`Tuple2>` ?


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-1515) [Gelly] Enable access to aggregators and broadcast sets in vertex-centric iteration

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1515?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14322886#comment-14322886
 ] 

ASF GitHub Bot commented on FLINK-1515:
---

Github user vasia commented on the pull request:

https://github.com/apache/flink/pull/402#issuecomment-74525590
  
Hi @martinkiefer! Thanks a lot for this pull request!
The change looks good :)

I was only thinking that it would be nice to have an example that 
demonstrates how to use broadcast sets and / or aggregators now. It could 
simply be a slight modification of an existing example, e.g. SSSP with a 
counter of how many nodes have distance less than some value or similar. Could 
you make this addition?


> [Gelly] Enable access to aggregators and broadcast sets in vertex-centric 
> iteration
> ---
>
> Key: FLINK-1515
> URL: https://issues.apache.org/jira/browse/FLINK-1515
> Project: Flink
>  Issue Type: Improvement
>  Components: Gelly
>Reporter: Vasia Kalavri
>Assignee: Martin Kiefer
>
> Currently, aggregators and broadcast sets cannot be accessed through Gelly's  
> {{runVertexCentricIteration}} method. The functionality is already present in 
> the {{VertexCentricIteration}} and we just need to expose it.
> This could be done like this: We create a method 
> {{createVertexCentricIteration}}, which will return a 
> {{VertexCentricIteration}} object and we change {{runVertexCentricIteration}} 
> to accept this as a parameter (and return the graph after running this 
> iteration).
> The user can configure the {{VertexCentricIteration}} by directly calling the 
> public methods {{registerAggregator}}, {{setName}}, etc.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: [FLINK-1515]Splitted runVertexCentricIteration...

2015-02-16 Thread vasia
Github user vasia commented on the pull request:

https://github.com/apache/flink/pull/402#issuecomment-74525590
  
Hi @martinkiefer! Thanks a lot for this pull request!
The change looks good :)

I was only thinking that it would be nice to have an example that 
demonstrates how to use broadcast sets and / or aggregators now. It could 
simply be a slight modification of an existing example, e.g. SSSP with a 
counter of how many nodes have distance less than some value or similar. Could 
you make this addition?


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] flink pull request: [FLINK-1556] Corrects faulty JobClient behavio...

2015-02-16 Thread tillrohrmann
GitHub user tillrohrmann opened a pull request:

https://github.com/apache/flink/pull/406

[FLINK-1556] Corrects faulty JobClient behaviour in case of a submission 
failure

If an error occurred during job submission, a ```SubmissionFailure``` is 
sent to the ```JobClient```. As a reaction, the ```JobClient``` terminated 
itself and sent the failure to the ```Client```. However, this does not 
necessarily mean that the job has reached a terminal state, because the failing 
procedure is executed asynchronously.

The ```JobClient``` now waits until it receives a ```JobResult``` message 
indicating that the job has completed and all resources are properly returned.

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/tillrohrmann/flink minorFixes

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/406.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #406


commit 2f32e9c6b87b8e295f792c04306d78fbb858f80d
Author: Till Rohrmann 
Date:   2015-02-16T09:17:21Z

[FLINK-1556] [runtime] Corrects faulty JobClient behaviour in case of a 
submission failure




---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-1556) JobClient does not wait until a job failed completely if submission exception

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1556?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14322884#comment-14322884
 ] 

ASF GitHub Bot commented on FLINK-1556:
---

GitHub user tillrohrmann opened a pull request:

https://github.com/apache/flink/pull/406

[FLINK-1556] Corrects faulty JobClient behaviour in case of a submission 
failure

If an error occurred during job submission, a ```SubmissionFailure``` is 
sent to the ```JobClient```. As a reaction, the ```JobClient``` terminated 
itself and sent the failure to the ```Client```. However, this does not 
necessarily mean that the job has reached a terminal state, because the failing 
procedure is executed asynchronously.

The ```JobClient``` now waits until it receives a ```JobResult``` message 
indicating that the job has completed and all resources are properly returned.

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/tillrohrmann/flink minorFixes

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/406.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #406


commit 2f32e9c6b87b8e295f792c04306d78fbb858f80d
Author: Till Rohrmann 
Date:   2015-02-16T09:17:21Z

[FLINK-1556] [runtime] Corrects faulty JobClient behaviour in case of a 
submission failure




> JobClient does not wait until a job failed completely if submission exception
> -
>
> Key: FLINK-1556
> URL: https://issues.apache.org/jira/browse/FLINK-1556
> Project: Flink
>  Issue Type: Bug
>Reporter: Till Rohrmann
>
> If an exception occurs during job submission the {{JobClient}} received a 
> {{SubmissionFailure}}. Upon receiving this message, the {{JobClient}} 
> terminates itself and returns the error to the {{Client}}. This indicates to 
> the user that the job has been completely failed which is not necessarily 
> true. 
> If the user directly after such a failure submits another job, then it might 
> be the case that not all slots of the formerly failed job are returned. This 
> can lead to a {{NoRessourceAvailableException}}.
> We can solve this problem by waiting for the completion of the job failure in 
> the {{JobClient}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (FLINK-1556) JobClient does not wait until a job failed completely if submission exception

2015-02-16 Thread Till Rohrmann (JIRA)
Till Rohrmann created FLINK-1556:


 Summary: JobClient does not wait until a job failed completely if 
submission exception
 Key: FLINK-1556
 URL: https://issues.apache.org/jira/browse/FLINK-1556
 Project: Flink
  Issue Type: Bug
Reporter: Till Rohrmann


If an exception occurs during job submission the {{JobClient}} received a 
{{SubmissionFailure}}. Upon receiving this message, the {{JobClient}} 
terminates itself and returns the error to the {{Client}}. This indicates to 
the user that the job has been completely failed which is not necessarily true. 

If the user directly after such a failure submits another job, then it might be 
the case that not all slots of the formerly failed job are returned. This can 
lead to a {{NoRessourceAvailableException}}.

We can solve this problem by waiting for the completion of the job failure in 
the {{JobClient}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1554) LocalFlinkMiniCluster cannot start multiple local TaskManagers

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1554?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14322860#comment-14322860
 ] 

ASF GitHub Bot commented on FLINK-1554:
---

Github user rmetzger commented on the pull request:

https://github.com/apache/flink/pull/403#issuecomment-74523049
  
Perfect. I've justed the code and its working now.


> LocalFlinkMiniCluster cannot start multiple local TaskManagers
> --
>
> Key: FLINK-1554
> URL: https://issues.apache.org/jira/browse/FLINK-1554
> Project: Flink
>  Issue Type: Bug
>Reporter: Till Rohrmann
>
> The {{LocalFlinkMiniCluster}} cannot currently start multiple 
> {{TaskManager}}s in the same {{ActorSystem}}, because it assigns them the 
> same actor name.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: [FLINK-1554] Allows the LocalFlinkMiniCluster ...

2015-02-16 Thread rmetzger
Github user rmetzger commented on the pull request:

https://github.com/apache/flink/pull/403#issuecomment-74523049
  
Perfect. I've justed the code and its working now.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Created] (FLINK-1555) Add utility to log the serializers of composite types

2015-02-16 Thread Robert Metzger (JIRA)
Robert Metzger created FLINK-1555:
-

 Summary: Add utility to log the serializers of composite types
 Key: FLINK-1555
 URL: https://issues.apache.org/jira/browse/FLINK-1555
 Project: Flink
  Issue Type: Improvement
Reporter: Robert Metzger
Priority: Minor


Users affected by poor performance might want to understand how Flink is 
serializing their data.

Therefore, it would be cool to have a tool utility which logs the serializers 
like this:
{{SerializerUtils.getSerializers(TypeInformation t);}}
to get 
{code}
PojoSerializer
TupleSerializer
  IntSer
  DateSer
  GenericTypeSer(java.sql.Date)
PojoSerializer
  GenericTypeSer(HashMap)
{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-947) Add support for "Named Datasets"

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-947?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14322816#comment-14322816
 ] 

ASF GitHub Bot commented on FLINK-947:
--

GitHub user aljoscha opened a pull request:

https://github.com/apache/flink/pull/405

[FLINK-947] Add a declarative expression API

This one is quite big. So you should check out the documentation, skaldic, 
examples and test cases to see how the API works.

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/aljoscha/flink linq

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/405.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #405


commit 147525ced43db6690a64fbae1395dbd258b8901d
Author: Aljoscha Krettek 
Date:   2014-10-03T16:25:15Z

Change translateToDataflow to return Operator

Before, translateToDataflow of SingleInputOperator could only return
a single input operator of the lower layer, same for TwoInputOperator.

This change allows translateToDataflow to return more kinds of
operators.

commit 58b5b9ec6e65855bfd71287deb6352dfc4498451
Author: Aljoscha Krettek 
Date:   2014-10-23T16:09:38Z

Add methods to CompositeType for querying field types and names

commit ac29ee3ad36a72d7c41549f38da1a00e66d85041
Author: Aljoscha Krettek 
Date:   2014-10-01T11:12:18Z

[FLINK-947] Add a declarative expression API




> Add support for "Named Datasets"
> 
>
> Key: FLINK-947
> URL: https://issues.apache.org/jira/browse/FLINK-947
> Project: Flink
>  Issue Type: New Feature
>  Components: Java API
>Reporter: Aljoscha Krettek
>Assignee: Aljoscha Krettek
>Priority: Minor
>
> This would create an API that is a mix between SQL like declarativity and the 
> power of user defined functions. Example user code could look like this:
> {code:Java}
> NamedDataSet one = ...
> NamedDataSet two = ...
> NamedDataSet result = one.join(two).where("key").equalTo("otherKey")
>   .project("a", "b", "c")
>   .map( (UserTypeIn in) -> return new UserTypeOut(...) )
>   .print();
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: [FLINK-947] Add a declarative expression API

2015-02-16 Thread aljoscha
GitHub user aljoscha opened a pull request:

https://github.com/apache/flink/pull/405

[FLINK-947] Add a declarative expression API

This one is quite big. So you should check out the documentation, skaldic, 
examples and test cases to see how the API works.

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/aljoscha/flink linq

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/405.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #405


commit 147525ced43db6690a64fbae1395dbd258b8901d
Author: Aljoscha Krettek 
Date:   2014-10-03T16:25:15Z

Change translateToDataflow to return Operator

Before, translateToDataflow of SingleInputOperator could only return
a single input operator of the lower layer, same for TwoInputOperator.

This change allows translateToDataflow to return more kinds of
operators.

commit 58b5b9ec6e65855bfd71287deb6352dfc4498451
Author: Aljoscha Krettek 
Date:   2014-10-23T16:09:38Z

Add methods to CompositeType for querying field types and names

commit ac29ee3ad36a72d7c41549f38da1a00e66d85041
Author: Aljoscha Krettek 
Date:   2014-10-01T11:12:18Z

[FLINK-947] Add a declarative expression API




---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Closed] (FLINK-1539) Runtime context not initialized when running streaming PojoExample

2015-02-16 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/FLINK-1539?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Márton Balassi closed FLINK-1539.
-
Resolution: Fixed

Fixed via 4470207

> Runtime context not initialized when running streaming PojoExample
> --
>
> Key: FLINK-1539
> URL: https://issues.apache.org/jira/browse/FLINK-1539
> Project: Flink
>  Issue Type: Bug
>  Components: Streaming
>Affects Versions: 0.9
>Reporter: Márton Balassi
>Assignee: Gyula Fora
>
> When running streaming PojoExample received the following exception:
> Exception in thread "main" java.lang.IllegalStateException: The runtime 
> context has not been initialized.
>   at 
> org.apache.flink.api.common.functions.AbstractRichFunction.getRuntimeContext(AbstractRichFunction.java:49)
>   at 
> org.apache.flink.streaming.api.function.aggregation.SumAggregator$PojoSumAggregator.(SumAggregator.java:149)
>   at 
> org.apache.flink.streaming.api.function.aggregation.SumAggregator.getSumFunction(SumAggregator.java:52)
>   at 
> org.apache.flink.streaming.api.datastream.DataStream.sum(DataStream.java:632)
>   at 
> org.apache.flink.streaming.examples.wordcount.PojoExample.main(PojoExample.java:65)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:483)
>   at com.intellij.rt.execution.application.AppMain.main(AppMain.java:134)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Closed] (FLINK-1176) Add operator for global reduce/aggregations

2015-02-16 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/FLINK-1176?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Márton Balassi closed FLINK-1176.
-
Resolution: Implemented

Implemented via 1146f64 and c560d76.

> Add operator for global reduce/aggregations
> ---
>
> Key: FLINK-1176
> URL: https://issues.apache.org/jira/browse/FLINK-1176
> Project: Flink
>  Issue Type: Improvement
>  Components: Streaming
>Reporter: Gyula Fora
>Assignee: Gyula Fora
>Priority: Minor
>
> Currently streaming reduce operators (like .reduce or .window(..).reduce, 
> sum, etc.) only allow local reduces/aggregations. While one can achieve 
> global aggregation by setting the parallelism of these operators to 1 it can 
> cause a serious bottleneck in the program. 
> For window/batch reduces and aggregations a global reducer can be added, with 
> parallelism 1,  to further reduce partial results. This global reduce option 
> could be introduced as an optional boolean parameter for the intended 
> operators.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Closed] (FLINK-1231) Add test for streaming remote execution

2015-02-16 Thread JIRA

 [ 
https://issues.apache.org/jira/browse/FLINK-1231?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Márton Balassi closed FLINK-1231.
-
Resolution: Fixed

Implemented via 26ae979

> Add test for streaming remote execution
> ---
>
> Key: FLINK-1231
> URL: https://issues.apache.org/jira/browse/FLINK-1231
> Project: Flink
>  Issue Type: Test
>  Components: Streaming
>Affects Versions: 0.7.1-incubating
>Reporter: Márton Balassi
>
> Add a test case that mocks the remote job execution. 
> PackagedProgramEndToEndITCase serves as an example.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1231) Add test for streaming remote execution

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1231?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14322808#comment-14322808
 ] 

ASF GitHub Bot commented on FLINK-1231:
---

Github user asfgit closed the pull request at:

https://github.com/apache/flink/pull/303


> Add test for streaming remote execution
> ---
>
> Key: FLINK-1231
> URL: https://issues.apache.org/jira/browse/FLINK-1231
> Project: Flink
>  Issue Type: Test
>  Components: Streaming
>Affects Versions: 0.7.1-incubating
>Reporter: Márton Balassi
>
> Add a test case that mocks the remote job execution. 
> PackagedProgramEndToEndITCase serves as an example.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: StreamWindow abstraction + modular window comp...

2015-02-16 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/flink/pull/395


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] flink pull request: [FLINK-1231] Add test for streaming remote exe...

2015-02-16 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/flink/pull/303


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-1454) CliFrontend blocks for 100 seconds when submitting to a non-existent JobManager

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1454?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14322803#comment-14322803
 ] 

ASF GitHub Bot commented on FLINK-1454:
---

GitHub user StephanEwen opened a pull request:

https://github.com/apache/flink/pull/404

[FLINK-1454] [job client] Improve error handling for failed connections 
between JobClient and JobManager

  - Decreases time until connection failure is discovered.
  - Improves error messages
  - adds / extends tests
  - Removes implicits in runtime, makes actor system and timeout always an 
explicit parameter
  - Removes old redundant config constants

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/StephanEwen/incubator-flink client_connect

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/404.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #404


commit 9365ba8cb3859d7a04a301d3417e418914ce813c
Author: Stephan Ewen 
Date:   2015-02-16T12:37:22Z

[FLINK-1454] [job client] Improve error handling for failed connections 
between JobClient and JobManager




> CliFrontend blocks for 100 seconds when submitting to a non-existent 
> JobManager
> ---
>
> Key: FLINK-1454
> URL: https://issues.apache.org/jira/browse/FLINK-1454
> Project: Flink
>  Issue Type: Bug
>  Components: JobManager
>Affects Versions: 0.9
>Reporter: Robert Metzger
>
> When a user tries to submit a job to a job manager which doesn't exist at 
> all, the CliFrontend blocks for 100 seconds.
> Ideally, Akka would fail because it can not connect to the given 
> hostname:port.
>  
> {code}
> ./bin/flink run ./examples/flink-java-examples-0.9-SNAPSHOT-WordCount.jar -c 
> foo.Baz
> org.apache.flink.client.program.ProgramInvocationException: The main method 
> caused an error.
>   at 
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:449)
>   at 
> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:350)
>   at org.apache.flink.client.program.Client.run(Client.java:242)
>   at 
> org.apache.flink.client.CliFrontend.executeProgram(CliFrontend.java:389)
>   at org.apache.flink.client.CliFrontend.run(CliFrontend.java:362)
>   at 
> org.apache.flink.client.CliFrontend.parseParameters(CliFrontend.java:1078)
>   at org.apache.flink.client.CliFrontend.main(CliFrontend.java:1102)
> Caused by: java.util.concurrent.TimeoutException: Futures timed out after 
> [100 seconds]
>   at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>   at 
> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
>   at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
>   at 
> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
>   at scala.concurrent.Await$.result(package.scala:107)
>   at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
>   at 
> org.apache.flink.runtime.client.JobClient$.uploadJarFiles(JobClient.scala:169)
>   at 
> org.apache.flink.runtime.client.JobClient.uploadJarFiles(JobClient.scala)
>   at org.apache.flink.client.program.Client.run(Client.java:314)
>   at org.apache.flink.client.program.Client.run(Client.java:296)
>   at org.apache.flink.client.program.Client.run(Client.java:290)
>   at 
> org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:55)
>   at 
> org.apache.flink.examples.java.wordcount.WordCount.main(WordCount.java:82)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:606)
>   at 
> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:434)
>   ... 6 more
> The exception above occurred while trying to run your command.
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: [FLINK-1454] [job client] Improve error handli...

2015-02-16 Thread StephanEwen
GitHub user StephanEwen opened a pull request:

https://github.com/apache/flink/pull/404

[FLINK-1454] [job client] Improve error handling for failed connections 
between JobClient and JobManager

  - Decreases time until connection failure is discovered.
  - Improves error messages
  - adds / extends tests
  - Removes implicits in runtime, makes actor system and timeout always an 
explicit parameter
  - Removes old redundant config constants

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/StephanEwen/incubator-flink client_connect

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/404.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #404


commit 9365ba8cb3859d7a04a301d3417e418914ce813c
Author: Stephan Ewen 
Date:   2015-02-16T12:37:22Z

[FLINK-1454] [job client] Improve error handling for failed connections 
between JobClient and JobManager




---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-1554) LocalFlinkMiniCluster cannot start multiple local TaskManagers

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1554?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14322797#comment-14322797
 ] 

ASF GitHub Bot commented on FLINK-1554:
---

GitHub user tillrohrmann opened a pull request:

https://github.com/apache/flink/pull/403

[FLINK-1554] Allows the LocalFlinkMiniCluster to start multiple TaskManager 
in the same ActorSystem

Allows the ```LocalFlinkMiniCluster``` to start multiple ```TaskManager``` 
in the same ```ActorSystem``` by assigning them unique actor names.

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/tillrohrmann/flink fixLocalFlinkMiniCluster

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/403.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #403


commit cecb45d45f05c0aa86f4ae780e8b49830f3ea6ac
Author: Till Rohrmann 
Date:   2015-02-16T14:04:42Z

[FLINK-1554] [runtime] Allows the LocalFlinkMiniCluster to start multiple 
TaskManager in the same ActorSystem.




> LocalFlinkMiniCluster cannot start multiple local TaskManagers
> --
>
> Key: FLINK-1554
> URL: https://issues.apache.org/jira/browse/FLINK-1554
> Project: Flink
>  Issue Type: Bug
>Reporter: Till Rohrmann
>
> The {{LocalFlinkMiniCluster}} cannot currently start multiple 
> {{TaskManager}}s in the same {{ActorSystem}}, because it assigns them the 
> same actor name.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: [FLINK-1554] Allows the LocalFlinkMiniCluster ...

2015-02-16 Thread tillrohrmann
GitHub user tillrohrmann opened a pull request:

https://github.com/apache/flink/pull/403

[FLINK-1554] Allows the LocalFlinkMiniCluster to start multiple TaskManager 
in the same ActorSystem

Allows the ```LocalFlinkMiniCluster``` to start multiple ```TaskManager``` 
in the same ```ActorSystem``` by assigning them unique actor names.

You can merge this pull request into a Git repository by running:

$ git pull https://github.com/tillrohrmann/flink fixLocalFlinkMiniCluster

Alternatively you can review and apply these changes as the patch at:

https://github.com/apache/flink/pull/403.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

This closes #403


commit cecb45d45f05c0aa86f4ae780e8b49830f3ea6ac
Author: Till Rohrmann 
Date:   2015-02-16T14:04:42Z

[FLINK-1554] [runtime] Allows the LocalFlinkMiniCluster to start multiple 
TaskManager in the same ActorSystem.




---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Created] (FLINK-1554) LocalFlinkMiniCluster cannot start multiple local TaskManagers

2015-02-16 Thread Till Rohrmann (JIRA)
Till Rohrmann created FLINK-1554:


 Summary: LocalFlinkMiniCluster cannot start multiple local 
TaskManagers
 Key: FLINK-1554
 URL: https://issues.apache.org/jira/browse/FLINK-1554
 Project: Flink
  Issue Type: Bug
Reporter: Till Rohrmann


The {{LocalFlinkMiniCluster}} cannot currently start multiple {{TaskManager}}s 
in the same {{ActorSystem}}, because it assigns them the same actor name.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1542) Test case at BlobUtilsTest should not assume user could not create new item in root directory

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1542?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14322791#comment-14322791
 ] 

ASF GitHub Bot commented on FLINK-1542:
---

Github user uce commented on the pull request:

https://github.com/apache/flink/pull/396#issuecomment-74511789
  
Thanks, Henry. The change looks good to me. :+1:


> Test case at BlobUtilsTest should not assume user could not create new item 
> in root directory
> -
>
> Key: FLINK-1542
> URL: https://issues.apache.org/jira/browse/FLINK-1542
> Project: Flink
>  Issue Type: Bug
>Reporter: Henry Saputra
>Assignee: Henry Saputra
>
> The test at BlobUtilsTest #testExceptionOnCreateStorageDirectoryFailure 
> assume that current user cannot create new directory at root.
> In my MacOSX, my user has admin group so the test will always fail.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: [FLINK-1542] Test case at BlobUtilsTest should...

2015-02-16 Thread uce
Github user uce commented on the pull request:

https://github.com/apache/flink/pull/396#issuecomment-74511789
  
Thanks, Henry. The change looks good to me. :+1:


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-1549) Add proper exception handling for YarnJobManager

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1549?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14322787#comment-14322787
 ] 

ASF GitHub Bot commented on FLINK-1549:
---

Github user mxm commented on the pull request:

https://github.com/apache/flink/pull/397#issuecomment-74511630
  
Looks good. Very much needed for our YARN deployment.


> Add proper exception handling for YarnJobManager
> 
>
> Key: FLINK-1549
> URL: https://issues.apache.org/jira/browse/FLINK-1549
> Project: Flink
>  Issue Type: Bug
>Reporter: Till Rohrmann
>
> The YarnJobManager actor thread calls methods which can throw an exception. 
> These exceptions should be caught and properly handled.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: [FLINK-1549] Adds proper exception handling to...

2015-02-16 Thread mxm
Github user mxm commented on the pull request:

https://github.com/apache/flink/pull/397#issuecomment-74511630
  
Looks good. Very much needed for our YARN deployment.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-1552) Allow secondary sorts in Vertex Centric Iteration

2015-02-16 Thread Vasia Kalavri (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1552?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14322766#comment-14322766
 ] 

Vasia Kalavri commented on FLINK-1552:
--

This would be very useful to have! Would you like to work on it [~martinkiefer]?

> Allow secondary sorts in Vertex Centric Iteration
> -
>
> Key: FLINK-1552
> URL: https://issues.apache.org/jira/browse/FLINK-1552
> Project: Flink
>  Issue Type: Wish
>  Components: Gelly
>Reporter: Martin Kiefer
>Priority: Minor
>
> The `VertexCentricIteration` class holds the logic to transform a 
> `VertexUpdateFunction` and a `MessagingFunction` into an iteration with two 
> CoGroup operators working on the set of messages and edges. Graph algorithms 
> can profit from implying an order on the edges or messages based on their 
> value and/or the vertex ID. This can be implemented easily making use of 
> secondary sorts. I would suggest extending the `VertexCentricIteration` to 
> allow to specify these kind of orderings optionally.
> For example, this comes handy when it is necessary to find the edges with the 
> minimum or maximum value or the algorithm requires to pick the edge with 
> lower vertex ID for edges with equal value. Similar use cases might be found 
> for orders on the messages. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1417) Automatically register nested types at Kryo

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1417?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14322743#comment-14322743
 ] 

ASF GitHub Bot commented on FLINK-1417:
---

Github user aljoscha commented on the pull request:

https://github.com/apache/flink/pull/393#issuecomment-74505787
  
Looks good except for the one remark I had.

Did you run any performance tests on this? What is the overhead of using an 
Avro POJO vs. a plain POJO vs. using the Kryo Serializer?


> Automatically register nested types at Kryo
> ---
>
> Key: FLINK-1417
> URL: https://issues.apache.org/jira/browse/FLINK-1417
> Project: Flink
>  Issue Type: Improvement
>  Components: Java API, Scala API
>Reporter: Stephan Ewen
>Assignee: Robert Metzger
> Fix For: 0.9
>
>
> Currently, the {{GenericTypeInfo}} registers the class of the type at Kryo. 
> In order to get the best performance, it should recursively walk the classes 
> and make sure that it registered all contained subtypes.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1528) Add local clustering coefficient library method and example

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1528?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14322741#comment-14322741
 ] 

ASF GitHub Bot commented on FLINK-1528:
---

Github user balidani commented on the pull request:

https://github.com/apache/flink/pull/400#issuecomment-74505722
  
Hi!

Sorry for the messed up commit history, hopefully my next PR will be better.

I have the following exception when running 
`LocalClusteringCoefficientExample`

Exception in thread "main" 
org.apache.flink.api.common.functions.InvalidTypesException: 
Type of TypeVariable 'K' in 'class 
flink.graphs.library.LocalClusteringCoefficient
$NeighborhoodEdgesFunction' could not be determined. 
This is most likely a type erasure problem. The type extraction 
currently supports types 
with generic variables only in cases where all variables in the return 
type can be deduced 
from the input type(s).

I have no idea how to fix this though. For anyone else reading this, here's 
@vasia's reply from the old `flink-graph` thread: 
https://github.com/project-flink/flink-graph/pull/48#issuecomment-70472784


> Add local clustering coefficient library method and example
> ---
>
> Key: FLINK-1528
> URL: https://issues.apache.org/jira/browse/FLINK-1528
> Project: Flink
>  Issue Type: Task
>  Components: Gelly
>Reporter: Vasia Kalavri
>Assignee: Daniel Bali
>
> Add a gelly library method and example to compute the local clustering 
> coefficient.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: [FLINK-1417] Automatically register types with...

2015-02-16 Thread aljoscha
Github user aljoscha commented on the pull request:

https://github.com/apache/flink/pull/393#issuecomment-74505787
  
Looks good except for the one remark I had.

Did you run any performance tests on this? What is the overhead of using an 
Avro POJO vs. a plain POJO vs. using the Kryo Serializer?


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] flink pull request: [FLINK-1528] Added local clustering coefficien...

2015-02-16 Thread balidani
Github user balidani commented on the pull request:

https://github.com/apache/flink/pull/400#issuecomment-74505722
  
Hi!

Sorry for the messed up commit history, hopefully my next PR will be better.

I have the following exception when running 
`LocalClusteringCoefficientExample`

Exception in thread "main" 
org.apache.flink.api.common.functions.InvalidTypesException: 
Type of TypeVariable 'K' in 'class 
flink.graphs.library.LocalClusteringCoefficient
$NeighborhoodEdgesFunction' could not be determined. 
This is most likely a type erasure problem. The type extraction 
currently supports types 
with generic variables only in cases where all variables in the return 
type can be deduced 
from the input type(s).

I have no idea how to fix this though. For anyone else reading this, here's 
@vasia's reply from the old `flink-graph` thread: 
https://github.com/project-flink/flink-graph/pull/48#issuecomment-70472784


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-1417) Automatically register nested types at Kryo

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1417?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14322736#comment-14322736
 ] 

ASF GitHub Bot commented on FLINK-1417:
---

Github user aljoscha commented on a diff in the pull request:

https://github.com/apache/flink/pull/393#discussion_r24749284
  
--- Diff: 
flink-core/src/main/java/org/apache/flink/api/common/ExecutionConfig.java ---
@@ -292,7 +360,76 @@ public void registerKryoType(Class type) {
/**
 * Returns the registered POJO types.
 */
-   public Set> getRegisteredPojoTypes() {
+   public List> getRegisteredPojoTypes() {
return registeredPojoTypes;
}
+
+
+   public boolean isDisableAutoTypeRegistration() {
+   return disableAutoTypeRegistration;
+   }
+
+   /**
+* Control whether Flink is automatically registering all types in the 
user programs with
+* Kryo.
+*
+* @param disableAutoTypeRegistration
+*/
+   public void setDisableAutoTypeRegistration(boolean 
disableAutoTypeRegistration) {
--- End diff --

I would prefer "disableAutoTypeRegistration" here. And then 
"isAutoTypeRegistrationDisabled", above.


> Automatically register nested types at Kryo
> ---
>
> Key: FLINK-1417
> URL: https://issues.apache.org/jira/browse/FLINK-1417
> Project: Flink
>  Issue Type: Improvement
>  Components: Java API, Scala API
>Reporter: Stephan Ewen
>Assignee: Robert Metzger
> Fix For: 0.9
>
>
> Currently, the {{GenericTypeInfo}} registers the class of the type at Kryo. 
> In order to get the best performance, it should recursively walk the classes 
> and make sure that it registered all contained subtypes.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: [FLINK-1417] Automatically register types with...

2015-02-16 Thread aljoscha
Github user aljoscha commented on a diff in the pull request:

https://github.com/apache/flink/pull/393#discussion_r24749284
  
--- Diff: 
flink-core/src/main/java/org/apache/flink/api/common/ExecutionConfig.java ---
@@ -292,7 +360,76 @@ public void registerKryoType(Class type) {
/**
 * Returns the registered POJO types.
 */
-   public Set> getRegisteredPojoTypes() {
+   public List> getRegisteredPojoTypes() {
return registeredPojoTypes;
}
+
+
+   public boolean isDisableAutoTypeRegistration() {
+   return disableAutoTypeRegistration;
+   }
+
+   /**
+* Control whether Flink is automatically registering all types in the 
user programs with
+* Kryo.
+*
+* @param disableAutoTypeRegistration
+*/
+   public void setDisableAutoTypeRegistration(boolean 
disableAutoTypeRegistration) {
--- End diff --

I would prefer "disableAutoTypeRegistration" here. And then 
"isAutoTypeRegistrationDisabled", above.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] flink pull request: Fix standalone plan visualizer for 0.8 branch

2015-02-16 Thread asfgit
Github user asfgit closed the pull request at:

https://github.com/apache/flink/pull/401


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Closed] (FLINK-1362) Spurious errors in execution state messaging between TaskManager and JobManager

2015-02-16 Thread Till Rohrmann (JIRA)

 [ 
https://issues.apache.org/jira/browse/FLINK-1362?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Till Rohrmann closed FLINK-1362.

Resolution: Fixed

Fixed via 4046819b380b8dfa57d52c6d314f389546a159a3

> Spurious errors in execution state messaging between TaskManager and 
> JobManager
> ---
>
> Key: FLINK-1362
> URL: https://issues.apache.org/jira/browse/FLINK-1362
> Project: Flink
>  Issue Type: Bug
>  Components: Distributed Runtime
>Affects Versions: 0.9
>Reporter: Stephan Ewen
>
> In various tasks, it is possible that a notification from the TaskManager to 
> the JobManager fails with the following exception:
> {code}
> Exception in thread "" java.lang.NullPointerException
>   at 
> org.apache.flink.runtime.taskmanager.TaskManager.notifyExecutionStateChange(TaskManager.scala:281)
>   at org.apache.flink.runtime.taskmanager.Task.markFailed(Task.java:199)
>   at 
> org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:307)
>   at java.lang.Thread.run(Thread.java:745)
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (FLINK-1542) Test case at BlobUtilsTest should not assume user could not create new item in root directory

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1542?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14322699#comment-14322699
 ] 

ASF GitHub Bot commented on FLINK-1542:
---

Github user hsaputra commented on the pull request:

https://github.com/apache/flink/pull/396#issuecomment-74500203
  
Ping for review?


> Test case at BlobUtilsTest should not assume user could not create new item 
> in root directory
> -
>
> Key: FLINK-1542
> URL: https://issues.apache.org/jira/browse/FLINK-1542
> Project: Flink
>  Issue Type: Bug
>Reporter: Henry Saputra
>Assignee: Henry Saputra
>
> The test at BlobUtilsTest #testExceptionOnCreateStorageDirectoryFailure 
> assume that current user cannot create new directory at root.
> In my MacOSX, my user has admin group so the test will always fail.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: [FLINK-1542] Test case at BlobUtilsTest should...

2015-02-16 Thread hsaputra
Github user hsaputra commented on the pull request:

https://github.com/apache/flink/pull/396#issuecomment-74500203
  
Ping for review?


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-1528) Add local clustering coefficient library method and example

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1528?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14322649#comment-14322649
 ] 

ASF GitHub Bot commented on FLINK-1528:
---

Github user vasia commented on the pull request:

https://github.com/apache/flink/pull/400#issuecomment-74493011
  
Hi @balidani! Thanks for the PR. The commit history is a bit messed up, but 
we'll fix that. It's just the last commit that's relevant, right?

If I recall correctly, you are having some type issues. Could you please 
give us some details on what is the error you are getting and what seems to be 
causing it? Thanks!


> Add local clustering coefficient library method and example
> ---
>
> Key: FLINK-1528
> URL: https://issues.apache.org/jira/browse/FLINK-1528
> Project: Flink
>  Issue Type: Task
>  Components: Gelly
>Reporter: Vasia Kalavri
>Assignee: Daniel Bali
>
> Add a gelly library method and example to compute the local clustering 
> coefficient.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: [FLINK-1528] Added local clustering coefficien...

2015-02-16 Thread vasia
Github user vasia commented on the pull request:

https://github.com/apache/flink/pull/400#issuecomment-74493011
  
Hi @balidani! Thanks for the PR. The commit history is a bit messed up, but 
we'll fix that. It's just the last commit that's relevant, right?

If I recall correctly, you are having some type issues. Could you please 
give us some details on what is the error you are getting and what seems to be 
causing it? Thanks!


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] flink pull request: [FLINK-1231] Add test for streaming remote exe...

2015-02-16 Thread mbalassi
Github user mbalassi commented on the pull request:

https://github.com/apache/flink/pull/303#issuecomment-74480487
  
Sorry for the late reply. Thanks for the great PR, I wanted to merge it for 
quite some time. It turns out that this test is redundant given the 6bec228 
I've pushed last week. As a result of that I would like to close this PR 
without merging your commit.

Sorry, it is on me: I've just recently realized that the difference between 
submitting through the `Client` and the `PackagedProgram` have very little 
differences.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Commented] (FLINK-1231) Add test for streaming remote execution

2015-02-16 Thread ASF GitHub Bot (JIRA)

[ 
https://issues.apache.org/jira/browse/FLINK-1231?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14322565#comment-14322565
 ] 

ASF GitHub Bot commented on FLINK-1231:
---

Github user mbalassi commented on the pull request:

https://github.com/apache/flink/pull/303#issuecomment-74480487
  
Sorry for the late reply. Thanks for the great PR, I wanted to merge it for 
quite some time. It turns out that this test is redundant given the 6bec228 
I've pushed last week. As a result of that I would like to close this PR 
without merging your commit.

Sorry, it is on me: I've just recently realized that the difference between 
submitting through the `Client` and the `PackagedProgram` have very little 
differences.


> Add test for streaming remote execution
> ---
>
> Key: FLINK-1231
> URL: https://issues.apache.org/jira/browse/FLINK-1231
> Project: Flink
>  Issue Type: Test
>  Components: Streaming
>Affects Versions: 0.7.1-incubating
>Reporter: Márton Balassi
>
> Add a test case that mocks the remote job execution. 
> PackagedProgramEndToEndITCase serves as an example.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Closed] (FLINK-1551) Fix History view of JobManager web interface

2015-02-16 Thread Till Rohrmann (JIRA)

 [ 
https://issues.apache.org/jira/browse/FLINK-1551?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Till Rohrmann closed FLINK-1551.

Resolution: Fixed

Fixed via 055f6dc3eb8705dbbfb04e44965a85cbfc0b10db

> Fix History view of JobManager web interface
> 
>
> Key: FLINK-1551
> URL: https://issues.apache.org/jira/browse/FLINK-1551
> Project: Flink
>  Issue Type: Bug
>  Components: JobManager
>Affects Versions: 0.9
>Reporter: Robert Metzger
>Priority: Critical
>
> The JobManager web history view is not working.
> {{scala.concurrent.impl.Promise$DefaultPromise cannot be cast to 
> org.apache.flink.runtime.messages.JobManagerMessages$JobResponse}}
> According to git blame, the code has been touched in FLINK-1508.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[GitHub] flink pull request: StreamWindow abstraction + modular window comp...

2015-02-16 Thread rmetzger
Github user rmetzger commented on a diff in the pull request:

https://github.com/apache/flink/pull/395#discussion_r24737554
  
--- Diff: 
flink-staging/flink-streaming/flink-streaming-core/src/main/java/org/apache/flink/streaming/api/windowing/StreamWindowTypeInfo.java
 ---
@@ -0,0 +1,74 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUStreamRecordWARRANTIES OR CONDITIONS OF ANY KIND, either 
express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.streaming.api.windowing;
+
+import org.apache.flink.api.common.ExecutionConfig;
+import org.apache.flink.api.common.typeinfo.TypeInformation;
+import org.apache.flink.api.common.typeutils.TypeSerializer;
+
+public class StreamWindowTypeInfo extends 
TypeInformation> {
+
+   private static final long serialVersionUID = 1L;
+   TypeInformation innerType;
+
+   public StreamWindowTypeInfo(TypeInformation innerType) {
+   this.innerType = innerType;
+   }
+
+   public TypeInformation getInnerType() {
+   return innerType;
+   }
+
+   @Override
+   public boolean isBasicType() {
+   return innerType.isBasicType();
+   }
+
+   @Override
+   public boolean isTupleType() {
+   return innerType.isTupleType();
+   }
+
+   @Override
+   public int getArity() {
+   return innerType.getArity();
+   }
+
+   @Override
+   public Class> getTypeClass() {
+   // TODO Auto-generated method stub
--- End diff --

hehe


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] flink pull request: StreamWindow abstraction + modular window comp...

2015-02-16 Thread mbalassi
Github user mbalassi commented on a diff in the pull request:

https://github.com/apache/flink/pull/395#discussion_r24737330
  
--- Diff: docs/streaming_guide.md ---
@@ -857,7 +895,7 @@ After installing Docker an image can be pulled for each 
connector. Containers ca
 For the easiest set up create a jar with all the dependencies of the 
*flink-streaming-connectors* project.
 
 ~~~batch
-cd /PATH/TO/GIT/flink/flink-staging/flink-streaming-connectors
+cd /PATH/TO/GIT/incubator-flink/flink-addons/flink-streaming-connectors
 mvn assembly:assembly
--- End diff --

An amend here would be nice. :) I'll do it while merging.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[GitHub] flink pull request: StreamWindow abstraction + modular window comp...

2015-02-16 Thread mbalassi
Github user mbalassi commented on a diff in the pull request:

https://github.com/apache/flink/pull/395#discussion_r24737233
  
--- Diff: docs/streaming_guide.md ---
@@ -383,9 +389,40 @@ 
dataStream.groupBy(groupingField).window(Count.of(100)).every(…).max(field);
 
 This will create separate windows for different keys and apply the trigger 
and eviction policies on a per group basis.
 
+ Applying multiple transformations on a window
+Using the `WindowedDataStream` abstraction we can apply several 
transformations one after another on the discretised streams without having to 
re-discretise it:
+
+~~~java

+dataStream.window(Count.of(1000)).groupBy(firstKey).mapWindow(…).groupBy(secondKey).reduceWindow(…).flatten()
+~~~
+
+The above call would create global windows of 1000 elements group it by 
the first key and then apply a mapWindow transformation. The resulting windowed 
stream will then be grouped by the second key and further reduced. The results 
of the reduce transformation are then flattened.
--- End diff --

Nice example.


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at infrastruct...@apache.org or file a JIRA ticket
with INFRA.
---


[jira] [Created] (FLINK-1553) Kafka connectors rework

2015-02-16 Thread JIRA
Gábor Hermann created FLINK-1553:


 Summary: Kafka connectors rework
 Key: FLINK-1553
 URL: https://issues.apache.org/jira/browse/FLINK-1553
 Project: Flink
  Issue Type: Bug
  Components: Streaming
Reporter: Gábor Hermann
Assignee: Gábor Hermann


The Kafka connectors should be reworked and tested because it is not fully 
functioning (e.g. the source can only read partitions from the start).



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)