[jira] [Created] (FLINK-6809) side outputs documentation: wrong variable name in java example code
Petr Novotnik created FLINK-6809: Summary: side outputs documentation: wrong variable name in java example code Key: FLINK-6809 URL: https://issues.apache.org/jira/browse/FLINK-6809 Project: Flink Issue Type: Bug Components: Documentation Affects Versions: 1.3.0 Reporter: Petr Novotnik Priority: Trivial The first parameter to the {{processElement}} method in the example for side-outputs [here|https://ci.apache.org/projects/flink/flink-docs-release-1.3/dev/stream/side_output.html] is wrongly named {{input}}, but should read {{value}}. -- This message was sent by Atlassian JIRA (v6.3.15#6346)
[jira] [Created] (FLINK-6468) release 1.2.1 tag in git
Petr Novotnik created FLINK-6468: Summary: release 1.2.1 tag in git Key: FLINK-6468 URL: https://issues.apache.org/jira/browse/FLINK-6468 Project: Flink Issue Type: Bug Components: Build System Affects Versions: 1.2.1 Reporter: Petr Novotnik Priority: Minor It appears that the `release-1.2.1` tag in missing in the git repository. It would be great to have it. -- This message was sent by Atlassian JIRA (v6.3.15#6346)
[jira] [Created] (FLINK-6296) Retrieving execution plan fails succeeding execution attempt
Petr Novotnik created FLINK-6296: Summary: Retrieving execution plan fails succeeding execution attempt Key: FLINK-6296 URL: https://issues.apache.org/jira/browse/FLINK-6296 Project: Flink Issue Type: Bug Components: DataSet API, Job-Submission, YARN Affects Versions: 1.2.0 Reporter: Petr Novotnik Hello, calling {{ExecutionEnvironment#getExecutionPlan()}} prior to {{ExecutionEnvironment#execute()}} makes the later fail with the following message if I try to run my job on yarn, i.e. {{flink run -m yarn-cluster ...}}: {noformat} > Caused by: java.lang.RuntimeException: No new data sinks have been defined > since the last execution. The last execution refers to the latest call to > 'execute()', 'count()', 'collect()', or 'print()'. > at > org.apache.flink.api.java.ExecutionEnvironment.createProgramPlan(ExecutionEnvironment.java:1050) > at > org.apache.flink.api.java.ExecutionEnvironment.createProgramPlan(ExecutionEnvironment.java:1032) > at > org.apache.flink.client.program.ContextEnvironment.execute(ContextEnvironment.java:59) > at > org.apache.flink.api.java.ExecutionEnvironment.execute(ExecutionEnvironment.java:926) {noformat} It works fine when executing the job locally (and it works fine locally or even on yarn using the DataStream API, i.e. {{StreamExecutionEnvironment}}). So far I could track it down, the problem seems the "clearing" behaviour of {{ContextEnvironment#createProgramPlan(..)}}. This makes me wonder why {{#getExecutionPlan}} uses the one parameter version of {{#createProgramPlan(..)}}. Is "clearing" the sinks really required as part of {{getExecutionPlan()}}? I just check, in Flink 1.1.5, {{#getExecutionPlan(..)}} invokes {{#createProgramPlan("unnamed job", false)}}, hence, it would be good to document as a breaking change [in the migration guide|https://ci.apache.org/projects/flink/flink-docs-release-1.2/ops/upgrading.html] since the resulting error message is rather misleading. P. -- This message was sent by Atlassian JIRA (v6.3.15#6346)
[jira] [Created] (FLINK-6285) resolve hadoop-compatibility confusion
Petr Novotnik created FLINK-6285: Summary: resolve hadoop-compatibility confusion Key: FLINK-6285 URL: https://issues.apache.org/jira/browse/FLINK-6285 Project: Flink Issue Type: Bug Affects Versions: 1.2.0 Reporter: Petr Novotnik As of Flink 1.2.0, the binary distribution does not include classes from the `hadoop-compatibility` dependency anymore. ``` flink-1.2.0> for i in lib/*.jar; jar tf $i | grep WritableTypeInfo; end flink-1.2.0 [1]> # the above finds nothing ``` Therefore, it is necessary to copy the compatibility jar to flink's installation `lib/` directory (or a sub-directory) if one wishes to use hadoop input formats. Merely packaging the compatibility jar as part of an application's "fat jar" does not suffice, as code in [TypeExtractor#createHadoopWritableTypeInfo](https://github.com/apache/flink/blob/master/flink-core/src/main/java/org/apache/flink/api/java/typeutils/TypeExtractor.java#L1988)'s relies on being able to see the compatibility classes through the classloader `TypeExtractor` itself was loaded by. On yarn this seems not to be the case (e.g. when running the application through `flink run -m yarn-cluster ...`). * Ideally, we'd fix the class loading issue, such that flink's installation does not need to be altered, due to the need of a particular application. * Alternatively, we could include the hadoop-compatibility jar as part of the binary distribution and provide corresponding instructions, [1] and [2] seem to be good places. [1] https://ci.apache.org/projects/flink/flink-docs-release-1.2/dev/batch/hadoop_compatibility.html [2] https://ci.apache.org/projects/flink/flink-docs-release-1.2/dev/migration.html -- This message was sent by Atlassian JIRA (v6.3.15#6346)