[GitHub] [flink] SteNicholas commented on a change in pull request #13983: [FLINK-19989][python] Add collect operation in Python DataStream API

2020-11-27 Thread GitBox


SteNicholas commented on a change in pull request #13983:
URL: https://github.com/apache/flink/pull/13983#discussion_r531542160



##
File path: flink-python/pyflink/datastream/data_stream.py
##
@@ -123,7 +124,7 @@ def set_max_parallelism(self, max_parallelism: int) -> 
'DataStream':
 self._j_data_stream.setMaxParallelism(max_parallelism)
 return self
 
-def get_type(self) -> TypeInformation:
+def get_type(self) -> WrapperTypeInfo:

Review comment:
   @dianfu Because type checker of PyFlink cause this change. I didn't want 
to change `TypeInformation` to  `WrapperTypeInfo`. IMO,  the basic type is 
`WrapperTypeInfo` not `TypeInformation `.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] SteNicholas commented on a change in pull request #13983: [FLINK-19989][python] Add collect operation in Python DataStream API

2020-11-27 Thread GitBox


SteNicholas commented on a change in pull request #13983:
URL: https://github.com/apache/flink/pull/13983#discussion_r531982527



##
File path: 
flink-python/src/main/java/org/apache/flink/api/common/python/PythonBridgeUtils.java
##
@@ -266,6 +278,92 @@ private static Object getPickledBytesFromJavaObject(Object 
obj, LogicalType data
}
}
 
+   public static Object getPickledBytesFromJavaObject(Object obj, 
TypeInformation dataType) throws IOException {
+   Pickler pickler = new Pickler();
+   initialize();
+   if (obj == null) {
+   return new byte[0];
+   } else {
+   if (dataType instanceof SqlTimeTypeInfo) {
+   SqlTimeTypeInfo sqlTimeTypeInfo = 
SqlTimeTypeInfo.getInfoFor(dataType.getTypeClass());
+   if (sqlTimeTypeInfo == DATE) {
+   long time;
+   if (obj instanceof LocalDate) {
+   time = ((LocalDate) 
(obj)).toEpochDay();
+   } else {
+   time = ((Date) 
obj).toLocalDate().toEpochDay();

Review comment:
   @dianfu , for `DateType` and `DATE`, the object could both be pickled 
directly, but for `DateType` this has the type check to pickle the Date object. 
IMO, this should be unified for `DateType` and `DATE`.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14165: [FLINK-19687][table] Support to get execution plan from StatementSet

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14165:
URL: https://github.com/apache/flink/pull/14165#issuecomment-732020040


   
   ## CI report:
   
   * 7c554a8e388693a7f243f211726bbb763dddcb60 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10278)
 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10280)
 
   * df3a80a44692c8b13d0026673b463a451d44608d Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10283)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14165: [FLINK-19687][table] Support to get execution plan from StatementSet

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14165:
URL: https://github.com/apache/flink/pull/14165#issuecomment-732020040


   
   ## CI report:
   
   * 7c554a8e388693a7f243f211726bbb763dddcb60 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10278)
 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10280)
 
   * df3a80a44692c8b13d0026673b463a451d44608d UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-20288) Correct documentation about savepoint self-contained

2020-11-27 Thread Congxian Qiu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-20288?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17239908#comment-17239908
 ] 

Congxian Qiu commented on FLINK-20288:
--

[~yunta] thanks for creating this issue, maybe this is duplicated with 
FLINK-19381?

> Correct documentation about savepoint self-contained
> 
>
> Key: FLINK-20288
> URL: https://issues.apache.org/jira/browse/FLINK-20288
> Project: Flink
>  Issue Type: Bug
>  Components: Documentation, Runtime / Checkpointing
>Affects Versions: 1.11.0
>Reporter: Yun Tang
>Assignee: Yun Tang
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.12.0, 1.11.4
>
>
> Savepoint self-contained has been supported while the documentation still 
> remain as not supported, we should fix that description.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #8952: [FLINK-10868][flink-yarn] Add failure rater for resource manager

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #8952:
URL: https://github.com/apache/flink/pull/8952#issuecomment-513724324


   
   ## CI report:
   
   * b59b8959f9bd54ae3a7bc4923d2d1aa5b780c331 Travis: 
[SUCCESS](https://travis-ci.com/flink-ci/flink/builds/147257697) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4798)
 
   * 58784051e00030de475f361d314c1c0473ff2b8a Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10279)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] klion26 commented on pull request #13488: [FLINK-19381][docs] Fix docs for savepoint relocation

2020-11-27 Thread GitBox


klion26 commented on pull request #13488:
URL: https://github.com/apache/flink/pull/13488#issuecomment-735047072


   kindly ping @NicoK 



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] klion26 commented on pull request #14175: [FLINK-20288][docs] Correct documentation about savepoint self-contained

2020-11-27 Thread GitBox


klion26 commented on pull request #14175:
URL: https://github.com/apache/flink/pull/14175#issuecomment-735047044


   @Myasuka thanks for your contribution, maybe this is duplicated with 
FLINK-19381/[PR](https://github.com/apache/flink/pull/13488) ?



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Resolved] (FLINK-20384) Broken Link in deployment/ha/kubernetes_ha.zh.md

2020-11-27 Thread Congxian Qiu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20384?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Congxian Qiu resolved FLINK-20384.
--
Fix Version/s: 1.12.0
   Resolution: Fixed

merged into master d5b15652fc85fe4b0929e7faf274b46c04b7e924
1.12 2e8d9b9489a1ea33cecdcfc4d84912d5c68c1bf0

> Broken Link in deployment/ha/kubernetes_ha.zh.md
> 
>
> Key: FLINK-20384
> URL: https://issues.apache.org/jira/browse/FLINK-20384
> Project: Flink
>  Issue Type: Bug
>  Components: Deployment / Kubernetes, Documentation
>Affects Versions: 1.12.0, 1.13.0
>Reporter: Huang Xingbo
>Assignee: Huang Xingbo
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.12.0
>
>
> When executing the script build_docs.sh, it will throw the following 
> exception:
> {code:java}
> Liquid Exception: Could not find document 
> 'deployment/resource-providers/standalone/kubernetes.md' in tag 'link'. Make 
> sure the document exists and the path is correct. in 
> deployment/ha/kubernetes_ha.zh.md Could not find document 
> 'deployment/resource-providers/standalone/kubernetes.md' in tag 'link'.
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (FLINK-20384) Broken Link in deployment/ha/kubernetes_ha.zh.md

2020-11-27 Thread Congxian Qiu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20384?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Congxian Qiu reassigned FLINK-20384:


Assignee: Huang Xingbo

> Broken Link in deployment/ha/kubernetes_ha.zh.md
> 
>
> Key: FLINK-20384
> URL: https://issues.apache.org/jira/browse/FLINK-20384
> Project: Flink
>  Issue Type: Bug
>  Components: Deployment / Kubernetes, Documentation
>Affects Versions: 1.12.0, 1.13.0
>Reporter: Huang Xingbo
>Assignee: Huang Xingbo
>Priority: Major
>  Labels: pull-request-available
>
> When executing the script build_docs.sh, it will throw the following 
> exception:
> {code:java}
> Liquid Exception: Could not find document 
> 'deployment/resource-providers/standalone/kubernetes.md' in tag 'link'. Make 
> sure the document exists and the path is correct. in 
> deployment/ha/kubernetes_ha.zh.md Could not find document 
> 'deployment/resource-providers/standalone/kubernetes.md' in tag 'link'.
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] klion26 merged pull request #14242: [FLINK-20384][doc] Fix Wrong Links in Some Chinese Docs

2020-11-27 Thread GitBox


klion26 merged pull request #14242:
URL: https://github.com/apache/flink/pull/14242


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14165: [FLINK-19687][table] Support to get execution plan from StatementSet

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14165:
URL: https://github.com/apache/flink/pull/14165#issuecomment-732020040


   
   ## CI report:
   
   * 7c554a8e388693a7f243f211726bbb763dddcb60 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10278)
 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10280)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] V1ncentzzZ commented on pull request #14165: [FLINK-19687][table] Support to get execution plan from StatementSet

2020-11-27 Thread GitBox


V1ncentzzZ commented on pull request #14165:
URL: https://github.com/apache/flink/pull/14165#issuecomment-735046296


   @flinkbot run azure



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #8952: [FLINK-10868][flink-yarn] Add failure rater for resource manager

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #8952:
URL: https://github.com/apache/flink/pull/8952#issuecomment-513724324


   
   ## CI report:
   
   * b59b8959f9bd54ae3a7bc4923d2d1aa5b780c331 Travis: 
[SUCCESS](https://travis-ci.com/flink-ci/flink/builds/147257697) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=4798)
 
   * 58784051e00030de475f361d314c1c0473ff2b8a UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] HuangZhenQiu commented on pull request #8952: [FLINK-10868][flink-yarn] Add failure rater for resource manager

2020-11-27 Thread GitBox


HuangZhenQiu commented on pull request #8952:
URL: https://github.com/apache/flink/pull/8952#issuecomment-735044796


   @tillrohrmann @zentol @azagrebin Rebased master. Please review it when you 
have time.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14165: [FLINK-19687][table] Support to get execution plan from StatementSet

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14165:
URL: https://github.com/apache/flink/pull/14165#issuecomment-732020040


   
   ## CI report:
   
   * 7c554a8e388693a7f243f211726bbb763dddcb60 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10278)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14165: [FLINK-19687][table] Support to get execution plan from StatementSet

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14165:
URL: https://github.com/apache/flink/pull/14165#issuecomment-732020040


   
   ## CI report:
   
   * 452c291a21493b2ae4e722de1879b5a94fca9ca9 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=9935)
 
   * 7c554a8e388693a7f243f211726bbb763dddcb60 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10278)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] V1ncentzzZ commented on pull request #14165: [FLINK-19687][table] Support to get execution plan from StatementSet

2020-11-27 Thread GitBox


V1ncentzzZ commented on pull request #14165:
URL: https://github.com/apache/flink/pull/14165#issuecomment-735038657


   Hi @wuchong , I removed entry `Streaming Execution Plan` and add some tests.
   
   Hi @godfreyhe , Can i submit another issue for docs?



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14165: [FLINK-19687][table] Support to get execution plan from StatementSet

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14165:
URL: https://github.com/apache/flink/pull/14165#issuecomment-732020040


   
   ## CI report:
   
   * 452c291a21493b2ae4e722de1879b5a94fca9ca9 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=9935)
 
   * 7c554a8e388693a7f243f211726bbb763dddcb60 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-20410) SQLClientSchemaRegistryITCase.testWriting failed with "Subject 'user_behavior' not found.; error code: 40401"

2020-11-27 Thread Dian Fu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-20410?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17239868#comment-17239868
 ] 

Dian Fu commented on FLINK-20410:
-

cc [~dwysakowicz]

> SQLClientSchemaRegistryITCase.testWriting failed with "Subject 
> 'user_behavior' not found.; error code: 40401"
> -
>
> Key: FLINK-20410
> URL: https://issues.apache.org/jira/browse/FLINK-20410
> Project: Flink
>  Issue Type: Bug
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
>Affects Versions: 1.12.0, 1.13.0
>Reporter: Dian Fu
>Priority: Critical
>  Labels: test-stability
> Fix For: 1.12.0
>
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=10276=logs=91bf6583-3fb2-592f-e4d4-d79d79c3230a=3425d8ba-5f03-540a-c64b-51b8481bf7d6
> {code}
> 2020-11-28T01:14:08.6444305Z Nov 28 01:14:08 [ERROR] 
> testWriting(org.apache.flink.tests.util.kafka.SQLClientSchemaRegistryITCase)  
> Time elapsed: 74.818 s  <<< ERROR!
> 2020-11-28T01:14:08.6445353Z Nov 28 01:14:08 
> io.confluent.kafka.schemaregistry.client.rest.exceptions.RestClientException: 
> Subject 'user_behavior' not found.; error code: 40401
> 2020-11-28T01:14:08.6446071Z Nov 28 01:14:08  at 
> io.confluent.kafka.schemaregistry.client.rest.RestService.sendHttpRequest(RestService.java:292)
> 2020-11-28T01:14:08.6446910Z Nov 28 01:14:08  at 
> io.confluent.kafka.schemaregistry.client.rest.RestService.httpRequest(RestService.java:352)
> 2020-11-28T01:14:08.6447522Z Nov 28 01:14:08  at 
> io.confluent.kafka.schemaregistry.client.rest.RestService.getAllVersions(RestService.java:769)
> 2020-11-28T01:14:08.6448352Z Nov 28 01:14:08  at 
> io.confluent.kafka.schemaregistry.client.rest.RestService.getAllVersions(RestService.java:760)
> 2020-11-28T01:14:08.6449091Z Nov 28 01:14:08  at 
> io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.getAllVersions(CachedSchemaRegistryClient.java:364)
> 2020-11-28T01:14:08.6449878Z Nov 28 01:14:08  at 
> org.apache.flink.tests.util.kafka.SQLClientSchemaRegistryITCase.testWriting(SQLClientSchemaRegistryITCase.java:195)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-20410) SQLClientSchemaRegistryITCase.testWriting failed with "Subject 'user_behavior' not found.; error code: 40401"

2020-11-27 Thread Dian Fu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20410?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dian Fu updated FLINK-20410:

Priority: Critical  (was: Major)

> SQLClientSchemaRegistryITCase.testWriting failed with "Subject 
> 'user_behavior' not found.; error code: 40401"
> -
>
> Key: FLINK-20410
> URL: https://issues.apache.org/jira/browse/FLINK-20410
> Project: Flink
>  Issue Type: Bug
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
>Affects Versions: 1.12.0
>Reporter: Dian Fu
>Priority: Critical
>  Labels: test-stability
> Fix For: 1.12.0
>
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=10276=logs=91bf6583-3fb2-592f-e4d4-d79d79c3230a=3425d8ba-5f03-540a-c64b-51b8481bf7d6
> {code}
> 2020-11-28T01:14:08.6444305Z Nov 28 01:14:08 [ERROR] 
> testWriting(org.apache.flink.tests.util.kafka.SQLClientSchemaRegistryITCase)  
> Time elapsed: 74.818 s  <<< ERROR!
> 2020-11-28T01:14:08.6445353Z Nov 28 01:14:08 
> io.confluent.kafka.schemaregistry.client.rest.exceptions.RestClientException: 
> Subject 'user_behavior' not found.; error code: 40401
> 2020-11-28T01:14:08.6446071Z Nov 28 01:14:08  at 
> io.confluent.kafka.schemaregistry.client.rest.RestService.sendHttpRequest(RestService.java:292)
> 2020-11-28T01:14:08.6446910Z Nov 28 01:14:08  at 
> io.confluent.kafka.schemaregistry.client.rest.RestService.httpRequest(RestService.java:352)
> 2020-11-28T01:14:08.6447522Z Nov 28 01:14:08  at 
> io.confluent.kafka.schemaregistry.client.rest.RestService.getAllVersions(RestService.java:769)
> 2020-11-28T01:14:08.6448352Z Nov 28 01:14:08  at 
> io.confluent.kafka.schemaregistry.client.rest.RestService.getAllVersions(RestService.java:760)
> 2020-11-28T01:14:08.6449091Z Nov 28 01:14:08  at 
> io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.getAllVersions(CachedSchemaRegistryClient.java:364)
> 2020-11-28T01:14:08.6449878Z Nov 28 01:14:08  at 
> org.apache.flink.tests.util.kafka.SQLClientSchemaRegistryITCase.testWriting(SQLClientSchemaRegistryITCase.java:195)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-20410) SQLClientSchemaRegistryITCase.testWriting failed with "Subject 'user_behavior' not found.; error code: 40401"

2020-11-27 Thread Dian Fu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20410?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dian Fu updated FLINK-20410:

Affects Version/s: 1.13.0

> SQLClientSchemaRegistryITCase.testWriting failed with "Subject 
> 'user_behavior' not found.; error code: 40401"
> -
>
> Key: FLINK-20410
> URL: https://issues.apache.org/jira/browse/FLINK-20410
> Project: Flink
>  Issue Type: Bug
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
>Affects Versions: 1.12.0, 1.13.0
>Reporter: Dian Fu
>Priority: Critical
>  Labels: test-stability
> Fix For: 1.12.0
>
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=10276=logs=91bf6583-3fb2-592f-e4d4-d79d79c3230a=3425d8ba-5f03-540a-c64b-51b8481bf7d6
> {code}
> 2020-11-28T01:14:08.6444305Z Nov 28 01:14:08 [ERROR] 
> testWriting(org.apache.flink.tests.util.kafka.SQLClientSchemaRegistryITCase)  
> Time elapsed: 74.818 s  <<< ERROR!
> 2020-11-28T01:14:08.6445353Z Nov 28 01:14:08 
> io.confluent.kafka.schemaregistry.client.rest.exceptions.RestClientException: 
> Subject 'user_behavior' not found.; error code: 40401
> 2020-11-28T01:14:08.6446071Z Nov 28 01:14:08  at 
> io.confluent.kafka.schemaregistry.client.rest.RestService.sendHttpRequest(RestService.java:292)
> 2020-11-28T01:14:08.6446910Z Nov 28 01:14:08  at 
> io.confluent.kafka.schemaregistry.client.rest.RestService.httpRequest(RestService.java:352)
> 2020-11-28T01:14:08.6447522Z Nov 28 01:14:08  at 
> io.confluent.kafka.schemaregistry.client.rest.RestService.getAllVersions(RestService.java:769)
> 2020-11-28T01:14:08.6448352Z Nov 28 01:14:08  at 
> io.confluent.kafka.schemaregistry.client.rest.RestService.getAllVersions(RestService.java:760)
> 2020-11-28T01:14:08.6449091Z Nov 28 01:14:08  at 
> io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.getAllVersions(CachedSchemaRegistryClient.java:364)
> 2020-11-28T01:14:08.6449878Z Nov 28 01:14:08  at 
> org.apache.flink.tests.util.kafka.SQLClientSchemaRegistryITCase.testWriting(SQLClientSchemaRegistryITCase.java:195)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-20410) SQLClientSchemaRegistryITCase.testWriting failed with "Subject 'user_behavior' not found.; error code: 40401"

2020-11-27 Thread Dian Fu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20410?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dian Fu updated FLINK-20410:

Labels: test-stability  (was: )

> SQLClientSchemaRegistryITCase.testWriting failed with "Subject 
> 'user_behavior' not found.; error code: 40401"
> -
>
> Key: FLINK-20410
> URL: https://issues.apache.org/jira/browse/FLINK-20410
> Project: Flink
>  Issue Type: Bug
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
>Affects Versions: 1.12.0
>Reporter: Dian Fu
>Priority: Major
>  Labels: test-stability
> Fix For: 1.12.0
>
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=10276=logs=91bf6583-3fb2-592f-e4d4-d79d79c3230a=3425d8ba-5f03-540a-c64b-51b8481bf7d6
> {code}
> 2020-11-28T01:14:08.6444305Z Nov 28 01:14:08 [ERROR] 
> testWriting(org.apache.flink.tests.util.kafka.SQLClientSchemaRegistryITCase)  
> Time elapsed: 74.818 s  <<< ERROR!
> 2020-11-28T01:14:08.6445353Z Nov 28 01:14:08 
> io.confluent.kafka.schemaregistry.client.rest.exceptions.RestClientException: 
> Subject 'user_behavior' not found.; error code: 40401
> 2020-11-28T01:14:08.6446071Z Nov 28 01:14:08  at 
> io.confluent.kafka.schemaregistry.client.rest.RestService.sendHttpRequest(RestService.java:292)
> 2020-11-28T01:14:08.6446910Z Nov 28 01:14:08  at 
> io.confluent.kafka.schemaregistry.client.rest.RestService.httpRequest(RestService.java:352)
> 2020-11-28T01:14:08.6447522Z Nov 28 01:14:08  at 
> io.confluent.kafka.schemaregistry.client.rest.RestService.getAllVersions(RestService.java:769)
> 2020-11-28T01:14:08.6448352Z Nov 28 01:14:08  at 
> io.confluent.kafka.schemaregistry.client.rest.RestService.getAllVersions(RestService.java:760)
> 2020-11-28T01:14:08.6449091Z Nov 28 01:14:08  at 
> io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.getAllVersions(CachedSchemaRegistryClient.java:364)
> 2020-11-28T01:14:08.6449878Z Nov 28 01:14:08  at 
> org.apache.flink.tests.util.kafka.SQLClientSchemaRegistryITCase.testWriting(SQLClientSchemaRegistryITCase.java:195)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-20410) SQLClientSchemaRegistryITCase.testWriting failed with "Subject 'user_behavior' not found.; error code: 40401"

2020-11-27 Thread Dian Fu (Jira)
Dian Fu created FLINK-20410:
---

 Summary: SQLClientSchemaRegistryITCase.testWriting failed with 
"Subject 'user_behavior' not found.; error code: 40401"
 Key: FLINK-20410
 URL: https://issues.apache.org/jira/browse/FLINK-20410
 Project: Flink
  Issue Type: Bug
  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
Affects Versions: 1.12.0
Reporter: Dian Fu
 Fix For: 1.12.0


https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=10276=logs=91bf6583-3fb2-592f-e4d4-d79d79c3230a=3425d8ba-5f03-540a-c64b-51b8481bf7d6

{code}
2020-11-28T01:14:08.6444305Z Nov 28 01:14:08 [ERROR] 
testWriting(org.apache.flink.tests.util.kafka.SQLClientSchemaRegistryITCase)  
Time elapsed: 74.818 s  <<< ERROR!
2020-11-28T01:14:08.6445353Z Nov 28 01:14:08 
io.confluent.kafka.schemaregistry.client.rest.exceptions.RestClientException: 
Subject 'user_behavior' not found.; error code: 40401
2020-11-28T01:14:08.6446071Z Nov 28 01:14:08at 
io.confluent.kafka.schemaregistry.client.rest.RestService.sendHttpRequest(RestService.java:292)
2020-11-28T01:14:08.6446910Z Nov 28 01:14:08at 
io.confluent.kafka.schemaregistry.client.rest.RestService.httpRequest(RestService.java:352)
2020-11-28T01:14:08.6447522Z Nov 28 01:14:08at 
io.confluent.kafka.schemaregistry.client.rest.RestService.getAllVersions(RestService.java:769)
2020-11-28T01:14:08.6448352Z Nov 28 01:14:08at 
io.confluent.kafka.schemaregistry.client.rest.RestService.getAllVersions(RestService.java:760)
2020-11-28T01:14:08.6449091Z Nov 28 01:14:08at 
io.confluent.kafka.schemaregistry.client.CachedSchemaRegistryClient.getAllVersions(CachedSchemaRegistryClient.java:364)
2020-11-28T01:14:08.6449878Z Nov 28 01:14:08at 
org.apache.flink.tests.util.kafka.SQLClientSchemaRegistryITCase.testWriting(SQLClientSchemaRegistryITCase.java:195)
{code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] 1996fanrui commented on pull request #13885: [FLINK-19911] add read buffer for input stream

2020-11-27 Thread GitBox


1996fanrui commented on pull request #13885:
URL: https://github.com/apache/flink/pull/13885#issuecomment-735033850


   
   
   
   > Hi @1996fanrui ,
   > 
   > that's an interesting discovery and investigation that you did there!
   > 
   > I think the approach on the filesystem level is also much better than the 
previous way. Let's try to not change any public API (FileSystem) as this would 
slow down the progress.
   > 
   > I'd probably focus on Hadoop file systems entirely (for now). What I'd 
propose is the following:
   > 
   > * Use `HadoopFsFactory#configure` to extract the buffer size and pass it 
to the `ctor` of all filesystems created by the factory.
   > * Use that default buffer size in `HadoopFileSystem#open(Path)` to call 
`#open(Path, int)`.
   > * `HadoopFileSystem#open(Path, int)` should use the buffer size both in 
the call to Hadoop and to wrap it as you did as in the `BufferedFSInputStream`. 
I dug a bit into the Hadoop code and noticed that the cache is by default just 
4kb. So even if we have cache on top of it with 64kb, we would still need to 
ask Hadoop several times.
   > 
   > So, that means you are not adding any new methods, but just modify 
existing ones.
   
   Hi @Myasuka @AHeise ,
   
   Based on the above design, there are some questions that I hope to be 
answered:
   
   - `buffer size` is passed to HadoopFileSystem through Constructor. It means 
that HadoopFileSystem needs to add a new Constructor: 
`HadoopFileSystem(FileSystem, bufferSize)`. HadoopFsFactory will call the new 
constructor.
   
   - The old constructor `HadoopFileSystem(FileSystem)` is still called by 
other FsFactory. So other FsFactory will not be able to improve performance. 
For example: OSSFileSystemFactory.
   
   Question: Do other FsFactory related to HadoopFileSystem need to improve 
performance? If needed, are there other better designs. (Can there be a way not 
to modify many FsFactory?)
   
   Thanks.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Created] (FLINK-20409) Migrate test_kubernetes_pyflink_application.sh

2020-11-27 Thread Huang Xingbo (Jira)
Huang Xingbo created FLINK-20409:


 Summary: Migrate test_kubernetes_pyflink_application.sh
 Key: FLINK-20409
 URL: https://issues.apache.org/jira/browse/FLINK-20409
 Project: Flink
  Issue Type: Sub-task
  Components: API / Python, Tests
Reporter: Huang Xingbo






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-20408) Migrate test_pyflink_yarn.sh

2020-11-27 Thread Huang Xingbo (Jira)
Huang Xingbo created FLINK-20408:


 Summary: Migrate test_pyflink_yarn.sh
 Key: FLINK-20408
 URL: https://issues.apache.org/jira/browse/FLINK-20408
 Project: Flink
  Issue Type: Sub-task
  Components: API / Python, Tests
Reporter: Huang Xingbo






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-20407) Migrate test_pyflink.sh

2020-11-27 Thread Huang Xingbo (Jira)
Huang Xingbo created FLINK-20407:


 Summary: Migrate test_pyflink.sh
 Key: FLINK-20407
 URL: https://issues.apache.org/jira/browse/FLINK-20407
 Project: Flink
  Issue Type: Sub-task
  Components: API / Python, Tests
Reporter: Huang Xingbo






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink-web] ClownfishYang commented on a change in pull request #394: [FLINK-13680] Translate "Common Rules" page into Chinese

2020-11-27 Thread GitBox


ClownfishYang commented on a change in pull request #394:
URL: https://github.com/apache/flink-web/pull/394#discussion_r531823381



##
File path: contributing/code-style-and-quality-common.zh.md
##
@@ -30,83 +32,88 @@ Each file must include the Apache license information as a 
header.
  */
 ```
 
-## 2. Tools
+
+
 
-We recommend to follow the [IDE Setup 
Guide](https://ci.apache.org/projects/flink/flink-docs-master/flinkDev/ide_setup.html#checkstyle-for-java)
 to get IDE tooling configured.
+## 2. 工具
+
+我们建议你按照 [IDE 
设置指南](https://ci.apache.org/projects/flink/flink-docs-master/flinkDev/ide_setup.html#checkstyle-for-java)
 配置 IDE 工具。
 
 
 
 
-### Warnings
+### 警告
+
+* 我们争取实现零警告
+* 尽管现有的代码中存在许多警告,但新的修改不应该有任何编译器警告
+* 如果不能用合理的方式处理警告(某些情况下使用泛型时)也应该添加注释以压制警告
+* 弃用方法时,检查是否会引入其他的警告
 
-* We strive for zero warnings
-* Even though there are many warnings in existing code, new changes should not 
add any additional compiler warnings
-* If it is not possible to address the warning in a sane way (in some cases 
when working with generics) add an annotation to suppress the warning
-* When deprecating methods, check that this does not introduce additional 
warnings
 
 
+
 
-## 3. Comments And Code Readability
+## 3. 注释和代码可读性
 
 
-### Comments
+### 注释
 
-**Golden rule: Comment as much as necessary to support code understanding, but 
don’t add redundant information.**
+**黄金法则: 尽可能多的注释以支持代码的理解,但不要添加多余的信息。**
 
-Think about
+思考
 
-* What is the code doing?
-* How does the code do this?
-* Why is the code like that?
+* What 代码在做什么?
+* How 代码怎么做到的?
+* Why 代码为什么是这样的?
 
-The code alone should explain as much as possible the “what” and the “how”
+代码本身应该尽可能的解释 “what” 和 “how”
 
-* Use JavaDocs to describe the roles of classes and the contracts of methods, 
in cases where the contract is not obvious or intuitive from the method name 
(the “what”).
-* The flow of the code should give a good description of the “how”.
-Think of variable and method names as part of the code documenting itself.
-* It often makes reading the code easier if larger blocks that form a unit are 
moved into a private method with a descriptive name of what that block is doing
+* 使用 JavaDocs 来描述类的作用和方法的协议,以防止不能从方法名看出协议(“what”)。
+* 代码流程应该能够很好的描述 “how”。

Review comment:
   ```suggestion
   * 代码流程应该能够很好的描述 “how”,能够将变量和方法名看作是代码文档的一部分。
   ```





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink-web] ClownfishYang commented on a change in pull request #394: [FLINK-13680] Translate "Common Rules" page into Chinese

2020-11-27 Thread GitBox


ClownfishYang commented on a change in pull request #394:
URL: https://github.com/apache/flink-web/pull/394#discussion_r531823305



##
File path: contributing/code-style-and-quality-common.zh.md
##
@@ -30,95 +32,101 @@ Each file must include the Apache license information as a 
header.
  */
 ```
 
-## 2. Tools
+
+
+
 
-We recommend to follow the [IDE Setup 
Guide](https://ci.apache.org/projects/flink/flink-docs-master/flinkDev/ide_setup.html#checkstyle-for-java)
 to get IDE tooling configured.
+## 2. 工具
+
+我们建议你按照 [IDE 
设置指南](https://ci.apache.org/projects/flink/flink-docs-master/flinkDev/ide_setup.html#checkstyle-for-java)
 配置 IDE 工具。
 
 
 
 
-### Warnings
+### 警告
+
+* 我们争取实现零警告
+* 尽管现有的代码中存在许多警告,但新的修改不应该有任何编译器警告
+* 如果不能用合理的方式处理警告(某些情况下使用泛型时)也应该添加注释以压制警告
+* 弃用方法时,检查是否会引入其他的警告
 
-* We strive for zero warnings
-* Even though there are many warnings in existing code, new changes should not 
add any additional compiler warnings
-* If it is not possible to address the warning in a sane way (in some cases 
when working with generics) add an annotation to suppress the warning
-* When deprecating methods, check that this does not introduce additional 
warnings
 
 
+
 
-## 3. Comments And Code Readability
+## 3. 注释和代码可读性
 
 
-### Comments
+### 注释
 
-**Golden rule: Comment as much as necessary to support code understanding, but 
don’t add redundant information.**
+**黄金法则: 尽可能多的注释以支持代码的理解,但不要添加多余的信息。**
 
-Think about
+思考
 
-* What is the code doing?
-* How does the code do this?
-* Why is the code like that?
+* What 代码在做什么?
+* How 代码怎么做到的?
+* Why 代码为什么是这样的?
 
-The code alone should explain as much as possible the “what” and the “how”
+代码本身应该尽可能的解释 “what” 和 “how”
 
-* Use JavaDocs to describe the roles of classes and the contracts of methods, 
in cases where the contract is not obvious or intuitive from the method name 
(the “what”).
-* The flow of the code should give a good description of the “how”.
-Think of variable and method names as part of the code documenting itself.
-* It often makes reading the code easier if larger blocks that form a unit are 
moved into a private method with a descriptive name of what that block is doing
+* 使用 JavaDocs 来描述类的作用和方法的协议,以防止不能从方法名看出协议(“what”)。
+* 代码流程应该能够很好的描述 “how”。
+将变量和方法名看作是代码文档的一部分。

Review comment:
   ```suggestion
   ```





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink-web] ClownfishYang commented on a change in pull request #394: [FLINK-13680] Translate "Common Rules" page into Chinese

2020-11-27 Thread GitBox


ClownfishYang commented on a change in pull request #394:
URL: https://github.com/apache/flink-web/pull/394#discussion_r531822683



##
File path: contributing/code-style-and-quality-common.zh.md
##
@@ -30,95 +32,101 @@ Each file must include the Apache license information as a 
header.
  */
 ```
 
-## 2. Tools
+
+
+
 
-We recommend to follow the [IDE Setup 
Guide](https://ci.apache.org/projects/flink/flink-docs-master/flinkDev/ide_setup.html#checkstyle-for-java)
 to get IDE tooling configured.
+## 2. 工具
+
+我们建议你按照 [IDE 
设置指南](https://ci.apache.org/projects/flink/flink-docs-master/flinkDev/ide_setup.html#checkstyle-for-java)
 配置 IDE 工具。
 
 
 
 
-### Warnings
+### 警告
+
+* 我们争取实现零警告
+* 尽管现有的代码中存在许多警告,但新的修改不应该有任何编译器警告
+* 如果不能用合理的方式处理警告(某些情况下使用泛型时)也应该添加注释以压制警告
+* 弃用方法时,检查是否会引入其他的警告
 
-* We strive for zero warnings
-* Even though there are many warnings in existing code, new changes should not 
add any additional compiler warnings
-* If it is not possible to address the warning in a sane way (in some cases 
when working with generics) add an annotation to suppress the warning
-* When deprecating methods, check that this does not introduce additional 
warnings
 
 
+
 
-## 3. Comments And Code Readability
+## 3. 注释和代码可读性
 
 
-### Comments
+### 注释
 
-**Golden rule: Comment as much as necessary to support code understanding, but 
don’t add redundant information.**
+**黄金法则: 尽可能多的注释以支持代码的理解,但不要添加多余的信息。**
 
-Think about
+思考
 
-* What is the code doing?
-* How does the code do this?
-* Why is the code like that?
+* What 代码在做什么?
+* How 代码怎么做到的?
+* Why 代码为什么是这样的?
 
-The code alone should explain as much as possible the “what” and the “how”
+代码本身应该尽可能的解释 “what” 和 “how”
 
-* Use JavaDocs to describe the roles of classes and the contracts of methods, 
in cases where the contract is not obvious or intuitive from the method name 
(the “what”).
-* The flow of the code should give a good description of the “how”.
-Think of variable and method names as part of the code documenting itself.
-* It often makes reading the code easier if larger blocks that form a unit are 
moved into a private method with a descriptive name of what that block is doing
+* 使用 JavaDocs 来描述类的作用和方法的协议,以防止不能从方法名看出协议(“what”)。
+* 代码流程应该能够很好的描述 “how”。
+将变量和方法名看作是代码文档的一部分。
+* 如果将组成单元较大块的代码移动到 private 方法中,并且该方法具有描述性的名称,那么代码的可读性就会更强。
 
-In-code comments help explain the “why”
+代码内部的注释有助于解释 “why”
 
-* For example `// this specific code layout helps the JIT to better do this or 
that`
-* Or `// nulling out this field here means future write attempts are fail-fast`
-* Or `// for arguments with which this method is actually called, this 
seemingly naive approach works actually better than any optimized/smart version`
+* 例如 `// 这种特定的代码布局可以让 JIT 更好的进行工作`
+* 或 `// 此字段为空将会导致写入尝试 fail-fast`
+* 或 `// 用于实际调用该方法的参数,这种看似简单的方式实际上比任何优化/智能版本更好`
 
-In-code comments should not state redundant information about the “what” and 
“how” that is already obvious in the code itself.
+在代码注释中,不应该有关于 “what” 和 “how” 这么明显的冗余信息。
 
-JavaDocs should not state meaningless information (just to satisfy the 
Checkstyle checker).
+JavaDocs 不应该说明无意义的信息 (这么做只是为了满足 Checkstyle 的检查)。
 
-__Don’t:__
+__反例:__
 
 ```
 /**
- * The symbol expression.
+ * 符号表达式。
  */
 public class CommonSymbolExpression {}
 ```
-__Do:__
+__正例:__
 
 ```
 /**
- * An expression that wraps a single specific symbol.
- * A symbol could be a unit, an alias, a variable, etc.
+ * 包含单个特定符号的表达式。
+ * 符号可以是 Unit、Alias、Variable 等等。
  */
 public class CommonSymbolExpression {}
 ```
 
 
-### Branches and Nesting
+### 分支和嵌套
 
-Avoid deep nesting of scopes, by flipping the if condition and exiting early.
+通过对 if 条件取反并提前退出,避免超出嵌套深度的范围。
 
-__Don’t:__
+__反例:__
 
 ```
 if (a) {
-if (b) {
+if (b) { 

Review comment:
   ```suggestion
   if (b) {
   ```





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink-web] ClownfishYang commented on a change in pull request #394: [FLINK-13680] Translate "Common Rules" page into Chinese

2020-11-27 Thread GitBox


ClownfishYang commented on a change in pull request #394:
URL: https://github.com/apache/flink-web/pull/394#discussion_r531822294



##
File path: contributing/code-style-and-quality-common.zh.md
##
@@ -30,95 +32,101 @@ Each file must include the Apache license information as a 
header.
  */
 ```
 
-## 2. Tools
+
+
+

Review comment:
   ```suggestion
   ```





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink-statefun] sjwiesman closed pull request #180: [FLINK-18810][sdk] Golang remote functions SDK

2020-11-27 Thread GitBox


sjwiesman closed pull request #180:
URL: https://github.com/apache/flink-statefun/pull/180


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-18810) Golang remote functions SDK

2020-11-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18810?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-18810:
---
Labels: pull-request-available  (was: )

> Golang remote functions SDK
> ---
>
> Key: FLINK-18810
> URL: https://issues.apache.org/jira/browse/FLINK-18810
> Project: Flink
>  Issue Type: New Feature
>  Components: Stateful Functions
>Affects Versions: statefun-2.3.0
>Reporter: Francesco Guardiani
>Assignee: Seth Wiesman
>Priority: Major
>  Labels: pull-request-available
>
> Hi,
> I was wondering if there's already some WIP for a Golang SDK to create remote 
> functions. If not, I'm willing to give it a try.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink-statefun] sjwiesman opened a new pull request #180: [FLINK-18810][sdk] Golang remote functions SDK

2020-11-27 Thread GitBox


sjwiesman opened a new pull request #180:
URL: https://github.com/apache/flink-statefun/pull/180


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14238: [FLINK-20347] Rework YARN documentation page

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14238:
URL: https://github.com/apache/flink/pull/14238#issuecomment-734336545


   
   ## CI report:
   
   * 5b9ca356af610ffa7faf50a88372a5e22f7c7185 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10273)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14252: [BP-1.12][FLINK-20382][runtime] Fail hard when JobMaster cannot start scheduling of a job

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14252:
URL: https://github.com/apache/flink/pull/14252#issuecomment-734817896


   
   ## CI report:
   
   * 9bff590490e977520cf92133855bd8a889f31fb4 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10271)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14251: [FLINK-20382][runtime] Fail hard when JobMaster cannot start scheduling of a job

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14251:
URL: https://github.com/apache/flink/pull/14251#issuecomment-734807691


   
   ## CI report:
   
   * 69beddd215cef8366db8a9c9e71b0dad0469b22d Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10270)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14253: [BP-1.11][FLINK-20382][runtime] Fail hard when JobMaster cannot start scheduling of a job

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14253:
URL: https://github.com/apache/flink/pull/14253#issuecomment-734818578


   
   ## CI report:
   
   * 6389f5ba95deedd53497b087df8da73d49d12929 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10272)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Created] (FLINK-20406) Return the Checkpoint ID of the restored Checkpoint in CheckpointCoordinator.restoreLatestCheckpointedStateToSubtasks()

2020-11-27 Thread Stephan Ewen (Jira)
Stephan Ewen created FLINK-20406:


 Summary: Return the Checkpoint ID of the restored Checkpoint in 
CheckpointCoordinator.restoreLatestCheckpointedStateToSubtasks()
 Key: FLINK-20406
 URL: https://issues.apache.org/jira/browse/FLINK-20406
 Project: Flink
  Issue Type: Sub-task
  Components: Runtime / Checkpointing
Reporter: Stephan Ewen
Assignee: Stephan Ewen
 Fix For: 1.12.0


To allow the scheduler to notify Operator Coordinators of subtask restores 
(local failover), we need to know which checkpoint ID was restored. 

This change does not adjust the other restore methods of the Checkpoint 
Coordinator, because the fact that the Scheduler needs to be involved in the 
subtask restore notification at all is only due to a shortcoming of the 
Checkpoint Coordinator: The CC is not aware of subtask restores, it always 
restores all subtasks and relies on the fact that assigning state to a running 
execution attempt has no effect.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #14204: [FLINK-20325][build] Move docs_404_check to CI stage

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14204:
URL: https://github.com/apache/flink/pull/14204#issuecomment-733048767


   
   ## CI report:
   
   * 6550d3e1b01af94d4f652f993834b75272da1020 UNKNOWN
   * 38ed086f80ee2a94613a716d090574beed92aef4 UNKNOWN
   * 6226047b9ce026f6431f9b8f7eed7ab49b373170 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10268)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Assigned] (FLINK-20389) UnalignedCheckpointITCase failure caused by NullPointerException

2020-11-27 Thread Arvid Heise (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20389?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arvid Heise reassigned FLINK-20389:
---

Assignee: Matthias  (was: Arvid Heise)

> UnalignedCheckpointITCase failure caused by NullPointerException
> 
>
> Key: FLINK-20389
> URL: https://issues.apache.org/jira/browse/FLINK-20389
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing
>Affects Versions: 1.12.0, 1.13.0
>Reporter: Matthias
>Assignee: Matthias
>Priority: Critical
>  Labels: pull-request-available, test-stability
> Fix For: 1.12.0
>
>
> [Build|https://dev.azure.com/mapohl/flink/_build/results?buildId=118=results]
>  failed due to {{UnalignedCheckpointITCase}} caused by a 
> {{NullPointerException}}:
> {code:java}
> Test execute[Parallel cogroup, p = 
> 10](org.apache.flink.test.checkpointing.UnalignedCheckpointITCase) failed 
> with:
> org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
>   at 
> org.apache.flink.runtime.jobmaster.JobResult.toJobExecutionResult(JobResult.java:147)
>   at 
> org.apache.flink.runtime.minicluster.MiniClusterJobClient.lambda$getJobExecutionResult$2(MiniClusterJobClient.java:119)
>   at 
> java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java:616)
>   at 
> java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:591)
>   at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
>   at 
> java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1975)
>   at 
> org.apache.flink.runtime.rpc.akka.AkkaInvocationHandler.lambda$invokeRpc$0(AkkaInvocationHandler.java:229)
>   at 
> java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:774)
>   at 
> java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:750)
>   at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
>   at 
> java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1975)
>   at 
> org.apache.flink.runtime.concurrent.FutureUtils$1.onComplete(FutureUtils.java:996)
>   at akka.dispatch.OnComplete.internal(Future.scala:264)
>   at akka.dispatch.OnComplete.internal(Future.scala:261)
>   at akka.dispatch.japi$CallbackBridge.apply(Future.scala:191)
>   at akka.dispatch.japi$CallbackBridge.apply(Future.scala:188)
>   at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:36)
>   at 
> org.apache.flink.runtime.concurrent.Executors$DirectExecutionContext.execute(Executors.java:74)
>   at 
> scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:44)
>   at 
> scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:252)
>   at akka.pattern.PromiseActorRef.$bang(AskSupport.scala:572)
>   at 
> akka.pattern.PipeToSupport$PipeableFuture$$anonfun$pipeTo$1.applyOrElse(PipeToSupport.scala:22)
>   at 
> akka.pattern.PipeToSupport$PipeableFuture$$anonfun$pipeTo$1.applyOrElse(PipeToSupport.scala:21)
>   at scala.concurrent.Future$$anonfun$andThen$1.apply(Future.scala:436)
>   at scala.concurrent.Future$$anonfun$andThen$1.apply(Future.scala:435)
>   at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:36)
>   at 
> akka.dispatch.BatchingExecutor$AbstractBatch.processBatch(BatchingExecutor.scala:55)
>   at 
> akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run$1.apply$mcV$sp(BatchingExecutor.scala:91)
>   at 
> akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run$1.apply(BatchingExecutor.scala:91)
>   at 
> akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run$1.apply(BatchingExecutor.scala:91)
>   at 
> scala.concurrent.BlockContext$.withBlockContext(BlockContext.scala:72)
>   at 
> akka.dispatch.BatchingExecutor$BlockableBatch.run(BatchingExecutor.scala:90)
>   at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:40)
>   at 
> akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(ForkJoinExecutorConfigurator.scala:44)
>   at akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
>   at 
> akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
>   at akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
>   at 
> akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> Caused by: org.apache.flink.runtime.JobException: Recovery is suppressed by 
> FixedDelayRestartBackoffTimeStrategy(maxNumberRestartAttempts=5, 
> backoffTimeMS=100)
>   at 
> 

[jira] [Commented] (FLINK-20389) UnalignedCheckpointITCase failure caused by NullPointerException

2020-11-27 Thread Arvid Heise (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-20389?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17239815#comment-17239815
 ] 

Arvid Heise commented on FLINK-20389:
-

Improved UCITCase to not fail with NPE on unexpected exeception, for the actual 
exception, [~mapohl] is investigating.

Merged improvement into master as f5889bdc0f63ae97859743dc73f5c675dfa7bd2f and 
into 1.12 as 4744fb9de0.

> UnalignedCheckpointITCase failure caused by NullPointerException
> 
>
> Key: FLINK-20389
> URL: https://issues.apache.org/jira/browse/FLINK-20389
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing
>Affects Versions: 1.12.0, 1.13.0
>Reporter: Matthias
>Assignee: Arvid Heise
>Priority: Critical
>  Labels: pull-request-available, test-stability
> Fix For: 1.12.0
>
>
> [Build|https://dev.azure.com/mapohl/flink/_build/results?buildId=118=results]
>  failed due to {{UnalignedCheckpointITCase}} caused by a 
> {{NullPointerException}}:
> {code:java}
> Test execute[Parallel cogroup, p = 
> 10](org.apache.flink.test.checkpointing.UnalignedCheckpointITCase) failed 
> with:
> org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
>   at 
> org.apache.flink.runtime.jobmaster.JobResult.toJobExecutionResult(JobResult.java:147)
>   at 
> org.apache.flink.runtime.minicluster.MiniClusterJobClient.lambda$getJobExecutionResult$2(MiniClusterJobClient.java:119)
>   at 
> java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java:616)
>   at 
> java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:591)
>   at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
>   at 
> java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1975)
>   at 
> org.apache.flink.runtime.rpc.akka.AkkaInvocationHandler.lambda$invokeRpc$0(AkkaInvocationHandler.java:229)
>   at 
> java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:774)
>   at 
> java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:750)
>   at 
> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
>   at 
> java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1975)
>   at 
> org.apache.flink.runtime.concurrent.FutureUtils$1.onComplete(FutureUtils.java:996)
>   at akka.dispatch.OnComplete.internal(Future.scala:264)
>   at akka.dispatch.OnComplete.internal(Future.scala:261)
>   at akka.dispatch.japi$CallbackBridge.apply(Future.scala:191)
>   at akka.dispatch.japi$CallbackBridge.apply(Future.scala:188)
>   at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:36)
>   at 
> org.apache.flink.runtime.concurrent.Executors$DirectExecutionContext.execute(Executors.java:74)
>   at 
> scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:44)
>   at 
> scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:252)
>   at akka.pattern.PromiseActorRef.$bang(AskSupport.scala:572)
>   at 
> akka.pattern.PipeToSupport$PipeableFuture$$anonfun$pipeTo$1.applyOrElse(PipeToSupport.scala:22)
>   at 
> akka.pattern.PipeToSupport$PipeableFuture$$anonfun$pipeTo$1.applyOrElse(PipeToSupport.scala:21)
>   at scala.concurrent.Future$$anonfun$andThen$1.apply(Future.scala:436)
>   at scala.concurrent.Future$$anonfun$andThen$1.apply(Future.scala:435)
>   at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:36)
>   at 
> akka.dispatch.BatchingExecutor$AbstractBatch.processBatch(BatchingExecutor.scala:55)
>   at 
> akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run$1.apply$mcV$sp(BatchingExecutor.scala:91)
>   at 
> akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run$1.apply(BatchingExecutor.scala:91)
>   at 
> akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run$1.apply(BatchingExecutor.scala:91)
>   at 
> scala.concurrent.BlockContext$.withBlockContext(BlockContext.scala:72)
>   at 
> akka.dispatch.BatchingExecutor$BlockableBatch.run(BatchingExecutor.scala:90)
>   at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:40)
>   at 
> akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(ForkJoinExecutorConfigurator.scala:44)
>   at akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
>   at 
> akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
>   at akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
>   at 
> akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
> Caused by: org.apache.flink.runtime.JobException: Recovery is suppressed by 

[GitHub] [flink] AHeise merged pull request #14250: [FLINK-20389][tests] Fix UnalignedCheckpointITCase to work with unassigned splits.

2020-11-27 Thread GitBox


AHeise merged pull request #14250:
URL: https://github.com/apache/flink/pull/14250


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14254: [FLINK-20357][docs] Split HA documentation up into a general overview and the specific implementations

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14254:
URL: https://github.com/apache/flink/pull/14254#issuecomment-734942786


   
   ## CI report:
   
   * 9f8f123a323372be3edd985c83d8af11528dec16 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10269)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14238: [FLINK-20347] Rework YARN documentation page

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14238:
URL: https://github.com/apache/flink/pull/14238#issuecomment-734336545


   
   ## CI report:
   
   * 7806bb7e0dbca1263ddffb771dec367807f0590a Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10261)
 
   * 5b9ca356af610ffa7faf50a88372a5e22f7c7185 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10273)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14204: [FLINK-20325][build] Move docs_404_check to CI stage

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14204:
URL: https://github.com/apache/flink/pull/14204#issuecomment-733048767


   
   ## CI report:
   
   * 6550d3e1b01af94d4f652f993834b75272da1020 UNKNOWN
   * 38ed086f80ee2a94613a716d090574beed92aef4 UNKNOWN
   * b4d5fbae5b72a01ab9e94bc000373d6797339f9a Azure: 
[CANCELED](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10266)
 
   * 6226047b9ce026f6431f9b8f7eed7ab49b373170 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10268)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14199: [FLINK-20118][file connector] Extend FileSourceTextLinesITCase with JM/TM failover tests

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14199:
URL: https://github.com/apache/flink/pull/14199#issuecomment-732845197


   
   ## CI report:
   
   * 6ffe3e161a8ec42b7ce5da3c46b57587d98a1f90 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10267)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14253: [BP-1.11][FLINK-20382][runtime] Fail hard when JobMaster cannot start scheduling of a job

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14253:
URL: https://github.com/apache/flink/pull/14253#issuecomment-734818578


   
   ## CI report:
   
   * f7bda2acaafa5f56911c79b02929157ab89b63aa Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10259)
 
   * 6389f5ba95deedd53497b087df8da73d49d12929 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10272)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14251: [FLINK-20382][runtime] Fail hard when JobMaster cannot start scheduling of a job

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14251:
URL: https://github.com/apache/flink/pull/14251#issuecomment-734807691


   
   ## CI report:
   
   * aa9069b35f492d5b4b982476db0205c1b0fd1fd7 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10256)
 
   * 69beddd215cef8366db8a9c9e71b0dad0469b22d Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10270)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14252: [BP-1.12][FLINK-20382][runtime] Fail hard when JobMaster cannot start scheduling of a job

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14252:
URL: https://github.com/apache/flink/pull/14252#issuecomment-734817896


   
   ## CI report:
   
   * f246a660e140fd3998c00363624205be5fb9801e Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10258)
 
   * 9bff590490e977520cf92133855bd8a889f31fb4 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10271)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14238: [FLINK-20347] Rework YARN documentation page

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14238:
URL: https://github.com/apache/flink/pull/14238#issuecomment-734336545


   
   ## CI report:
   
   * 7806bb7e0dbca1263ddffb771dec367807f0590a Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10261)
 
   * 5b9ca356af610ffa7faf50a88372a5e22f7c7185 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14254: [FLINK-20357][docs] Split HA documentation up into a general overview and the specific implementations

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14254:
URL: https://github.com/apache/flink/pull/14254#issuecomment-734942786


   
   ## CI report:
   
   * 9f8f123a323372be3edd985c83d8af11528dec16 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10269)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14253: [BP-1.11][FLINK-20382][runtime] Fail hard when JobMaster cannot start scheduling of a job

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14253:
URL: https://github.com/apache/flink/pull/14253#issuecomment-734818578


   
   ## CI report:
   
   * f7bda2acaafa5f56911c79b02929157ab89b63aa Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10259)
 
   * 6389f5ba95deedd53497b087df8da73d49d12929 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14252: [BP-1.12][FLINK-20382][runtime] Fail hard when JobMaster cannot start scheduling of a job

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14252:
URL: https://github.com/apache/flink/pull/14252#issuecomment-734817896


   
   ## CI report:
   
   * f246a660e140fd3998c00363624205be5fb9801e Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10258)
 
   * 9bff590490e977520cf92133855bd8a889f31fb4 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14251: [FLINK-20382][runtime] Fail hard when JobMaster cannot start scheduling of a job

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14251:
URL: https://github.com/apache/flink/pull/14251#issuecomment-734807691


   
   ## CI report:
   
   * aa9069b35f492d5b4b982476db0205c1b0fd1fd7 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10256)
 
   * 69beddd215cef8366db8a9c9e71b0dad0469b22d UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] rmetzger commented on pull request #14238: [FLINK-20347] Rework YARN documentation page

2020-11-27 Thread GitBox


rmetzger commented on pull request #14238:
URL: https://github.com/apache/flink/pull/14238#issuecomment-734947167


   Thanks a lot for the through review. I've addressed all comments.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] rmetzger commented on a change in pull request #14238: [FLINK-20347] Rework YARN documentation page

2020-11-27 Thread GitBox


rmetzger commented on a change in pull request #14238:
URL: https://github.com/apache/flink/pull/14238#discussion_r531726061



##
File path: docs/deployment/resource-providers/yarn.md
##
@@ -0,0 +1,200 @@
+---
+title:  "Apache Hadoop YARN"
+nav-title: YARN
+nav-parent_id: resource_providers
+nav-pos: 4
+---
+
+
+* This will be replaced by the TOC
+{:toc}
+
+## Getting Started
+
+This *Getting Started* section guides you through setting up a fully 
functional Flink Cluster on YARN.
+
+### Introduction
+
+[Apache Hadoop 
YARN](https://hadoop.apache.org/docs/current/hadoop-yarn/hadoop-yarn-site/YARN.html)
 is a resource provider popular with many data processing frameworks.
+Flink services are submitted to YARN's ResourceManager, which spawns 
containers on machines managed by YARN NodeManagers. Flink deploys its 
JobManager and TaskManager instances into such containers.
+
+Flink can dynamically allocate and de-allocate TaskManager resources depending 
on the number of processing slots required by the job(s) running on the 
JobManager.
+
+### Preparation
+
+This *Getting Started* section assumes a functional YARN environment, starting 
from version 2.4.1. YARN environments are provided most conveniently through 
services such as Amazon EMR, Google Cloud DataProc or products like Cloudera. 
[Manually setting up a YARN environment 
locally](https://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/SingleCluster.html)
 or [on a 
cluster](https://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/ClusterSetup.html)
 is not recommended for following through this *Getting Started* tutorial. 
+
+
+- Make sure your YARN cluster is ready for accepting Flink applications by 
running `yarn top`. It should show no error messages.
+- Download a recent Flink distribution from the [download page]({{ 
site.download_url }}) and unpack it.
+- **Important** Make sure that the `HADOOP_CLASSPATH` environment variable is 
set up (it can be checked by running `echo $HADOOP_CLASSPATH`). If not, set it 
up using 
+
+{% highlight bash %}
+export HADOOP_CLASSPATH=`hadoop classpath`
+{% endhighlight %}
+
+
+### Starting a Flink Session on YARN
+
+Once you've made sure that the `HADOOP_CLASSPATH` environment variable is set, 
we can launch a Flink on YARN session, and submit an example job:
+
+{% highlight bash %}
+
+# we assume to be in the root directory of the unzipped Flink distribution
+
+# (0) export HADOOP_CLASSPATH
+export HADOOP_CLASSPATH=`hadoop classpath`
+
+# (1) Start YARN Session
+./bin/yarn-session.sh --detached
+
+# (2) You can now access the Flink Web Interface through the URL printed in 
the last lines of the command output, or through the YARN ResourceManager web 
UI.
+
+# (3) Submit example job
+./bin/flink run ./examples/streaming/TopSpeedWindowing.jar
+
+# (4) Stop YARN session (replace the application id based on the output of the 
yarn-session.sh command)
+echo "stop" | ./bin/yarn-session.sh -yid application_X_XXX
+{% endhighlight %}
+
+Congratulations! You have successfully run a Flink application by deploying 
Flink on YARN.
+
+
+## Deployment Modes Supported by Flink on YARN
+
+For production use, we recommend deploying Flink Applications in the [Per-job 
or Application Mode]({% link deployment/index.md %}#deployment-modes), as these 
modes provide a better isolation for the Applications.
+
+### Application Mode
+
+Application Mode will launch a Flink cluster on YARN, where the main() method 
of the application jar gets executed on the JobManager in YARN.
+The cluster will shut down as soon as the application has finished. You can 
manually stop the cluster using `yarn application -kill ` or by 
cancelling the Flink job.
+
+{% highlight bash %}
+./bin/flink run-application -t yarn-application 
./examples/streaming/TopSpeedWindowing.jar
+{% endhighlight %}
+
+To unlock the full potential of the application mode, consider using it with 
the `yarn.provided.lib.dirs` configuration option
+and pre-upload your application jar to a location accessible by all nodes in 
your cluster. In this case, the 
+command could look like: 
+
+{% highlight bash %}
+./bin/flink run-application -t yarn-application \
+   -Dyarn.provided.lib.dirs="hdfs://myhdfs/my-remote-flink-dist-dir" \
+   hdfs://myhdfs/jars/my-application.jar
+{% endhighlight %}
+
+The above will allow the job submission to be extra lightweight as the needed 
Flink jars and the application jar
+are  going to be picked up by the specified remote locations rather than be 
shipped to the cluster by the 
+client.
+
+### Per-Job Cluster Mode
+
+The Per-job Cluster mode will launch a Flink cluster on YARN, then run the 
provided application jar locally and finally submit the JobGraph to the 
JobManager on YARN. If you pass the `--detached` argument, the client will stop 
once the submission is accepted.
+
+The YARN cluster will stop once the job has stopped.
+
+{% highlight bash %}
+./bin/flink run -m 

[jira] [Commented] (FLINK-20390) Programmatic access to the back-pressure

2020-11-27 Thread Steven Zhen Wu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-20390?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17239806#comment-17239806
 ] 

Steven Zhen Wu commented on FLINK-20390:


not sure what is the input/source. If it is Kafka, would Kafka consumer lag 
(either in the number of msgs or in the wall clock time) be a good trigger. if 
it is lagging too much, it means that the job can't keep up with the input 
load. Then start to drop/sample msgs.

> Programmatic access to the back-pressure
> 
>
> Key: FLINK-20390
> URL: https://issues.apache.org/jira/browse/FLINK-20390
> Project: Flink
>  Issue Type: New Feature
>  Components: API / Core
>Reporter: Gaël Renoux
>Priority: Major
>
> It would be useful to access the back-pressure monitoring from within 
> functions.
> Here is our use case: we have a real-time Flink job, which takes decisions 
> based on input data. Sometimes, we have traffic spikes on the input and the 
> decisions process cannot processe records fast enough. Back-pressure starts 
> mounting, all the way back to the Source. What we want to do is to start 
> dropping records in this case, because it's better to make decisions based on 
> just a sample of the data rather than accumulate too much lag.
> Right now, the only way is to have a filter with a hard-limit on the number 
> of records per-interval-of-time, and to drop records once we are over this 
> limit. However, this requires a lot of tuning to find out what the correct 
> limit is, especially since it may depend on the nature of the inputs (some 
> decisions take longer to make than others). It's also heavily dependent on 
> the buffers: the limit needs to be low enough that all records that pass the 
> limit can fit in the downstream buffers, or the back-pressure will will go 
> back past the filtering task and we're back to square one. Finally, it's not 
> very resilient to change: whenever we scale the infrastructure up, we need to 
> redo the whole tuning thing.
> With programmatic access to the back-pressure, we could simply start dropping 
> records based on its current level. No tuning, and adjusted to the actual 
> issue. For performance, I assume it would be better if it reused the existing 
> back-pressure monitoring mechanism, rather than looking directly into the 
> buffer. A sampling of the back-pressure should be enough, and if more 
> precision is needed you can simply change the existing back-pressure 
> configuration.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-20405) The LAG function in over window is not implemented correctly

2020-11-27 Thread Stephan Ewen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20405?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen updated FLINK-20405:
-
Summary: The LAG function in over window is not implemented correctly  
(was: The LAG function in over window is not implements correctly)

> The LAG function in over window is not implemented correctly
> 
>
> Key: FLINK-20405
> URL: https://issues.apache.org/jira/browse/FLINK-20405
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Runtime
>Affects Versions: 1.12.0
>Reporter: Leonard Xu
>Priority: Major
>
> For LAG(input, offset, default) function in over window, it always return 
> current row's input no matter how the offset is set.
> After see the codegen code of the function, I think the implementation is not 
> correct and need to correct.
> {code:java}
> // the offset and default value is never used
> public UnboundedOverAggregateHelper$24(java.lang.Object[] references) throws 
> Exception {constant$14 = ((int) 1);   
>  constant$14isNull = false;   
>  constant$15 = ((org.apache.flink.table.data.binary.BinaryStringData) 
> str$13);constant$15isNull = false;
> typeSerializer$19 = 
> (((org.apache.flink.table.runtime.typeutils.StringDataSerializer) 
> references[0]));  }
> public void accumulate(org.apache.flink.table.data.RowData accInput) throws 
> Exception {
> org.apache.flink.table.data.binary.BinaryStringData field$21;
> boolean isNull$21;
> org.apache.flink.table.data.binary.BinaryStringData field$22;
> isNull$21 = accInput.isNullAt(2);field$21 = 
> org.apache.flink.table.data.binary.BinaryStringData.EMPTY_UTF8;if 
> (!isNull$21) {  field$21 = 
> ((org.apache.flink.table.data.binary.BinaryStringData) 
> accInput.getString(2));}field$22 = field$21;  
>   if (!isNull$21) {  field$22 = 
> (org.apache.flink.table.data.binary.BinaryStringData) 
> (typeSerializer$19.copy(field$22));}  
>   if (agg0_leadlag != field$22) {  agg0_leadlag = 
> ((org.apache.flink.table.data.binary.BinaryStringData) 
> typeSerializer$19.copy(field$22));}   ;   
>  agg0_leadlagIsNull = isNull$21; }
> {code}
>  
> The question comes from user mail list[1]
> [1] 
> http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/FlinkSQL-kafka-gt-dedup-gt-kafka-td39335.html



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot commented on pull request #14254: [FLINK-20357][docs] Split HA documentation up into a general overview and the specific implementations

2020-11-27 Thread GitBox


flinkbot commented on pull request #14254:
URL: https://github.com/apache/flink/pull/14254#issuecomment-734942786


   
   ## CI report:
   
   * 9f8f123a323372be3edd985c83d8af11528dec16 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14249: [FLINK-20391] Set FORWARD_EDGES_PIPELINED for BATCH ExecutionMode

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14249:
URL: https://github.com/apache/flink/pull/14249#issuecomment-734789081


   
   ## CI report:
   
   * 2ce421de97821f5fa47a965362d3c08c4971fa92 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10260)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] tillrohrmann commented on a change in pull request #14251: [FLINK-20382][runtime] Fail hard when JobMaster cannot start scheduling of a job

2020-11-27 Thread GitBox


tillrohrmann commented on a change in pull request #14251:
URL: https://github.com/apache/flink/pull/14251#discussion_r531718698



##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/jobmaster/JobMaster.java
##
@@ -932,7 +932,7 @@ private void resetAndStartScheduler() throws Exception {
);
}
 
-   schedulerAssignedFuture.thenRun(this::startScheduling);
+   
FutureUtils.assertNoException(schedulerAssignedFuture.thenRun(this::startScheduling));

Review comment:
   The intention of this PR is to fix the problem of swallowed exceptions 
and not the job lifecycle management based on operator coordinator failures. I 
don't know whether restarting a job makes sense if the operator coordinator has 
failed to start. To me it looks as if the operator coordinator integration 
hasn't been completed yet. I would consider this as a follow up task.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] rmetzger commented on a change in pull request #14238: [FLINK-20347] Rework YARN documentation page

2020-11-27 Thread GitBox


rmetzger commented on a change in pull request #14238:
URL: https://github.com/apache/flink/pull/14238#discussion_r531718518



##
File path: docs/deployment/resource-providers/yarn.md
##
@@ -0,0 +1,200 @@
+---
+title:  "Apache Hadoop YARN"
+nav-title: YARN
+nav-parent_id: resource_providers
+nav-pos: 4
+---
+
+
+* This will be replaced by the TOC
+{:toc}
+
+## Getting Started
+
+This *Getting Started* section guides you through setting up a fully 
functional Flink Cluster on YARN.
+
+### Introduction
+
+[Apache Hadoop 
YARN](https://hadoop.apache.org/docs/current/hadoop-yarn/hadoop-yarn-site/YARN.html)
 is a resource provider popular with many data processing frameworks.
+Flink services are submitted to YARN's ResourceManager, which spawns 
containers on machines managed by YARN NodeManagers. Flink deploys its 
JobManager and TaskManager instances into such containers.
+
+Flink can dynamically allocate and de-allocate TaskManager resources depending 
on the number of processing slots required by the job(s) running on the 
JobManager.
+
+### Preparation
+
+This *Getting Started* section assumes a functional YARN environment, starting 
from version 2.4.1. YARN environments are provided most conveniently through 
services such as Amazon EMR, Google Cloud DataProc or products like Cloudera. 
[Manually setting up a YARN environment 
locally](https://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/SingleCluster.html)
 or [on a 
cluster](https://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/ClusterSetup.html)
 is not recommended for following through this *Getting Started* tutorial. 
+
+
+- Make sure your YARN cluster is ready for accepting Flink applications by 
running `yarn top`. It should show no error messages.
+- Download a recent Flink distribution from the [download page]({{ 
site.download_url }}) and unpack it.
+- **Important** Make sure that the `HADOOP_CLASSPATH` environment variable is 
set up (it can be checked by running `echo $HADOOP_CLASSPATH`). If not, set it 
up using 
+
+{% highlight bash %}
+export HADOOP_CLASSPATH=`hadoop classpath`
+{% endhighlight %}
+
+
+### Starting a Flink Session on YARN
+
+Once you've made sure that the `HADOOP_CLASSPATH` environment variable is set, 
we can launch a Flink on YARN session, and submit an example job:
+
+{% highlight bash %}
+
+# we assume to be in the root directory of the unzipped Flink distribution
+
+# (0) export HADOOP_CLASSPATH
+export HADOOP_CLASSPATH=`hadoop classpath`
+
+# (1) Start YARN Session
+./bin/yarn-session.sh --detached
+
+# (2) You can now access the Flink Web Interface through the URL printed in 
the last lines of the command output, or through the YARN ResourceManager web 
UI.
+
+# (3) Submit example job
+./bin/flink run ./examples/streaming/TopSpeedWindowing.jar
+
+# (4) Stop YARN session (replace the application id based on the output of the 
yarn-session.sh command)
+echo "stop" | ./bin/yarn-session.sh -yid application_X_XXX
+{% endhighlight %}
+
+Congratulations! You have successfully run a Flink application by deploying 
Flink on YARN.
+
+
+## Deployment Modes Supported by Flink on YARN
+
+For production use, we recommend deploying Flink Applications in the [Per-job 
or Application Mode]({% link deployment/index.md %}#deployment-modes), as these 
modes provide a better isolation for the Applications.
+
+### Application Mode
+
+Application Mode will launch a Flink cluster on YARN, where the main() method 
of the application jar gets executed on the JobManager in YARN.
+The cluster will shut down as soon as the application has finished. You can 
manually stop the cluster using `yarn application -kill ` or by 
cancelling the Flink job.
+
+{% highlight bash %}
+./bin/flink run-application -t yarn-application 
./examples/streaming/TopSpeedWindowing.jar
+{% endhighlight %}
+
+To unlock the full potential of the application mode, consider using it with 
the `yarn.provided.lib.dirs` configuration option
+and pre-upload your application jar to a location accessible by all nodes in 
your cluster. In this case, the 
+command could look like: 
+
+{% highlight bash %}
+./bin/flink run-application -t yarn-application \
+   -Dyarn.provided.lib.dirs="hdfs://myhdfs/my-remote-flink-dist-dir" \
+   hdfs://myhdfs/jars/my-application.jar
+{% endhighlight %}
+
+The above will allow the job submission to be extra lightweight as the needed 
Flink jars and the application jar
+are  going to be picked up by the specified remote locations rather than be 
shipped to the cluster by the 
+client.
+
+### Per-Job Cluster Mode
+
+The Per-job Cluster mode will launch a Flink cluster on YARN, then run the 
provided application jar locally and finally submit the JobGraph to the 
JobManager on YARN. If you pass the `--detached` argument, the client will stop 
once the submission is accepted.
+
+The YARN cluster will stop once the job has stopped.
+
+{% highlight bash %}
+./bin/flink run -m 

[GitHub] [flink] tillrohrmann commented on a change in pull request #14251: [FLINK-20382][runtime] Fail hard when JobMaster cannot start scheduling of a job

2020-11-27 Thread GitBox


tillrohrmann commented on a change in pull request #14251:
URL: https://github.com/apache/flink/pull/14251#discussion_r531717894



##
File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/util/FatalExitExceptionHandler.java
##
@@ -32,6 +32,7 @@
private static final Logger LOG = 
LoggerFactory.getLogger(FatalExitExceptionHandler.class);
 
public static final FatalExitExceptionHandler INSTANCE = new 
FatalExitExceptionHandler();
+   public static final int EXIT_CODE = -17;

Review comment:
   I would see this an orthogonal change.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #14254: [FLINK-20357][docs] Split HA documentation up into a general overview and the specific implementations

2020-11-27 Thread GitBox


flinkbot commented on pull request #14254:
URL: https://github.com/apache/flink/pull/14254#issuecomment-734938232


   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit 9f8f123a323372be3edd985c83d8af11528dec16 (Fri Nov 27 
17:55:04 UTC 2020)
   
✅no warnings
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-20357) Rework HA documentation page

2020-11-27 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20357?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-20357:
---
Labels: pull-request-available  (was: )

> Rework HA documentation page
> 
>
> Key: FLINK-20357
> URL: https://issues.apache.org/jira/browse/FLINK-20357
> Project: Flink
>  Issue Type: Sub-task
>  Components: Documentation, Runtime / Coordination
>Affects Versions: 1.12.0
>Reporter: Till Rohrmann
>Assignee: Till Rohrmann
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.12.0
>
>
> We need to rework the HA documentation page. The first step is to split the 
> existing documentation into general concepts as an overview page and HA 
> service implementation specific sub pages. For the implementation specific 
> sub pages we need to add Zookeeper and the K8s HA services.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] tillrohrmann opened a new pull request #14254: [FLINK-20357][docs] Split HA documentation up into a general overview and the specific implementations

2020-11-27 Thread GitBox


tillrohrmann opened a new pull request #14254:
URL: https://github.com/apache/flink/pull/14254


   This commit splits the HA documentation up into a general overview and the 
specific implementations:
   * ZooKeeper HA services
   * Kubernetes HA services
   
   Moreover, this commit moves resource-provider specific documentation to the 
respective resource-provider
   documentation. This is done in order to not lose this information and it 
should be properly incorporated once the resource-provider documentation is 
updated.
   
   cc @rmetzger, @XComp, @wangyang0918 
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14247: [FLINK-20175][bp-1.11] Avro Confluent Registry SQL format does not support adding nullable columns

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14247:
URL: https://github.com/apache/flink/pull/14247#issuecomment-734770560


   
   ## CI report:
   
   * 70cf3a3d81f9d5499e5008c5c5eca93a4cdb0c7b Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10264)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] rmetzger commented on a change in pull request #14238: [FLINK-20347] Rework YARN documentation page

2020-11-27 Thread GitBox


rmetzger commented on a change in pull request #14238:
URL: https://github.com/apache/flink/pull/14238#discussion_r531709904



##
File path: docs/deployment/resource-providers/yarn.md
##
@@ -0,0 +1,200 @@
+---
+title:  "Apache Hadoop YARN"
+nav-title: YARN
+nav-parent_id: resource_providers
+nav-pos: 4
+---
+
+
+* This will be replaced by the TOC
+{:toc}
+
+## Getting Started
+
+This *Getting Started* section guides you through setting up a fully 
functional Flink Cluster on YARN.
+
+### Introduction
+
+[Apache Hadoop 
YARN](https://hadoop.apache.org/docs/current/hadoop-yarn/hadoop-yarn-site/YARN.html)
 is a resource provider popular with many data processing frameworks.
+Flink services are submitted to YARN's ResourceManager, which spawns 
containers on machines managed by YARN NodeManagers. Flink deploys its 
JobManager and TaskManager instances into such containers.
+
+Flink can dynamically allocate and de-allocate TaskManager resources depending 
on the number of processing slots required by the job(s) running on the 
JobManager.
+
+### Preparation
+
+This *Getting Started* section assumes a functional YARN environment, starting 
from version 2.4.1. YARN environments are provided most conveniently through 
services such as Amazon EMR, Google Cloud DataProc or products like Cloudera. 
[Manually setting up a YARN environment 
locally](https://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/SingleCluster.html)
 or [on a 
cluster](https://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/ClusterSetup.html)
 is not recommended for following through this *Getting Started* tutorial. 
+
+
+- Make sure your YARN cluster is ready for accepting Flink applications by 
running `yarn top`. It should show no error messages.
+- Download a recent Flink distribution from the [download page]({{ 
site.download_url }}) and unpack it.
+- **Important** Make sure that the `HADOOP_CLASSPATH` environment variable is 
set up (it can be checked by running `echo $HADOOP_CLASSPATH`). If not, set it 
up using 
+
+{% highlight bash %}
+export HADOOP_CLASSPATH=`hadoop classpath`
+{% endhighlight %}
+
+
+### Starting a Flink Session on YARN

Review comment:
   I think it only makes sense to mention `HADOOP_CONF_DIR`. Afaik 
`HADOOP_HOME` is from Hadoop 1.x times, and deprecated.
   
   I will mention  `HADOOP_CONF_DIR` in the reference section about the 
configuration, to not confuse new users in the beginners section.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14204: [FLINK-20325][build] Move docs_404_check to CI stage

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14204:
URL: https://github.com/apache/flink/pull/14204#issuecomment-733048767


   
   ## CI report:
   
   * 6550d3e1b01af94d4f652f993834b75272da1020 UNKNOWN
   * 38ed086f80ee2a94613a716d090574beed92aef4 UNKNOWN
   * 983518fa6cdff38b3264506a9ca54178699c9525 Azure: 
[CANCELED](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10257)
 
   * b4d5fbae5b72a01ab9e94bc000373d6797339f9a Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10266)
 
   * 6226047b9ce026f6431f9b8f7eed7ab49b373170 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10268)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] tillrohrmann commented on a change in pull request #14238: [FLINK-20347] Rework YARN documentation page

2020-11-27 Thread GitBox


tillrohrmann commented on a change in pull request #14238:
URL: https://github.com/apache/flink/pull/14238#discussion_r531702295



##
File path: docs/deployment/resource-providers/yarn.md
##
@@ -0,0 +1,200 @@
+---
+title:  "Apache Hadoop YARN"
+nav-title: YARN
+nav-parent_id: resource_providers
+nav-pos: 4
+---
+
+
+* This will be replaced by the TOC
+{:toc}
+
+## Getting Started
+
+This *Getting Started* section guides you through setting up a fully 
functional Flink Cluster on YARN.
+
+### Introduction
+
+[Apache Hadoop 
YARN](https://hadoop.apache.org/docs/current/hadoop-yarn/hadoop-yarn-site/YARN.html)
 is a resource provider popular with many data processing frameworks.
+Flink services are submitted to YARN's ResourceManager, which spawns 
containers on machines managed by YARN NodeManagers. Flink deploys its 
JobManager and TaskManager instances into such containers.
+
+Flink can dynamically allocate and de-allocate TaskManager resources depending 
on the number of processing slots required by the job(s) running on the 
JobManager.
+
+### Preparation
+
+This *Getting Started* section assumes a functional YARN environment, starting 
from version 2.4.1. YARN environments are provided most conveniently through 
services such as Amazon EMR, Google Cloud DataProc or products like Cloudera. 
[Manually setting up a YARN environment 
locally](https://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/SingleCluster.html)
 or [on a 
cluster](https://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/ClusterSetup.html)
 is not recommended for following through this *Getting Started* tutorial. 
+
+
+- Make sure your YARN cluster is ready for accepting Flink applications by 
running `yarn top`. It should show no error messages.
+- Download a recent Flink distribution from the [download page]({{ 
site.download_url }}) and unpack it.
+- **Important** Make sure that the `HADOOP_CLASSPATH` environment variable is 
set up (it can be checked by running `echo $HADOOP_CLASSPATH`). If not, set it 
up using 
+
+{% highlight bash %}
+export HADOOP_CLASSPATH=`hadoop classpath`
+{% endhighlight %}
+
+
+### Starting a Flink Session on YARN
+
+Once you've made sure that the `HADOOP_CLASSPATH` environment variable is set, 
we can launch a Flink on YARN session, and submit an example job:
+
+{% highlight bash %}
+
+# we assume to be in the root directory of the unzipped Flink distribution
+
+# (0) export HADOOP_CLASSPATH
+export HADOOP_CLASSPATH=`hadoop classpath`
+
+# (1) Start YARN Session
+./bin/yarn-session.sh --detached
+
+# (2) You can now access the Flink Web Interface through the URL printed in 
the last lines of the command output, or through the YARN ResourceManager web 
UI.
+
+# (3) Submit example job
+./bin/flink run ./examples/streaming/TopSpeedWindowing.jar
+
+# (4) Stop YARN session (replace the application id based on the output of the 
yarn-session.sh command)
+echo "stop" | ./bin/yarn-session.sh -yid application_X_XXX
+{% endhighlight %}
+
+Congratulations! You have successfully run a Flink application by deploying 
Flink on YARN.
+
+
+## Deployment Modes Supported by Flink on YARN
+
+For production use, we recommend deploying Flink Applications in the [Per-job 
or Application Mode]({% link deployment/index.md %}#deployment-modes), as these 
modes provide a better isolation for the Applications.
+
+### Application Mode
+
+Application Mode will launch a Flink cluster on YARN, where the main() method 
of the application jar gets executed on the JobManager in YARN.
+The cluster will shut down as soon as the application has finished. You can 
manually stop the cluster using `yarn application -kill ` or by 
cancelling the Flink job.
+
+{% highlight bash %}
+./bin/flink run-application -t yarn-application 
./examples/streaming/TopSpeedWindowing.jar
+{% endhighlight %}
+
+To unlock the full potential of the application mode, consider using it with 
the `yarn.provided.lib.dirs` configuration option
+and pre-upload your application jar to a location accessible by all nodes in 
your cluster. In this case, the 
+command could look like: 
+
+{% highlight bash %}
+./bin/flink run-application -t yarn-application \
+   -Dyarn.provided.lib.dirs="hdfs://myhdfs/my-remote-flink-dist-dir" \
+   hdfs://myhdfs/jars/my-application.jar
+{% endhighlight %}
+
+The above will allow the job submission to be extra lightweight as the needed 
Flink jars and the application jar
+are  going to be picked up by the specified remote locations rather than be 
shipped to the cluster by the 
+client.
+
+### Per-Job Cluster Mode
+
+The Per-job Cluster mode will launch a Flink cluster on YARN, then run the 
provided application jar locally and finally submit the JobGraph to the 
JobManager on YARN. If you pass the `--detached` argument, the client will stop 
once the submission is accepted.
+
+The YARN cluster will stop once the job has stopped.
+
+{% highlight bash %}
+./bin/flink run -m 

[GitHub] [flink] rmetzger commented on a change in pull request #14238: [FLINK-20347] Rework YARN documentation page

2020-11-27 Thread GitBox


rmetzger commented on a change in pull request #14238:
URL: https://github.com/apache/flink/pull/14238#discussion_r531698001



##
File path: docs/deployment/resource-providers/yarn.md
##
@@ -0,0 +1,200 @@
+---
+title:  "Apache Hadoop YARN"
+nav-title: YARN
+nav-parent_id: resource_providers
+nav-pos: 4
+---
+
+
+* This will be replaced by the TOC
+{:toc}
+
+## Getting Started
+
+This *Getting Started* section guides you through setting up a fully 
functional Flink Cluster on YARN.
+
+### Introduction
+
+[Apache Hadoop 
YARN](https://hadoop.apache.org/docs/current/hadoop-yarn/hadoop-yarn-site/YARN.html)
 is a resource provider popular with many data processing frameworks.
+Flink services are submitted to YARN's ResourceManager, which spawns 
containers on machines managed by YARN NodeManagers. Flink deploys its 
JobManager and TaskManager instances into such containers.
+
+Flink can dynamically allocate and de-allocate TaskManager resources depending 
on the number of processing slots required by the job(s) running on the 
JobManager.
+
+### Preparation
+
+This *Getting Started* section assumes a functional YARN environment, starting 
from version 2.4.1. YARN environments are provided most conveniently through 
services such as Amazon EMR, Google Cloud DataProc or products like Cloudera. 
[Manually setting up a YARN environment 
locally](https://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/SingleCluster.html)
 or [on a 
cluster](https://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/ClusterSetup.html)
 is not recommended for following through this *Getting Started* tutorial. 
+
+
+- Make sure your YARN cluster is ready for accepting Flink applications by 
running `yarn top`. It should show no error messages.
+- Download a recent Flink distribution from the [download page]({{ 
site.download_url }}) and unpack it.
+- **Important** Make sure that the `HADOOP_CLASSPATH` environment variable is 
set up (it can be checked by running `echo $HADOOP_CLASSPATH`). If not, set it 
up using 
+
+{% highlight bash %}
+export HADOOP_CLASSPATH=`hadoop classpath`
+{% endhighlight %}
+
+
+### Starting a Flink Session on YARN
+
+Once you've made sure that the `HADOOP_CLASSPATH` environment variable is set, 
we can launch a Flink on YARN session, and submit an example job:
+
+{% highlight bash %}
+
+# we assume to be in the root directory of the unzipped Flink distribution
+
+# (0) export HADOOP_CLASSPATH
+export HADOOP_CLASSPATH=`hadoop classpath`
+
+# (1) Start YARN Session
+./bin/yarn-session.sh --detached
+
+# (2) You can now access the Flink Web Interface through the URL printed in 
the last lines of the command output, or through the YARN ResourceManager web 
UI.
+
+# (3) Submit example job
+./bin/flink run ./examples/streaming/TopSpeedWindowing.jar
+
+# (4) Stop YARN session (replace the application id based on the output of the 
yarn-session.sh command)
+echo "stop" | ./bin/yarn-session.sh -yid application_X_XXX
+{% endhighlight %}
+
+Congratulations! You have successfully run a Flink application by deploying 
Flink on YARN.
+
+
+## Deployment Modes Supported by Flink on YARN
+
+For production use, we recommend deploying Flink Applications in the [Per-job 
or Application Mode]({% link deployment/index.md %}#deployment-modes), as these 
modes provide a better isolation for the Applications.
+
+### Application Mode
+
+Application Mode will launch a Flink cluster on YARN, where the main() method 
of the application jar gets executed on the JobManager in YARN.
+The cluster will shut down as soon as the application has finished. You can 
manually stop the cluster using `yarn application -kill ` or by 
cancelling the Flink job.
+
+{% highlight bash %}
+./bin/flink run-application -t yarn-application 
./examples/streaming/TopSpeedWindowing.jar
+{% endhighlight %}
+
+To unlock the full potential of the application mode, consider using it with 
the `yarn.provided.lib.dirs` configuration option
+and pre-upload your application jar to a location accessible by all nodes in 
your cluster. In this case, the 
+command could look like: 
+
+{% highlight bash %}
+./bin/flink run-application -t yarn-application \
+   -Dyarn.provided.lib.dirs="hdfs://myhdfs/my-remote-flink-dist-dir" \
+   hdfs://myhdfs/jars/my-application.jar
+{% endhighlight %}
+
+The above will allow the job submission to be extra lightweight as the needed 
Flink jars and the application jar
+are  going to be picked up by the specified remote locations rather than be 
shipped to the cluster by the 
+client.
+
+### Per-Job Cluster Mode
+
+The Per-job Cluster mode will launch a Flink cluster on YARN, then run the 
provided application jar locally and finally submit the JobGraph to the 
JobManager on YARN. If you pass the `--detached` argument, the client will stop 
once the submission is accepted.
+
+The YARN cluster will stop once the job has stopped.
+
+{% highlight bash %}
+./bin/flink run -m 

[jira] [Commented] (FLINK-20396) Replace "OperatorCoordinator.subtaskFailed()" with "subtaskRestored()"

2020-11-27 Thread Stephan Ewen (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-20396?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17239774#comment-17239774
 ] 

Stephan Ewen commented on FLINK-20396:
--

In the first version, I would add the {{subtaskReset()}} method in addition to 
the {{subtaskFailed()}} method because of two reasons:
  - {{subtaskFailed()}} can potentially do slightly faster cleanup (for example 
unregistering readers)
  - It is complex to communicate a failure cause to {{subtaskReset()}}. If we 
want to make 

We can decide to remove {{subtaskFailed()}} in the future, if we feel we don't 
need the slightly faster notification, or the failure reason.

If we want to retain the failure reason, but do not care about the slightly 
faster notification, we can consolidate the two into a single method 
{{subtaskReset(int subtask, long checkpointId, Throwable failureCause);}}.
The {{OperatorCoordinatorHolder}} can remember the exceptions per subtask 
between failure and restore to pass them to the coordinator.
Because that needs potentially noticeably more heap memory (retain many 
exceptions with stack traces) I would suggest to do that change "if needed" and 
not immediately.

> Replace "OperatorCoordinator.subtaskFailed()" with "subtaskRestored()"
> --
>
> Key: FLINK-20396
> URL: https://issues.apache.org/jira/browse/FLINK-20396
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.11.2
>Reporter: Stephan Ewen
>Assignee: Stephan Ewen
>Priority: Blocker
> Fix For: 1.12.0, 1.11.3
>
>
> There are no strong order guarantees between 
> {{OperatorCoordinator.subtaskFailed()}} and 
> {{OperatorCoordinator.notifyCheckpointComplete()}}.
> It can happen that a checkpoint completes after the notification for task 
> failure is sent:
>   - {{OperatorCoordinator.checkpoint()}}
>   - {{OperatorCoordinator.subtaskFailed()}}
>   - {{OperatorCoordinator.checkpointComplete()}}
> The subtask failure here does not know whether the previous checkpoint 
> completed or not. It cannot decide what state the subtask will be in after 
> recovery.
> There is no easy fix right now to strictly guarantee the order of the method 
> calls, so alternatively we need to provide the necessary information to 
> reason about the status of tasks.
> We should replace {{OperatorCoordinator.subtaskFailed(int subtask)}} with 
> {{OperatorCoordinator.subtaskRestored(int subtask, long checkpoint)}}. That 
> implementations get the explicit checkpoint ID for the subtask recovery, and 
> can align that with the IDs of checkpoints that were taken.
> It is still (in rare cases) possible that for a specific checkpoint C, 
> {{OperatorCoordinator.subtaskRestored(subtaskIndex, C)) comes before 
> {{OperatorCoordinator.checkpointComplete(C)}}.
> h3. Background
> The Checkpointing Procedure is partially asynchronous on the {{JobManager}} / 
> {{CheckpointCoordinator}}: After all subtasks acknowledged the checkpoint, 
> the finalization (writing out metadata and registering the checkpoint in 
> ZooKeeper) happens in an I/O thread, and the checkpoint completes after that.
> This sequence of events can happen:
>   - tasks acks checkpoint
>   - checkpoint fully acknowledged, finalization starts
>   - task fails
>   - task failure notification is dispatched
>   - checkpoint completes.
> For task failures and checkpoint completion, no order is defined.
> However, for task restore and checkpoint completion, the order is well 
> defined: When a task is restored, pending checkpoints are either canceled or 
> complete. None can be within finalization. That is currently guaranteed with 
> a lock in the {{CheckpointCoordinator}}.
> (An implication of that being that restores can be blocking operations in the 
> scheduler, which is not ideal from the perspective of making the scheduler 
> async/non-blocking, but it is currently essential for correctness).



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (FLINK-20396) Replace "OperatorCoordinator.subtaskFailed()" with "subtaskRestored()"

2020-11-27 Thread Stephan Ewen (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-20396?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17239774#comment-17239774
 ] 

Stephan Ewen edited comment on FLINK-20396 at 11/27/20, 4:43 PM:
-

In the first version, I would add the {{subtaskReset()}} method in addition to 
the {{subtaskFailed()}} method because of two reasons:
  - {{subtaskFailed()}} can potentially do slightly faster cleanup (for example 
unregistering readers)
  - It is complex to communicate a failure cause to {{subtaskReset()}}. If we 
want to make actions dependent on exception types, we need to handle that in 
the {{subtaskFailed()}} method.

We can decide to remove {{subtaskFailed()}} in the future, if we feel we don't 
need the slightly faster notification, or the failure reason.

If we want to retain the failure reason, but do not care about the slightly 
faster notification, we can consolidate the two into a single method 
{{subtaskReset(int subtask, long checkpointId, Throwable failureCause);}}.
The {{OperatorCoordinatorHolder}} can remember the exceptions per subtask 
between failure and restore to pass them to the coordinator.
Because that needs potentially noticeably more heap memory (retain many 
exceptions with stack traces) I would suggest to do that change "if needed" and 
not immediately.


was (Author: stephanewen):
In the first version, I would add the {{subtaskReset()}} method in addition to 
the {{subtaskFailed()}} method because of two reasons:
  - {{subtaskFailed()}} can potentially do slightly faster cleanup (for example 
unregistering readers)
  - It is complex to communicate a failure cause to {{subtaskReset()}}. If we 
want to make 

We can decide to remove {{subtaskFailed()}} in the future, if we feel we don't 
need the slightly faster notification, or the failure reason.

If we want to retain the failure reason, but do not care about the slightly 
faster notification, we can consolidate the two into a single method 
{{subtaskReset(int subtask, long checkpointId, Throwable failureCause);}}.
The {{OperatorCoordinatorHolder}} can remember the exceptions per subtask 
between failure and restore to pass them to the coordinator.
Because that needs potentially noticeably more heap memory (retain many 
exceptions with stack traces) I would suggest to do that change "if needed" and 
not immediately.

> Replace "OperatorCoordinator.subtaskFailed()" with "subtaskRestored()"
> --
>
> Key: FLINK-20396
> URL: https://issues.apache.org/jira/browse/FLINK-20396
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.11.2
>Reporter: Stephan Ewen
>Assignee: Stephan Ewen
>Priority: Blocker
> Fix For: 1.12.0, 1.11.3
>
>
> There are no strong order guarantees between 
> {{OperatorCoordinator.subtaskFailed()}} and 
> {{OperatorCoordinator.notifyCheckpointComplete()}}.
> It can happen that a checkpoint completes after the notification for task 
> failure is sent:
>   - {{OperatorCoordinator.checkpoint()}}
>   - {{OperatorCoordinator.subtaskFailed()}}
>   - {{OperatorCoordinator.checkpointComplete()}}
> The subtask failure here does not know whether the previous checkpoint 
> completed or not. It cannot decide what state the subtask will be in after 
> recovery.
> There is no easy fix right now to strictly guarantee the order of the method 
> calls, so alternatively we need to provide the necessary information to 
> reason about the status of tasks.
> We should replace {{OperatorCoordinator.subtaskFailed(int subtask)}} with 
> {{OperatorCoordinator.subtaskRestored(int subtask, long checkpoint)}}. That 
> implementations get the explicit checkpoint ID for the subtask recovery, and 
> can align that with the IDs of checkpoints that were taken.
> It is still (in rare cases) possible that for a specific checkpoint C, 
> {{OperatorCoordinator.subtaskRestored(subtaskIndex, C)) comes before 
> {{OperatorCoordinator.checkpointComplete(C)}}.
> h3. Background
> The Checkpointing Procedure is partially asynchronous on the {{JobManager}} / 
> {{CheckpointCoordinator}}: After all subtasks acknowledged the checkpoint, 
> the finalization (writing out metadata and registering the checkpoint in 
> ZooKeeper) happens in an I/O thread, and the checkpoint completes after that.
> This sequence of events can happen:
>   - tasks acks checkpoint
>   - checkpoint fully acknowledged, finalization starts
>   - task fails
>   - task failure notification is dispatched
>   - checkpoint completes.
> For task failures and checkpoint completion, no order is defined.
> However, for task restore and checkpoint completion, the order is well 
> defined: When a task is restored, pending checkpoints are either canceled or 
> complete. None can be within 

[GitHub] [flink] flinkbot edited a comment on pull request #14204: [FLINK-20325][build] Move docs_404_check to CI stage

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14204:
URL: https://github.com/apache/flink/pull/14204#issuecomment-733048767


   
   ## CI report:
   
   * 6550d3e1b01af94d4f652f993834b75272da1020 UNKNOWN
   * 38ed086f80ee2a94613a716d090574beed92aef4 UNKNOWN
   * d97f8149d6fcb9985cd75c4cd8f30b9cd10f14e5 Azure: 
[CANCELED](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10233)
 
   * 983518fa6cdff38b3264506a9ca54178699c9525 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10257)
 
   * b4d5fbae5b72a01ab9e94bc000373d6797339f9a Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10266)
 
   * 6226047b9ce026f6431f9b8f7eed7ab49b373170 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10268)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14234: Make FileSink Support migrating from StreamingFileSink

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14234:
URL: https://github.com/apache/flink/pull/14234#issuecomment-734232999


   
   ## CI report:
   
   * c89575f22a01309ffb329e1d0ecdc07b3c99fc91 UNKNOWN
   * 3f569564dae3e5ee5a9b76ad86dc3acf9685d9df Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10255)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14204: [FLINK-20325][build] Move docs_404_check to CI stage

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14204:
URL: https://github.com/apache/flink/pull/14204#issuecomment-733048767


   
   ## CI report:
   
   * 6550d3e1b01af94d4f652f993834b75272da1020 UNKNOWN
   * 38ed086f80ee2a94613a716d090574beed92aef4 UNKNOWN
   * d97f8149d6fcb9985cd75c4cd8f30b9cd10f14e5 Azure: 
[CANCELED](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10233)
 
   * 983518fa6cdff38b3264506a9ca54178699c9525 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10257)
 
   * b4d5fbae5b72a01ab9e94bc000373d6797339f9a Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10266)
 
   * 6226047b9ce026f6431f9b8f7eed7ab49b373170 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14157: [FLINK-19969] CLI print run-application help msg

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14157:
URL: https://github.com/apache/flink/pull/14157#issuecomment-731272625


   
   ## CI report:
   
   * f8b909654df226298fd5f5c21ece16ebb55d5ea1 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10254)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Created] (FLINK-20405) The LAG function in over window is not implements correctly

2020-11-27 Thread Leonard Xu (Jira)
Leonard Xu created FLINK-20405:
--

 Summary: The LAG function in over window is not implements 
correctly
 Key: FLINK-20405
 URL: https://issues.apache.org/jira/browse/FLINK-20405
 Project: Flink
  Issue Type: Bug
  Components: Table SQL / Runtime
Affects Versions: 1.12.0
Reporter: Leonard Xu


For LAG(input, offset, default) function in over window, it always return 
current row's input no matter how the offset is set.

After see the codegen code of the function, I think the implementation is not 
correct and need to correct.
{code:java}

// the offset and default value is never used
public UnboundedOverAggregateHelper$24(java.lang.Object[] references) throws 
Exception {constant$14 = ((int) 1); 
   constant$14isNull = false;
constant$15 = ((org.apache.flink.table.data.binary.BinaryStringData) str$13);   
 constant$15isNull = false;typeSerializer$19 = 
(((org.apache.flink.table.runtime.typeutils.StringDataSerializer) 
references[0]));  }

public void accumulate(org.apache.flink.table.data.RowData accInput) throws 
Exception {
org.apache.flink.table.data.binary.BinaryStringData field$21;
boolean isNull$21;
org.apache.flink.table.data.binary.BinaryStringData field$22;
isNull$21 = accInput.isNullAt(2);field$21 = 
org.apache.flink.table.data.binary.BinaryStringData.EMPTY_UTF8;if 
(!isNull$21) {  field$21 = 
((org.apache.flink.table.data.binary.BinaryStringData) accInput.getString(2));  
  }field$22 = field$21;if (!isNull$21) {
  field$22 = (org.apache.flink.table.data.binary.BinaryStringData) 
(typeSerializer$19.copy(field$22));}
if (agg0_leadlag != field$22) {  agg0_leadlag = 
((org.apache.flink.table.data.binary.BinaryStringData) 
typeSerializer$19.copy(field$22));}   ;
agg0_leadlagIsNull = isNull$21; }
{code}
 

The question comes from user mail list[1]

[1] 
http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/FlinkSQL-kafka-gt-dedup-gt-kafka-td39335.html



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #14249: [FLINK-20391] Set FORWARD_EDGES_PIPELINED for BATCH ExecutionMode

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14249:
URL: https://github.com/apache/flink/pull/14249#issuecomment-734789081


   
   ## CI report:
   
   * 8092920b062657c9f8fed976a7b1792411a92a59 Azure: 
[CANCELED](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10252)
 
   * 2ce421de97821f5fa47a965362d3c08c4971fa92 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10260)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14250: [FLINK-20389][tests] Fix UnalignedCheckpointITCase to work with unassigned splits.

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14250:
URL: https://github.com/apache/flink/pull/14250#issuecomment-734789670


   
   ## CI report:
   
   * 12672b400e4237bb5ec5ada14fae4c2f6596b03d Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10253)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14199: [FLINK-20118][file connector] Extend FileSourceTextLinesITCase with JM/TM failover tests

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14199:
URL: https://github.com/apache/flink/pull/14199#issuecomment-732845197


   
   ## CI report:
   
   * 8139d3ee5e5fd595279a4a4c6e7849d1a9dfc6b5 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10049)
 
   * 6ffe3e161a8ec42b7ce5da3c46b57587d98a1f90 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10267)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] Shawn-Hx commented on pull request #14235: [FLINK-20298][docs] Replace usage of {{ site.baseurl }} in documentation

2020-11-27 Thread GitBox


Shawn-Hx commented on pull request #14235:
URL: https://github.com/apache/flink/pull/14235#issuecomment-734892872


   Thanks for the review @dawidwys 
   I need some time to be familiar with the grammar of Jekyll.
   I wonder if I could open another JIRA ticket to solve the remaining problems 
because this PR has changed a large number of files, and if others change these 
files there might be lots of conflicts to resolve.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] rmetzger commented on a change in pull request #14238: [FLINK-20347] Rework YARN documentation page

2020-11-27 Thread GitBox


rmetzger commented on a change in pull request #14238:
URL: https://github.com/apache/flink/pull/14238#discussion_r531669932



##
File path: docs/deployment/resource-providers/yarn.md
##
@@ -0,0 +1,200 @@
+---
+title:  "Apache Hadoop YARN"
+nav-title: YARN
+nav-parent_id: resource_providers
+nav-pos: 4
+---
+
+
+* This will be replaced by the TOC
+{:toc}
+
+## Getting Started
+
+This *Getting Started* section guides you through setting up a fully 
functional Flink Cluster on YARN.
+
+### Introduction
+
+[Apache Hadoop 
YARN](https://hadoop.apache.org/docs/current/hadoop-yarn/hadoop-yarn-site/YARN.html)
 is a resource provider popular with many data processing frameworks.
+Flink services are submitted to YARN's ResourceManager, which spawns 
containers on machines managed by YARN NodeManagers. Flink deploys its 
JobManager and TaskManager instances into such containers.
+
+Flink can dynamically allocate and de-allocate TaskManager resources depending 
on the number of processing slots required by the job(s) running on the 
JobManager.
+
+### Preparation
+
+This *Getting Started* section assumes a functional YARN environment, starting 
from version 2.4.1. YARN environments are provided most conveniently through 
services such as Amazon EMR, Google Cloud DataProc or products like Cloudera. 
[Manually setting up a YARN environment 
locally](https://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/SingleCluster.html)
 or [on a 
cluster](https://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/ClusterSetup.html)
 is not recommended for following through this *Getting Started* tutorial. 
+
+
+- Make sure your YARN cluster is ready for accepting Flink applications by 
running `yarn top`. It should show no error messages.
+- Download a recent Flink distribution from the [download page]({{ 
site.download_url }}) and unpack it.
+- **Important** Make sure that the `HADOOP_CLASSPATH` environment variable is 
set up (it can be checked by running `echo $HADOOP_CLASSPATH`). If not, set it 
up using 
+
+{% highlight bash %}
+export HADOOP_CLASSPATH=`hadoop classpath`
+{% endhighlight %}
+
+
+### Starting a Flink Session on YARN
+
+Once you've made sure that the `HADOOP_CLASSPATH` environment variable is set, 
we can launch a Flink on YARN session, and submit an example job:
+
+{% highlight bash %}
+
+# we assume to be in the root directory of the unzipped Flink distribution
+
+# (0) export HADOOP_CLASSPATH
+export HADOOP_CLASSPATH=`hadoop classpath`
+
+# (1) Start YARN Session
+./bin/yarn-session.sh --detached
+
+# (2) You can now access the Flink Web Interface through the URL printed in 
the last lines of the command output, or through the YARN ResourceManager web 
UI.
+
+# (3) Submit example job
+./bin/flink run ./examples/streaming/TopSpeedWindowing.jar
+
+# (4) Stop YARN session (replace the application id based on the output of the 
yarn-session.sh command)
+echo "stop" | ./bin/yarn-session.sh -yid application_X_XXX
+{% endhighlight %}
+
+Congratulations! You have successfully run a Flink application by deploying 
Flink on YARN.
+
+
+## Deployment Modes Supported by Flink on YARN
+
+For production use, we recommend deploying Flink Applications in the [Per-job 
or Application Mode]({% link deployment/index.md %}#deployment-modes), as these 
modes provide a better isolation for the Applications.
+
+### Application Mode
+
+Application Mode will launch a Flink cluster on YARN, where the main() method 
of the application jar gets executed on the JobManager in YARN.
+The cluster will shut down as soon as the application has finished. You can 
manually stop the cluster using `yarn application -kill ` or by 
cancelling the Flink job.
+
+{% highlight bash %}
+./bin/flink run-application -t yarn-application 
./examples/streaming/TopSpeedWindowing.jar
+{% endhighlight %}
+
+To unlock the full potential of the application mode, consider using it with 
the `yarn.provided.lib.dirs` configuration option
+and pre-upload your application jar to a location accessible by all nodes in 
your cluster. In this case, the 
+command could look like: 
+
+{% highlight bash %}
+./bin/flink run-application -t yarn-application \
+   -Dyarn.provided.lib.dirs="hdfs://myhdfs/my-remote-flink-dist-dir" \
+   hdfs://myhdfs/jars/my-application.jar
+{% endhighlight %}
+
+The above will allow the job submission to be extra lightweight as the needed 
Flink jars and the application jar
+are  going to be picked up by the specified remote locations rather than be 
shipped to the cluster by the 
+client.
+
+### Per-Job Cluster Mode
+
+The Per-job Cluster mode will launch a Flink cluster on YARN, then run the 
provided application jar locally and finally submit the JobGraph to the 
JobManager on YARN. If you pass the `--detached` argument, the client will stop 
once the submission is accepted.
+
+The YARN cluster will stop once the job has stopped.
+
+{% highlight bash %}
+./bin/flink run -m 

[jira] [Updated] (FLINK-20404) ZooKeeper quorum fails to start due to missing log4j library

2020-11-27 Thread Stephan Ewen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20404?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen updated FLINK-20404:
-
Priority: Blocker  (was: Major)

> ZooKeeper quorum fails to start due to missing log4j library
> 
>
> Key: FLINK-20404
> URL: https://issues.apache.org/jira/browse/FLINK-20404
> Project: Flink
>  Issue Type: Bug
>Affects Versions: 1.11.2
>Reporter: Pedro Miguel Rainho Chaves
>Priority: Blocker
> Fix For: 1.12.0, 1.11.3
>
>
> Upon starting a zookeeper quorum using flink's bootstrapped zookeeper, it 
> throws the following exception.
>  
> {code:java}
> 2020-11-27 13:13:38,371 ERROR 
> org.apache.flink.runtime.zookeeper.FlinkZooKeeperQuorumPeer  [] - Error 
> running ZooKeeper quorum peer: org/apache/log4j/jmx/HierarchyDynamicMBean
> java.lang.NoClassDefFoundError: org/apache/log4j/jmx/HierarchyDynamicMBean
> at 
> org.apache.flink.shaded.zookeeper3.org.apache.zookeeper.jmx.ManagedUtil.registerLog4jMBeans(ManagedUtil.java:51)
>  ~[flink-shaded-zookeeper-3.4.14.jar:3.4.14-11.0]
> at 
> org.apache.flink.shaded.zookeeper3.org.apache.zookeeper.server.quorum.QuorumPeerMain.runFromConfig(QuorumPeerMain.java:125)
>  ~[flink-shaded-zookeeper-3.4.14.jar:3.4.14-11.0]
> at 
> org.apache.flink.runtime.zookeeper.FlinkZooKeeperQuorumPeer.runFlinkZkQuorumPeer(FlinkZooKeeperQuorumPeer.java:123)
>  ~[flink-dist_2.11-1.11.2.jar:1.11.2]
> at 
> org.apache.flink.runtime.zookeeper.FlinkZooKeeperQuorumPeer.main(FlinkZooKeeperQuorumPeer.java:79)
>  [flink-dist_2.11-1.11.2.jar:1.11.2]
> Caused by: java.lang.ClassNotFoundException: 
> org.apache.log4j.jmx.HierarchyDynamicMBean
> at java.net.URLClassLoader.findClass(URLClassLoader.java:382) 
> ~[?:1.8.0_262]
> at java.lang.ClassLoader.loadClass(ClassLoader.java:418) 
> ~[?:1.8.0_262]
> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352) 
> ~[?:1.8.0_262]
> at java.lang.ClassLoader.loadClass(ClassLoader.java:351) 
> ~[?:1.8.0_262]
> ... 4 more
> {code}
> This happens because the new flink version is missing a log4j library. This 
> can be solved by adding log4j-1.2.17.jar to the classpath, nonetheless the 
> bootstrapped zookeepeer version should be compatible with the log4j2 
> libraries that come with flink's default installation.
>  
> *Steps to reproduce:*
>  # Fresh install of flink version 1.11.2 
>  # Change the zookeeper config to start as a quorum
> {code:java}
> server.1=:2888:3888
> server.2=:2888:3888{code}
>  # Start zookeeper
>  # /bin/zookeeper.sh start-foreground 1



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-20404) ZooKeeper quorum fails to start due to missing log4j library

2020-11-27 Thread Stephan Ewen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-20404?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Stephan Ewen updated FLINK-20404:
-
Fix Version/s: 1.11.3
   1.12.0

> ZooKeeper quorum fails to start due to missing log4j library
> 
>
> Key: FLINK-20404
> URL: https://issues.apache.org/jira/browse/FLINK-20404
> Project: Flink
>  Issue Type: Bug
>Affects Versions: 1.11.2
>Reporter: Pedro Miguel Rainho Chaves
>Priority: Major
> Fix For: 1.12.0, 1.11.3
>
>
> Upon starting a zookeeper quorum using flink's bootstrapped zookeeper, it 
> throws the following exception.
>  
> {code:java}
> 2020-11-27 13:13:38,371 ERROR 
> org.apache.flink.runtime.zookeeper.FlinkZooKeeperQuorumPeer  [] - Error 
> running ZooKeeper quorum peer: org/apache/log4j/jmx/HierarchyDynamicMBean
> java.lang.NoClassDefFoundError: org/apache/log4j/jmx/HierarchyDynamicMBean
> at 
> org.apache.flink.shaded.zookeeper3.org.apache.zookeeper.jmx.ManagedUtil.registerLog4jMBeans(ManagedUtil.java:51)
>  ~[flink-shaded-zookeeper-3.4.14.jar:3.4.14-11.0]
> at 
> org.apache.flink.shaded.zookeeper3.org.apache.zookeeper.server.quorum.QuorumPeerMain.runFromConfig(QuorumPeerMain.java:125)
>  ~[flink-shaded-zookeeper-3.4.14.jar:3.4.14-11.0]
> at 
> org.apache.flink.runtime.zookeeper.FlinkZooKeeperQuorumPeer.runFlinkZkQuorumPeer(FlinkZooKeeperQuorumPeer.java:123)
>  ~[flink-dist_2.11-1.11.2.jar:1.11.2]
> at 
> org.apache.flink.runtime.zookeeper.FlinkZooKeeperQuorumPeer.main(FlinkZooKeeperQuorumPeer.java:79)
>  [flink-dist_2.11-1.11.2.jar:1.11.2]
> Caused by: java.lang.ClassNotFoundException: 
> org.apache.log4j.jmx.HierarchyDynamicMBean
> at java.net.URLClassLoader.findClass(URLClassLoader.java:382) 
> ~[?:1.8.0_262]
> at java.lang.ClassLoader.loadClass(ClassLoader.java:418) 
> ~[?:1.8.0_262]
> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352) 
> ~[?:1.8.0_262]
> at java.lang.ClassLoader.loadClass(ClassLoader.java:351) 
> ~[?:1.8.0_262]
> ... 4 more
> {code}
> This happens because the new flink version is missing a log4j library. This 
> can be solved by adding log4j-1.2.17.jar to the classpath, nonetheless the 
> bootstrapped zookeepeer version should be compatible with the log4j2 
> libraries that come with flink's default installation.
>  
> *Steps to reproduce:*
>  # Fresh install of flink version 1.11.2 
>  # Change the zookeeper config to start as a quorum
> {code:java}
> server.1=:2888:3888
> server.2=:2888:3888{code}
>  # Start zookeeper
>  # /bin/zookeeper.sh start-foreground 1



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] rmetzger commented on a change in pull request #14238: [FLINK-20347] Rework YARN documentation page

2020-11-27 Thread GitBox


rmetzger commented on a change in pull request #14238:
URL: https://github.com/apache/flink/pull/14238#discussion_r531666867



##
File path: docs/deployment/resource-providers/yarn.md
##
@@ -0,0 +1,200 @@
+---
+title:  "Apache Hadoop YARN"
+nav-title: YARN
+nav-parent_id: resource_providers
+nav-pos: 4
+---
+
+
+* This will be replaced by the TOC
+{:toc}
+
+## Getting Started
+
+This *Getting Started* section guides you through setting up a fully 
functional Flink Cluster on YARN.
+
+### Introduction
+
+[Apache Hadoop 
YARN](https://hadoop.apache.org/docs/current/hadoop-yarn/hadoop-yarn-site/YARN.html)
 is a resource provider popular with many data processing frameworks.
+Flink services are submitted to YARN's ResourceManager, which spawns 
containers on machines managed by YARN NodeManagers. Flink deploys its 
JobManager and TaskManager instances into such containers.
+
+Flink can dynamically allocate and de-allocate TaskManager resources depending 
on the number of processing slots required by the job(s) running on the 
JobManager.
+
+### Preparation
+
+This *Getting Started* section assumes a functional YARN environment, starting 
from version 2.4.1. YARN environments are provided most conveniently through 
services such as Amazon EMR, Google Cloud DataProc or products like Cloudera. 
[Manually setting up a YARN environment 
locally](https://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/SingleCluster.html)
 or [on a 
cluster](https://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/ClusterSetup.html)
 is not recommended for following through this *Getting Started* tutorial. 
+
+
+- Make sure your YARN cluster is ready for accepting Flink applications by 
running `yarn top`. It should show no error messages.
+- Download a recent Flink distribution from the [download page]({{ 
site.download_url }}) and unpack it.
+- **Important** Make sure that the `HADOOP_CLASSPATH` environment variable is 
set up (it can be checked by running `echo $HADOOP_CLASSPATH`). If not, set it 
up using 
+
+{% highlight bash %}
+export HADOOP_CLASSPATH=`hadoop classpath`
+{% endhighlight %}
+
+
+### Starting a Flink Session on YARN
+
+Once you've made sure that the `HADOOP_CLASSPATH` environment variable is set, 
we can launch a Flink on YARN session, and submit an example job:
+
+{% highlight bash %}
+
+# we assume to be in the root directory of the unzipped Flink distribution
+
+# (0) export HADOOP_CLASSPATH
+export HADOOP_CLASSPATH=`hadoop classpath`
+
+# (1) Start YARN Session
+./bin/yarn-session.sh --detached
+
+# (2) You can now access the Flink Web Interface through the URL printed in 
the last lines of the command output, or through the YARN ResourceManager web 
UI.
+
+# (3) Submit example job
+./bin/flink run ./examples/streaming/TopSpeedWindowing.jar
+
+# (4) Stop YARN session (replace the application id based on the output of the 
yarn-session.sh command)
+echo "stop" | ./bin/yarn-session.sh -yid application_X_XXX
+{% endhighlight %}
+
+Congratulations! You have successfully run a Flink application by deploying 
Flink on YARN.
+
+
+## Deployment Modes Supported by Flink on YARN
+
+For production use, we recommend deploying Flink Applications in the [Per-job 
or Application Mode]({% link deployment/index.md %}#deployment-modes), as these 
modes provide a better isolation for the Applications.
+
+### Application Mode
+
+Application Mode will launch a Flink cluster on YARN, where the main() method 
of the application jar gets executed on the JobManager in YARN.
+The cluster will shut down as soon as the application has finished. You can 
manually stop the cluster using `yarn application -kill ` or by 
cancelling the Flink job.
+
+{% highlight bash %}
+./bin/flink run-application -t yarn-application 
./examples/streaming/TopSpeedWindowing.jar
+{% endhighlight %}
+
+To unlock the full potential of the application mode, consider using it with 
the `yarn.provided.lib.dirs` configuration option
+and pre-upload your application jar to a location accessible by all nodes in 
your cluster. In this case, the 
+command could look like: 
+
+{% highlight bash %}
+./bin/flink run-application -t yarn-application \
+   -Dyarn.provided.lib.dirs="hdfs://myhdfs/my-remote-flink-dist-dir" \
+   hdfs://myhdfs/jars/my-application.jar
+{% endhighlight %}
+
+The above will allow the job submission to be extra lightweight as the needed 
Flink jars and the application jar
+are  going to be picked up by the specified remote locations rather than be 
shipped to the cluster by the 
+client.
+
+### Per-Job Cluster Mode
+
+The Per-job Cluster mode will launch a Flink cluster on YARN, then run the 
provided application jar locally and finally submit the JobGraph to the 
JobManager on YARN. If you pass the `--detached` argument, the client will stop 
once the submission is accepted.
+
+The YARN cluster will stop once the job has stopped.
+
+{% highlight bash %}
+./bin/flink run -m 

[GitHub] [flink] XComp commented on a change in pull request #14238: [FLINK-20347] Rework YARN documentation page

2020-11-27 Thread GitBox


XComp commented on a change in pull request #14238:
URL: https://github.com/apache/flink/pull/14238#discussion_r531663140



##
File path: docs/deployment/resource-providers/yarn.md
##
@@ -0,0 +1,200 @@
+---
+title:  "Apache Hadoop YARN"
+nav-title: YARN
+nav-parent_id: resource_providers
+nav-pos: 4
+---
+
+
+* This will be replaced by the TOC
+{:toc}
+
+## Getting Started
+
+This *Getting Started* section guides you through setting up a fully 
functional Flink Cluster on YARN.
+
+### Introduction
+
+[Apache Hadoop 
YARN](https://hadoop.apache.org/docs/current/hadoop-yarn/hadoop-yarn-site/YARN.html)
 is a resource provider popular with many data processing frameworks.
+Flink services are submitted to YARN's ResourceManager, which spawns 
containers on machines managed by YARN NodeManagers. Flink deploys its 
JobManager and TaskManager instances into such containers.
+
+Flink can dynamically allocate and de-allocate TaskManager resources depending 
on the number of processing slots required by the job(s) running on the 
JobManager.
+
+### Preparation
+
+This *Getting Started* section assumes a functional YARN environment, starting 
from version 2.4.1. YARN environments are provided most conveniently through 
services such as Amazon EMR, Google Cloud DataProc or products like Cloudera. 
[Manually setting up a YARN environment 
locally](https://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/SingleCluster.html)
 or [on a 
cluster](https://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-common/ClusterSetup.html)
 is not recommended for following through this *Getting Started* tutorial. 
+
+
+- Make sure your YARN cluster is ready for accepting Flink applications by 
running `yarn top`. It should show no error messages.
+- Download a recent Flink distribution from the [download page]({{ 
site.download_url }}) and unpack it.
+- **Important** Make sure that the `HADOOP_CLASSPATH` environment variable is 
set up (it can be checked by running `echo $HADOOP_CLASSPATH`). If not, set it 
up using 
+
+{% highlight bash %}
+export HADOOP_CLASSPATH=`hadoop classpath`
+{% endhighlight %}
+
+
+### Starting a Flink Session on YARN
+
+Once you've made sure that the `HADOOP_CLASSPATH` environment variable is set, 
we can launch a Flink on YARN session, and submit an example job:
+
+{% highlight bash %}
+
+# we assume to be in the root directory of the unzipped Flink distribution
+
+# (0) export HADOOP_CLASSPATH
+export HADOOP_CLASSPATH=`hadoop classpath`
+
+# (1) Start YARN Session
+./bin/yarn-session.sh --detached
+
+# (2) You can now access the Flink Web Interface through the URL printed in 
the last lines of the command output, or through the YARN ResourceManager web 
UI.
+
+# (3) Submit example job
+./bin/flink run ./examples/streaming/TopSpeedWindowing.jar
+
+# (4) Stop YARN session (replace the application id based on the output of the 
yarn-session.sh command)
+echo "stop" | ./bin/yarn-session.sh -yid application_X_XXX
+{% endhighlight %}
+
+Congratulations! You have successfully run a Flink application by deploying 
Flink on YARN.
+
+
+## Deployment Modes Supported by Flink on YARN
+
+For production use, we recommend deploying Flink Applications in the [Per-job 
or Application Mode]({% link deployment/index.md %}#deployment-modes), as these 
modes provide a better isolation for the Applications.
+
+### Application Mode
+
+Application Mode will launch a Flink cluster on YARN, where the main() method 
of the application jar gets executed on the JobManager in YARN.
+The cluster will shut down as soon as the application has finished. You can 
manually stop the cluster using `yarn application -kill ` or by 
cancelling the Flink job.
+
+{% highlight bash %}
+./bin/flink run-application -t yarn-application 
./examples/streaming/TopSpeedWindowing.jar
+{% endhighlight %}
+
+To unlock the full potential of the application mode, consider using it with 
the `yarn.provided.lib.dirs` configuration option
+and pre-upload your application jar to a location accessible by all nodes in 
your cluster. In this case, the 
+command could look like: 
+
+{% highlight bash %}
+./bin/flink run-application -t yarn-application \
+   -Dyarn.provided.lib.dirs="hdfs://myhdfs/my-remote-flink-dist-dir" \
+   hdfs://myhdfs/jars/my-application.jar
+{% endhighlight %}
+
+The above will allow the job submission to be extra lightweight as the needed 
Flink jars and the application jar
+are  going to be picked up by the specified remote locations rather than be 
shipped to the cluster by the 
+client.
+
+### Per-Job Cluster Mode
+
+The Per-job Cluster mode will launch a Flink cluster on YARN, then run the 
provided application jar locally and finally submit the JobGraph to the 
JobManager on YARN. If you pass the `--detached` argument, the client will stop 
once the submission is accepted.
+
+The YARN cluster will stop once the job has stopped.
+
+{% highlight bash %}
+./bin/flink run -m 

[GitHub] [flink] flinkbot edited a comment on pull request #14199: [FLINK-20118][file connector] Extend FileSourceTextLinesITCase with JM/TM failover tests

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14199:
URL: https://github.com/apache/flink/pull/14199#issuecomment-732845197


   
   ## CI report:
   
   * 8139d3ee5e5fd595279a4a4c6e7849d1a9dfc6b5 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10049)
 
   * 6ffe3e161a8ec42b7ce5da3c46b57587d98a1f90 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14204: [FLINK-20325][build] Move docs_404_check to CI stage

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14204:
URL: https://github.com/apache/flink/pull/14204#issuecomment-733048767


   
   ## CI report:
   
   * 6550d3e1b01af94d4f652f993834b75272da1020 UNKNOWN
   * 38ed086f80ee2a94613a716d090574beed92aef4 UNKNOWN
   * d97f8149d6fcb9985cd75c4cd8f30b9cd10f14e5 Azure: 
[CANCELED](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10233)
 
   * 983518fa6cdff38b3264506a9ca54178699c9525 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10257)
 
   * b4d5fbae5b72a01ab9e94bc000373d6797339f9a Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10266)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] azagrebin commented on pull request #14199: [FLINK-20118][file connector] Extend FileSourceTextLinesITCase with JM/TM failover tests

2020-11-27 Thread GitBox


azagrebin commented on pull request #14199:
URL: https://github.com/apache/flink/pull/14199#issuecomment-734883917


   I added TM failure for batch case.
   
   The JM failure is tricky w/o synchronization with the job execution. The job 
gets done before JM failure.
   We could make `EmbeddedHaServices::getJobManagerLeaderService` public to see 
whether the JM is still available to restart but this is fragile to 
concurrency. Hence, I am not sure about the value of this test.
   
   I added JM failure test for batch with the job execution synchronization, 
abstracting it away, in the last commit. I can remove the last commit if it 
looks too complicated and there is no better idea how to do the JM failure test 
for batch in a simpler way.
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Created] (FLINK-20404) ZooKeeper quorum fails to start due to missing log4j library

2020-11-27 Thread Pedro Miguel Rainho Chaves (Jira)
Pedro Miguel Rainho Chaves created FLINK-20404:
--

 Summary: ZooKeeper quorum fails to start due to missing log4j 
library
 Key: FLINK-20404
 URL: https://issues.apache.org/jira/browse/FLINK-20404
 Project: Flink
  Issue Type: Bug
Affects Versions: 1.11.2
Reporter: Pedro Miguel Rainho Chaves


Upon starting a zookeeper quorum using flink's bootstrapped zookeeper, it 
throws the following exception.

 
{code:java}
2020-11-27 13:13:38,371 ERROR 
org.apache.flink.runtime.zookeeper.FlinkZooKeeperQuorumPeer  [] - Error running 
ZooKeeper quorum peer: org/apache/log4j/jmx/HierarchyDynamicMBean
java.lang.NoClassDefFoundError: org/apache/log4j/jmx/HierarchyDynamicMBean
at 
org.apache.flink.shaded.zookeeper3.org.apache.zookeeper.jmx.ManagedUtil.registerLog4jMBeans(ManagedUtil.java:51)
 ~[flink-shaded-zookeeper-3.4.14.jar:3.4.14-11.0]
at 
org.apache.flink.shaded.zookeeper3.org.apache.zookeeper.server.quorum.QuorumPeerMain.runFromConfig(QuorumPeerMain.java:125)
 ~[flink-shaded-zookeeper-3.4.14.jar:3.4.14-11.0]
at 
org.apache.flink.runtime.zookeeper.FlinkZooKeeperQuorumPeer.runFlinkZkQuorumPeer(FlinkZooKeeperQuorumPeer.java:123)
 ~[flink-dist_2.11-1.11.2.jar:1.11.2]
at 
org.apache.flink.runtime.zookeeper.FlinkZooKeeperQuorumPeer.main(FlinkZooKeeperQuorumPeer.java:79)
 [flink-dist_2.11-1.11.2.jar:1.11.2]
Caused by: java.lang.ClassNotFoundException: 
org.apache.log4j.jmx.HierarchyDynamicMBean
at java.net.URLClassLoader.findClass(URLClassLoader.java:382) 
~[?:1.8.0_262]
at java.lang.ClassLoader.loadClass(ClassLoader.java:418) ~[?:1.8.0_262]
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352) 
~[?:1.8.0_262]
at java.lang.ClassLoader.loadClass(ClassLoader.java:351) ~[?:1.8.0_262]
... 4 more
{code}
This happens because the new flink version is missing a log4j library. This can 
be solved by adding log4j-1.2.17.jar to the classpath, nonetheless the 
bootstrapped zookeepeer version should be compatible with the log4j2 libraries 
that come with flink's default installation.

 

*Steps to reproduce:*
 # Fresh install of flink version 1.11.2 
 # Change the zookeeper config to start as a quorum
{code:java}
server.1=:2888:3888
server.2=:2888:3888{code}

 # Start zookeeper
 # /bin/zookeeper.sh start-foreground 1



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #14234: Make FileSink Support migrating from StreamingFileSink

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14234:
URL: https://github.com/apache/flink/pull/14234#issuecomment-734232999


   
   ## CI report:
   
   * b461ede54e0564d0b1fcab49aef07c7b9a3c107a Azure: 
[CANCELED](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10247)
 
   * c89575f22a01309ffb329e1d0ecdc07b3c99fc91 UNKNOWN
   * 3f569564dae3e5ee5a9b76ad86dc3acf9685d9df Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10255)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14238: [FLINK-20347] Rework YARN documentation page

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14238:
URL: https://github.com/apache/flink/pull/14238#issuecomment-734336545


   
   ## CI report:
   
   * 7806bb7e0dbca1263ddffb771dec367807f0590a Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10261)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #14204: [FLINK-20325][build] Move docs_404_check to CI stage

2020-11-27 Thread GitBox


flinkbot edited a comment on pull request #14204:
URL: https://github.com/apache/flink/pull/14204#issuecomment-733048767


   
   ## CI report:
   
   * 6550d3e1b01af94d4f652f993834b75272da1020 UNKNOWN
   * 38ed086f80ee2a94613a716d090574beed92aef4 UNKNOWN
   * d97f8149d6fcb9985cd75c4cd8f30b9cd10f14e5 Azure: 
[CANCELED](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10233)
 
   * 983518fa6cdff38b3264506a9ca54178699c9525 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=10257)
 
   * b4d5fbae5b72a01ab9e94bc000373d6797339f9a UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Resolved] (FLINK-19852) Managed memory released check can block IterativeTask

2020-11-27 Thread Arvid Heise (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-19852?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arvid Heise resolved FLINK-19852.
-
Resolution: Fixed

> Managed memory released check can block IterativeTask
> -
>
> Key: FLINK-19852
> URL: https://issues.apache.org/jira/browse/FLINK-19852
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Task
>Affects Versions: 1.11.0, 1.10.2, 1.12.0, 1.11.1, 1.11.2
>Reporter: shaomeng.wang
>Assignee: Roman Khachatryan
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.12.0, 1.11.3
>
> Attachments: image-2020-10-28-17-48-28-395.png, 
> image-2020-10-28-17-48-48-583.png
>
>
> UnsafeMemoryBudget#reserveMemory, called on TempBarrier, needs time to wait 
> on GC of all allocated/released managed memory at every iteration.
>  
> stack:
> !image-2020-10-28-17-48-48-583.png!
> new TempBarrier in BatchTask
> !image-2020-10-28-17-48-28-395.png!
>  
> These will be very slow than before.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-19852) Managed memory released check can block IterativeTask

2020-11-27 Thread Arvid Heise (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-19852?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17239718#comment-17239718
 ] 

Arvid Heise commented on FLINK-19852:
-

Merged into 1.11 as 05a7875448b6f92450083c17153e7f21e7de2e31.

> Managed memory released check can block IterativeTask
> -
>
> Key: FLINK-19852
> URL: https://issues.apache.org/jira/browse/FLINK-19852
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Task
>Affects Versions: 1.11.0, 1.10.2, 1.12.0, 1.11.1, 1.11.2
>Reporter: shaomeng.wang
>Assignee: Roman Khachatryan
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.12.0, 1.11.3
>
> Attachments: image-2020-10-28-17-48-28-395.png, 
> image-2020-10-28-17-48-48-583.png
>
>
> UnsafeMemoryBudget#reserveMemory, called on TempBarrier, needs time to wait 
> on GC of all allocated/released managed memory at every iteration.
>  
> stack:
> !image-2020-10-28-17-48-48-583.png!
> new TempBarrier in BatchTask
> !image-2020-10-28-17-48-28-395.png!
>  
> These will be very slow than before.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] AHeise merged pull request #14221: (1.11) [FLINK-19852][task] Reuse TempBarrier memory between iterations

2020-11-27 Thread GitBox


AHeise merged pull request #14221:
URL: https://github.com/apache/flink/pull/14221


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Comment Edited] (FLINK-20392) Migrating bash e2e tests to Java/Docker

2020-11-27 Thread Dawid Wysakowicz (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-20392?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17239716#comment-17239716
 ] 

Dawid Wysakowicz edited comment on FLINK-20392 at 11/27/20, 2:57 PM:
-

Just to add to the FLINK-19997, I set up a FlinkContainer that runs a Flink 
cluster within a docker container with a subset of functionalities provided by 
the FlinkResource. It can be a good base to move the FlinkResource onto 
testcontainers entirely.


was (Author: dawidwys):
Just to add to the FLINK-19997, I set up a FlinkContainer that runs a Flink 
cluster within a docker container with a subset of functionalities provided by 
the FlinkResource. It can be a good base to move the FlinkResource onto 
testcontainers.

> Migrating bash e2e tests to Java/Docker
> ---
>
> Key: FLINK-20392
> URL: https://issues.apache.org/jira/browse/FLINK-20392
> Project: Flink
>  Issue Type: Test
>  Components: Test Infrastructure, Tests
>Reporter: Matthias
>Priority: Major
>
> This Jira issue serves as an umbrella ticket for single e2e test migration 
> tasks. This should enable us to migrate all bash-based e2e tests step-by-step.
> The goal is to utilize the e2e test framework (see 
> [flink-end-to-end-tests-common|https://github.com/apache/flink/tree/master/flink-end-to-end-tests/flink-end-to-end-tests-common]).
>  Ideally, the test should use Docker containers as much as possible 
> disconnect the execution from the environment. A good source to achieve that 
> is [testcontainers.org|https://www.testcontainers.org/].
> The related ML discussion is [Stop adding new bash-based e2e tests to 
> Flink|http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/DISCUSS-Stop-adding-new-bash-based-e2e-tests-to-Flink-td46607.html].



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


  1   2   3   4   >