[GitHub] [flink] flinkbot edited a comment on pull request #12422: [FLINK-17901][python] Add module interface in PyFlink

2020-06-02 Thread GitBox


flinkbot edited a comment on pull request #12422:
URL: https://github.com/apache/flink/pull/12422#issuecomment-636647115


   
   ## CI report:
   
   * 3a387da20f29d9b1f6baf1e57c6d9979d27dd81f Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2509)
 
   * a57beb047e52bacd60d9c19cbcd31bc9b1d88ec4 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2599)
 
   * 2a1da9cc0e95b2a429a577fd83955380e30e323c UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Created] (FLINK-18082) UnsignedTypeConversionITCase stalls in ch.vorburger.mariadb4j.DB.stop

2020-06-02 Thread Robert Metzger (Jira)
Robert Metzger created FLINK-18082:
--

 Summary: UnsignedTypeConversionITCase stalls in 
ch.vorburger.mariadb4j.DB.stop
 Key: FLINK-18082
 URL: https://issues.apache.org/jira/browse/FLINK-18082
 Project: Flink
  Issue Type: Bug
  Components: Connectors / JDBC
Affects Versions: 1.12.0
Reporter: Robert Metzger


CI: 
https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=2582=logs=d44f43ce-542c-597d-bf94-b0718c71e5e8=03dca39c-73e8-5aaf-601d-328ae5c35f20

{code}
2020-06-02T19:01:31.8486456Z 
==
2020-06-02T19:01:31.8487052Z Printing stack trace of Java process 8653
2020-06-02T19:01:31.8487424Z 
==
2020-06-02T19:01:31.8541169Z Picked up JAVA_TOOL_OPTIONS: 
-XX:+HeapDumpOnOutOfMemoryError
2020-06-02T19:01:32.1665740Z 2020-06-02 19:01:32
2020-06-02T19:01:32.1666470Z Full thread dump OpenJDK 64-Bit Server VM 
(25.242-b08 mixed mode):
2020-06-02T19:01:32.1666735Z 
2020-06-02T19:01:32.1667614Z "Attach Listener" #537 daemon prio=9 os_prio=0 
tid=0x7f61f8001000 nid=0x3b9f waiting on condition [0x]
2020-06-02T19:01:32.1668130Zjava.lang.Thread.State: RUNNABLE
2020-06-02T19:01:32.1668311Z 
2020-06-02T19:01:32.1668958Z "flink-akka.actor.default-dispatcher-193" #535 
prio=5 os_prio=0 tid=0x7f6034001000 nid=0x3af7 waiting on condition 
[0x7f61a25b8000]
2020-06-02T19:01:32.1669418Zjava.lang.Thread.State: TIMED_WAITING (parking)
2020-06-02T19:01:32.1669730Zat sun.misc.Unsafe.park(Native Method)
2020-06-02T19:01:32.1670301Z- parking to wait for  <0x80c51528> (a 
akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinPool)
2020-06-02T19:01:32.1670791Zat 
akka.dispatch.forkjoin.ForkJoinPool.idleAwaitWork(ForkJoinPool.java:2135)
2020-06-02T19:01:32.1671329Zat 
akka.dispatch.forkjoin.ForkJoinPool.scan(ForkJoinPool.java:2067)
2020-06-02T19:01:32.1671763Zat 
akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
2020-06-02T19:01:32.1672211Zat 
akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
2020-06-02T19:01:32.1672491Z 
2020-06-02T19:01:32.1673104Z "flink-akka.actor.default-dispatcher-191" #533 
prio=5 os_prio=0 tid=0x7f619801e000 nid=0x3ae1 waiting on condition 
[0x7f60770f1000]
2020-06-02T19:01:32.1673564Zjava.lang.Thread.State: WAITING (parking)
2020-06-02T19:01:32.1673839Zat sun.misc.Unsafe.park(Native Method)
2020-06-02T19:01:32.1674422Z- parking to wait for  <0x80c51528> (a 
akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinPool)
2020-06-02T19:01:32.1674865Zat 
akka.dispatch.forkjoin.ForkJoinPool.scan(ForkJoinPool.java:2075)
2020-06-02T19:01:32.1675305Zat 
akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
2020-06-02T19:01:32.1675751Zat 
akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
2020-06-02T19:01:32.1676046Z 
2020-06-02T19:01:32.1676669Z "jobmanager-future-thread-2" #466 daemon prio=5 
os_prio=0 tid=0x7f6124001000 nid=0x3795 waiting on condition 
[0x7f61a23b6000]
2020-06-02T19:01:32.1677316Zjava.lang.Thread.State: WAITING (parking)
2020-06-02T19:01:32.1677617Zat sun.misc.Unsafe.park(Native Method)
2020-06-02T19:01:32.1678220Z- parking to wait for  <0x816a4c90> (a 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
2020-06-02T19:01:32.1678702Zat 
java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
2020-06-02T19:01:32.1679209Zat 
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2039)
2020-06-02T19:01:32.1679822Zat 
java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:1088)
2020-06-02T19:01:32.1680422Zat 
java.util.concurrent.ScheduledThreadPoolExecutor$DelayedWorkQueue.take(ScheduledThreadPoolExecutor.java:809)
2020-06-02T19:01:32.1680962Zat 
java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1074)
2020-06-02T19:01:32.1681424Zat 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1134)
2020-06-02T19:01:32.1682062Zat 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
2020-06-02T19:01:32.1682445Zat java.lang.Thread.run(Thread.java:748)
2020-06-02T19:01:32.1682656Z 
2020-06-02T19:01:32.1683271Z "Flink-DispatcherRestEndpoint-thread-4" #349 
daemon prio=5 os_prio=0 tid=0x7f618c00a000 nid=0x29a4 waiting on condition 
[0x7f61a029f000]
2020-06-02T19:01:32.1683750Zjava.lang.Thread.State: TIMED_WAITING (parking)
2020-06-02T19:01:32.1684057Zat sun.misc.Unsafe.park(Native Method)
2020-06-02T19:01:32.1684648Z- parking to wait for  <0x81731ff8> 

[GitHub] [flink] godfreyhe commented on a change in pull request #12431: [FLINK-18055] [sql-client] Fix catalog/database does not exist in sql client

2020-06-02 Thread GitBox


godfreyhe commented on a change in pull request #12431:
URL: https://github.com/apache/flink/pull/12431#discussion_r434322529



##
File path: 
flink-table/flink-sql-client/src/main/java/org/apache/flink/table/client/cli/SqlCommandParser.java
##
@@ -132,11 +132,10 @@ private SqlCommandParser() {
cmd = SqlCommand.DROP_CATALOG;
} else if (operation instanceof UseCatalogOperation) {
cmd = SqlCommand.USE_CATALOG;
-   operands = new String[] { String.format("`%s`", 
((UseCatalogOperation) operation).getCatalogName()) };
+   operands = new String[] { ((UseCatalogOperation) 
operation).getCatalogName() };
} else if (operation instanceof UseDatabaseOperation) {
cmd = SqlCommand.USE;
-   UseDatabaseOperation op = ((UseDatabaseOperation) 
operation);
-   operands = new String[] { String.format("`%s`.`%s`", 
op.getCatalogName(), op.getDatabaseName()) };
+   operands = new String[] { ((UseDatabaseOperation) 
operation).getDatabaseName() };

Review comment:
   catalog name is unnecessary, only database is needed.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] godfreyhe commented on a change in pull request #12431: [FLINK-18055] [sql-client] Fix catalog/database does not exist in sql client

2020-06-02 Thread GitBox


godfreyhe commented on a change in pull request #12431:
URL: https://github.com/apache/flink/pull/12431#discussion_r434321402



##
File path: 
flink-table/flink-sql-client/src/main/java/org/apache/flink/table/client/cli/SqlCommandParser.java
##
@@ -132,11 +132,10 @@ private SqlCommandParser() {
cmd = SqlCommand.DROP_CATALOG;
} else if (operation instanceof UseCatalogOperation) {
cmd = SqlCommand.USE_CATALOG;
-   operands = new String[] { String.format("`%s`", 
((UseCatalogOperation) operation).getCatalogName()) };
+   operands = new String[] { ((UseCatalogOperation) 
operation).getCatalogName() };

Review comment:
   The quoting string ("\`")  is skipped by sql parser, so we need not to 
handle this. see the test (TestItem.validSql("use \`default\`;", 
SqlCommand.USE, "default")) in `SqlCommandParserTest#testCommands` 





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] godfreyhe commented on a change in pull request #12431: [FLINK-18055] [sql-client] Fix catalog/database does not exist in sql client

2020-06-02 Thread GitBox


godfreyhe commented on a change in pull request #12431:
URL: https://github.com/apache/flink/pull/12431#discussion_r434321402



##
File path: 
flink-table/flink-sql-client/src/main/java/org/apache/flink/table/client/cli/SqlCommandParser.java
##
@@ -132,11 +132,10 @@ private SqlCommandParser() {
cmd = SqlCommand.DROP_CATALOG;
} else if (operation instanceof UseCatalogOperation) {
cmd = SqlCommand.USE_CATALOG;
-   operands = new String[] { String.format("`%s`", 
((UseCatalogOperation) operation).getCatalogName()) };
+   operands = new String[] { ((UseCatalogOperation) 
operation).getCatalogName() };

Review comment:
   The quoting string \`  is skipped by sql parser, so we need not to 
handle this. see the test (`TestItem.validSql("use `default`;", SqlCommand.USE, 
"default")`) in `SqlCommandParserTest#testCommands` 

##
File path: 
flink-table/flink-sql-client/src/main/java/org/apache/flink/table/client/cli/SqlCommandParser.java
##
@@ -132,11 +132,10 @@ private SqlCommandParser() {
cmd = SqlCommand.DROP_CATALOG;
} else if (operation instanceof UseCatalogOperation) {
cmd = SqlCommand.USE_CATALOG;
-   operands = new String[] { String.format("`%s`", 
((UseCatalogOperation) operation).getCatalogName()) };
+   operands = new String[] { ((UseCatalogOperation) 
operation).getCatalogName() };

Review comment:
   The quoting string ("\`")  is skipped by sql parser, so we need not to 
handle this. see the test (`TestItem.validSql("use `default`;", SqlCommand.USE, 
"default")`) in `SqlCommandParserTest#testCommands` 





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] godfreyhe commented on a change in pull request #12431: [FLINK-18055] [sql-client] Fix catalog/database does not exist in sql client

2020-06-02 Thread GitBox


godfreyhe commented on a change in pull request #12431:
URL: https://github.com/apache/flink/pull/12431#discussion_r434321402



##
File path: 
flink-table/flink-sql-client/src/main/java/org/apache/flink/table/client/cli/SqlCommandParser.java
##
@@ -132,11 +132,10 @@ private SqlCommandParser() {
cmd = SqlCommand.DROP_CATALOG;
} else if (operation instanceof UseCatalogOperation) {
cmd = SqlCommand.USE_CATALOG;
-   operands = new String[] { String.format("`%s`", 
((UseCatalogOperation) operation).getCatalogName()) };
+   operands = new String[] { ((UseCatalogOperation) 
operation).getCatalogName() };

Review comment:
   The quoting string ("\`")  is skipped by sql parser, so we need not to 
handle this. see the test (`TestItem.validSql("use \`default\`;", 
SqlCommand.USE, "default")`) in `SqlCommandParserTest#testCommands` 





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-18081) Fix broken links in "Kerberos Authentication Setup and Configuration" doc

2020-06-02 Thread Yangze Guo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18081?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17124627#comment-17124627
 ] 

Yangze Guo commented on FLINK-18081:


Could someone assign this to me?

> Fix broken links in "Kerberos Authentication Setup and Configuration" doc
> -
>
> Key: FLINK-18081
> URL: https://issues.apache.org/jira/browse/FLINK-18081
> Project: Flink
>  Issue Type: Bug
>Reporter: Yangze Guo
>Priority: Major
>
> The {{config.html#kerberos-based-security}} is not valid now.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-18081) Fix broken links in "Kerberos Authentication Setup and Configuration" doc

2020-06-02 Thread Yangze Guo (Jira)
Yangze Guo created FLINK-18081:
--

 Summary: Fix broken links in "Kerberos Authentication Setup and 
Configuration" doc
 Key: FLINK-18081
 URL: https://issues.apache.org/jira/browse/FLINK-18081
 Project: Flink
  Issue Type: Bug
Reporter: Yangze Guo


The {{config.html#kerberos-based-security}} is not valid now.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] godfreyhe commented on a change in pull request #12431: [FLINK-18055] [sql-client] Fix catalog/database does not exist in sql client

2020-06-02 Thread GitBox


godfreyhe commented on a change in pull request #12431:
URL: https://github.com/apache/flink/pull/12431#discussion_r434321402



##
File path: 
flink-table/flink-sql-client/src/main/java/org/apache/flink/table/client/cli/SqlCommandParser.java
##
@@ -132,11 +132,10 @@ private SqlCommandParser() {
cmd = SqlCommand.DROP_CATALOG;
} else if (operation instanceof UseCatalogOperation) {
cmd = SqlCommand.USE_CATALOG;
-   operands = new String[] { String.format("`%s`", 
((UseCatalogOperation) operation).getCatalogName()) };
+   operands = new String[] { ((UseCatalogOperation) 
operation).getCatalogName() };

Review comment:
   The quoting string `  is skipped by sql parser, so we need not to handle 
this. see the test (`TestItem.validSql("use `default`;", SqlCommand.USE, 
"default")`) in `SqlCommandParserTest#testCommands` 





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] godfreyhe commented on a change in pull request #12431: [FLINK-18055] [sql-client] Fix catalog/database does not exist in sql client

2020-06-02 Thread GitBox


godfreyhe commented on a change in pull request #12431:
URL: https://github.com/apache/flink/pull/12431#discussion_r434321402



##
File path: 
flink-table/flink-sql-client/src/main/java/org/apache/flink/table/client/cli/SqlCommandParser.java
##
@@ -132,11 +132,10 @@ private SqlCommandParser() {
cmd = SqlCommand.DROP_CATALOG;
} else if (operation instanceof UseCatalogOperation) {
cmd = SqlCommand.USE_CATALOG;
-   operands = new String[] { String.format("`%s`", 
((UseCatalogOperation) operation).getCatalogName()) };
+   operands = new String[] { ((UseCatalogOperation) 
operation).getCatalogName() };

Review comment:
   The quoting string ` \` ` is skipped by sql parser, so we need not to 
handle this. see the test (`TestItem.validSql("use `default`;", SqlCommand.USE, 
"default")`) in `SqlCommandParserTest#testCommands` 





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] godfreyhe commented on a change in pull request #12431: [FLINK-18055] [sql-client] Fix catalog/database does not exist in sql client

2020-06-02 Thread GitBox


godfreyhe commented on a change in pull request #12431:
URL: https://github.com/apache/flink/pull/12431#discussion_r434321402



##
File path: 
flink-table/flink-sql-client/src/main/java/org/apache/flink/table/client/cli/SqlCommandParser.java
##
@@ -132,11 +132,10 @@ private SqlCommandParser() {
cmd = SqlCommand.DROP_CATALOG;
} else if (operation instanceof UseCatalogOperation) {
cmd = SqlCommand.USE_CATALOG;
-   operands = new String[] { String.format("`%s`", 
((UseCatalogOperation) operation).getCatalogName()) };
+   operands = new String[] { ((UseCatalogOperation) 
operation).getCatalogName() };

Review comment:
   The quoting string ` "`" ` is skipped by sql parser, so we need not to 
handle this. see the test (`TestItem.validSql("use `default`;", SqlCommand.USE, 
"default")`) in `SqlCommandParserTest#testCommands` 





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] godfreyhe commented on a change in pull request #12431: [FLINK-18055] [sql-client] Fix catalog/database does not exist in sql client

2020-06-02 Thread GitBox


godfreyhe commented on a change in pull request #12431:
URL: https://github.com/apache/flink/pull/12431#discussion_r434321402



##
File path: 
flink-table/flink-sql-client/src/main/java/org/apache/flink/table/client/cli/SqlCommandParser.java
##
@@ -132,11 +132,10 @@ private SqlCommandParser() {
cmd = SqlCommand.DROP_CATALOG;
} else if (operation instanceof UseCatalogOperation) {
cmd = SqlCommand.USE_CATALOG;
-   operands = new String[] { String.format("`%s`", 
((UseCatalogOperation) operation).getCatalogName()) };
+   operands = new String[] { ((UseCatalogOperation) 
operation).getCatalogName() };

Review comment:
   The quoting string ``` is skipped by sql parser, so we need not to 
handle this. see the test (`TestItem.validSql("use `default`;", SqlCommand.USE, 
"default")`) in `SqlCommandParserTest#testCommands` 

##
File path: 
flink-table/flink-sql-client/src/main/java/org/apache/flink/table/client/cli/SqlCommandParser.java
##
@@ -132,11 +132,10 @@ private SqlCommandParser() {
cmd = SqlCommand.DROP_CATALOG;
} else if (operation instanceof UseCatalogOperation) {
cmd = SqlCommand.USE_CATALOG;
-   operands = new String[] { String.format("`%s`", 
((UseCatalogOperation) operation).getCatalogName()) };
+   operands = new String[] { ((UseCatalogOperation) 
operation).getCatalogName() };

Review comment:
   The quoting string ` ` ` is skipped by sql parser, so we need not to 
handle this. see the test (`TestItem.validSql("use `default`;", SqlCommand.USE, 
"default")`) in `SqlCommandParserTest#testCommands` 





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-17851) FlinkKafkaProducerITCase.testResumeTransaction:108->assertRecord NoSuchElement

2020-06-02 Thread Robert Metzger (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-17851?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17124623#comment-17124623
 ] 

Robert Metzger commented on FLINK-17851:


https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=2586=logs=4be4ed2b-549a-533d-aa33-09e28e360cc8=0db94045-2aa0-53fa-f444-0130d6933518

> FlinkKafkaProducerITCase.testResumeTransaction:108->assertRecord NoSuchElement
> --
>
> Key: FLINK-17851
> URL: https://issues.apache.org/jira/browse/FLINK-17851
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kafka, Tests
>Affects Versions: 1.11.0
>Reporter: Robert Metzger
>Priority: Critical
>  Labels: test-stability
> Fix For: 1.11.0
>
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=1957=logs=d44f43ce-542c-597d-bf94-b0718c71e5e8=34f486e1-e1e4-5dd2-9c06-bfdd9b9c74a8
> {code}
> 2020-05-20T16:36:17.1033719Z [ERROR] Tests run: 10, Failures: 0, Errors: 1, 
> Skipped: 0, Time elapsed: 68.936 s <<< FAILURE! - in 
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducerITCase
> 2020-05-20T16:36:17.1035011Z [ERROR] 
> testResumeTransaction(org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducerITCase)
>   Time elapsed: 11.918 s  <<< ERROR!
> 2020-05-20T16:36:17.1036296Z java.util.NoSuchElementException
> 2020-05-20T16:36:17.1036802Z  at 
> org.apache.kafka.common.utils.AbstractIterator.next(AbstractIterator.java:52)
> 2020-05-20T16:36:17.1037553Z  at 
> org.apache.flink.shaded.guava18.com.google.common.collect.Iterators.getOnlyElement(Iterators.java:302)
> 2020-05-20T16:36:17.1038087Z  at 
> org.apache.flink.shaded.guava18.com.google.common.collect.Iterables.getOnlyElement(Iterables.java:289)
> 2020-05-20T16:36:17.1038654Z  at 
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducerITCase.assertRecord(FlinkKafkaProducerITCase.java:201)
> 2020-05-20T16:36:17.1039239Z  at 
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducerITCase.testResumeTransaction(FlinkKafkaProducerITCase.java:108)
> 2020-05-20T16:36:17.1039727Z  at 
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> 2020-05-20T16:36:17.1040131Z  at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> 2020-05-20T16:36:17.1040575Z  at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> 2020-05-20T16:36:17.1040989Z  at 
> java.lang.reflect.Method.invoke(Method.java:498)
> 2020-05-20T16:36:17.1041515Z  at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
> 2020-05-20T16:36:17.1041989Z  at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> 2020-05-20T16:36:17.1042468Z  at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
> 2020-05-20T16:36:17.1042908Z  at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> 2020-05-20T16:36:17.1043497Z  at 
> org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:298)
> 2020-05-20T16:36:17.1044096Z  at 
> org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:292)
> 2020-05-20T16:36:17.1044833Z  at 
> java.util.concurrent.FutureTask.run(FutureTask.java:266)
> 2020-05-20T16:36:17.1045171Z  at java.lang.Thread.run(Thread.java:748)
> 2020-05-20T16:36:17.1045358Z 
> 2020-05-20T16:36:17.6145018Z [INFO] 
> 2020-05-20T16:36:17.6145477Z [INFO] Results:
> 2020-05-20T16:36:17.6145654Z [INFO] 
> 2020-05-20T16:36:17.6145838Z [ERROR] Errors: 
> 2020-05-20T16:36:17.6146898Z [ERROR]   
> FlinkKafkaProducerITCase.testResumeTransaction:108->assertRecord:201 » 
> NoSuchElement
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-18037) The doc of StreamTaskNetworkInput.java may has a redundant 'status'

2020-06-02 Thread Jark Wu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18037?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17124620#comment-17124620
 ] 

Jark Wu commented on FLINK-18037:
-

I think so. 

> The doc of StreamTaskNetworkInput.java may has a redundant 'status'
> ---
>
> Key: FLINK-18037
> URL: https://issues.apache.org/jira/browse/FLINK-18037
> Project: Flink
>  Issue Type: Bug
>  Components: Documentation
>Affects Versions: 1.10.0, 1.10.1, 1.11.0
>Reporter: ZhuShang
>Priority: Major
>   Original Estimate: 1h
>  Remaining Estimate: 1h
>
> The doc of class StreamTaskNetworkInput as follows 
> {code:java}
> * Forwarding elements, watermarks, or status status elements must be 
> protected by synchronizing
> * on the given lock object. This ensures that we don't call methods on a
> * {@link StreamInputProcessor} concurrently with the timer callback or other 
> things.{code}
>  
>  Is one of the 'status' redundant?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-17260) StreamingKafkaITCase failure on Azure

2020-06-02 Thread Robert Metzger (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-17260?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17124621#comment-17124621
 ] 

Robert Metzger commented on FLINK-17260:


https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=2586=logs=08866332-78f7-59e4-4f7e-49a56faa3179=931b3127-d6ee-5f94-e204-48d51cd1c334

> StreamingKafkaITCase failure on Azure
> -
>
> Key: FLINK-17260
> URL: https://issues.apache.org/jira/browse/FLINK-17260
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kafka, Tests
>Affects Versions: 1.11.0, 1.12.0
>Reporter: Roman Khachatryan
>Priority: Blocker
>  Labels: test-stability
> Fix For: 1.11.0
>
>
> [https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_apis/build/builds/7544/logs/165]
>  
> {code:java}
> 2020-04-16T00:12:32.2848429Z [INFO] Running 
> org.apache.flink.tests.util.kafka.StreamingKafkaITCase
> 2020-04-16T00:14:47.9100927Z [ERROR] Tests run: 3, Failures: 1, Errors: 0, 
> Skipped: 0, Time elapsed: 135.621 s <<< FAILURE! - in 
> org.apache.flink.tests.util.k afka.StreamingKafkaITCase
> 2020-04-16T00:14:47.9103036Z [ERROR] testKafka[0: 
> kafka-version:0.10.2.0](org.apache.flink.tests.util.kafka.StreamingKafkaITCase)
>   Time elapsed: 46.222 s  <<<  FAILURE!
> 2020-04-16T00:14:47.9104033Z java.lang.AssertionError: 
> expected:<[elephant,27,64213]> but was:<[]>
> 2020-04-16T00:14:47.9104638Zat org.junit.Assert.fail(Assert.java:88)
> 2020-04-16T00:14:47.9105148Zat 
> org.junit.Assert.failNotEquals(Assert.java:834)
> 2020-04-16T00:14:47.9105701Zat 
> org.junit.Assert.assertEquals(Assert.java:118)
> 2020-04-16T00:14:47.9106239Zat 
> org.junit.Assert.assertEquals(Assert.java:144)
> 2020-04-16T00:14:47.9107177Zat 
> org.apache.flink.tests.util.kafka.StreamingKafkaITCase.testKafka(StreamingKafkaITCase.java:162)
> 2020-04-16T00:14:47.9107845Zat 
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> 2020-04-16T00:14:47.9108434Zat 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> 2020-04-16T00:14:47.9109318Zat 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> 2020-04-16T00:14:47.9109914Zat 
> java.lang.reflect.Method.invoke(Method.java:498)
> 2020-04-16T00:14:47.9110434Zat 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
> 2020-04-16T00:14:47.9110985Zat 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> 2020-04-16T00:14:47.9111548Zat 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
> 2020-04-16T00:14:47.9112083Zat 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> 2020-04-16T00:14:47.9112629Zat 
> org.apache.flink.util.ExternalResource$1.evaluate(ExternalResource.java:48)
> 2020-04-16T00:14:47.9113145Zat 
> org.apache.flink.util.ExternalResource$1.evaluate(ExternalResource.java:48)
> 2020-04-16T00:14:47.9113637Zat 
> org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55)
> 2020-04-16T00:14:47.9114072Zat 
> org.junit.rules.RunRules.evaluate(RunRules.java:20)
> 2020-04-16T00:14:47.9114490Zat 
> org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
> 2020-04-16T00:14:47.9115256Zat 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
> 2020-04-16T00:14:47.9115791Zat 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
> 2020-04-16T00:14:47.9116292Zat 
> org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> 2020-04-16T00:14:47.9116736Zat 
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> 2020-04-16T00:14:47.9117779Zat 
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> 2020-04-16T00:14:47.9118274Zat 
> org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> 2020-04-16T00:14:47.9118766Zat 
> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> 2020-04-16T00:14:47.9119204Zat 
> org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> 2020-04-16T00:14:47.9119625Zat 
> org.junit.runners.Suite.runChild(Suite.java:128)
> 2020-04-16T00:14:47.9120005Zat 
> org.junit.runners.Suite.runChild(Suite.java:27)
> 2020-04-16T00:14:47.9120428Zat 
> org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> 2020-04-16T00:14:47.9120876Zat 
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> 2020-04-16T00:14:47.9121350Zat 
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> 2020-04-16T00:14:47.9121805Zat 
> org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> 2020-04-16T00:14:47.9122273Zat 
> 

[jira] [Assigned] (FLINK-18037) The doc of StreamTaskNetworkInput.java may has a redundant 'status'

2020-06-02 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18037?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu reassigned FLINK-18037:
---

Assignee: ZhuShang

> The doc of StreamTaskNetworkInput.java may has a redundant 'status'
> ---
>
> Key: FLINK-18037
> URL: https://issues.apache.org/jira/browse/FLINK-18037
> Project: Flink
>  Issue Type: Bug
>  Components: Documentation
>Affects Versions: 1.10.0, 1.10.1, 1.11.0
>Reporter: ZhuShang
>Assignee: ZhuShang
>Priority: Major
>   Original Estimate: 1h
>  Remaining Estimate: 1h
>
> The doc of class StreamTaskNetworkInput as follows 
> {code:java}
> * Forwarding elements, watermarks, or status status elements must be 
> protected by synchronizing
> * on the given lock object. This ensures that we don't call methods on a
> * {@link StreamInputProcessor} concurrently with the timer callback or other 
> things.{code}
>  
>  Is one of the 'status' redundant?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-17824) "Resuming Savepoint" e2e stalls indefinitely

2020-06-02 Thread Robert Metzger (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-17824?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17124618#comment-17124618
 ] 

Robert Metzger commented on FLINK-17824:


Now with logs: 
https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=2586=logs=91bf6583-3fb2-592f-e4d4-d79d79c3230a=94459a52-42b6-5bfc-5d74-690b5d3c6de8
 (hadoop 3.1.3 profile, see FLINK-17404)

> "Resuming Savepoint" e2e stalls indefinitely 
> -
>
> Key: FLINK-17824
> URL: https://issues.apache.org/jira/browse/FLINK-17824
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing, Tests
>Affects Versions: 1.12.0
>Reporter: Robert Metzger
>Assignee: Robert Metzger
>Priority: Major
>  Labels: pull-request-available, test-stability
> Fix For: 1.12.0
>
>
> CI; 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=1887=logs=91bf6583-3fb2-592f-e4d4-d79d79c3230a=94459a52-42b6-5bfc-5d74-690b5d3c6de8
> {code}
> 2020-05-19T21:05:52.9696236Z 
> ==
> 2020-05-19T21:05:52.9696860Z Running 'Resuming Savepoint (file, async, scale 
> down) end-to-end test'
> 2020-05-19T21:05:52.9697243Z 
> ==
> 2020-05-19T21:05:52.9713094Z TEST_DATA_DIR: 
> /home/vsts/work/1/s/flink-end-to-end-tests/test-scripts/temp-test-directory-52970362751
> 2020-05-19T21:05:53.1194478Z Flink dist directory: 
> /home/vsts/work/1/s/flink-dist/target/flink-1.12-SNAPSHOT-bin/flink-1.12-SNAPSHOT
> 2020-05-19T21:05:53.2180375Z Starting cluster.
> 2020-05-19T21:05:53.9986167Z Starting standalonesession daemon on host 
> fv-az558.
> 2020-05-19T21:05:55.5997224Z Starting taskexecutor daemon on host fv-az558.
> 2020-05-19T21:05:55.6223837Z Waiting for Dispatcher REST endpoint to come 
> up...
> 2020-05-19T21:05:57.0552482Z Waiting for Dispatcher REST endpoint to come 
> up...
> 2020-05-19T21:05:57.9446865Z Waiting for Dispatcher REST endpoint to come 
> up...
> 2020-05-19T21:05:59.0098434Z Waiting for Dispatcher REST endpoint to come 
> up...
> 2020-05-19T21:06:00.0569710Z Dispatcher REST endpoint is up.
> 2020-05-19T21:06:07.7099937Z Job (a92a74de8446a80403798bb4806b73f3) is 
> running.
> 2020-05-19T21:06:07.7855906Z Waiting for job to process up to 200 records, 
> current progress: 114 records ...
> 2020-05-19T21:06:55.5755111Z 
> 2020-05-19T21:06:55.5756550Z 
> 
> 2020-05-19T21:06:55.5757225Z  The program finished with the following 
> exception:
> 2020-05-19T21:06:55.5757566Z 
> 2020-05-19T21:06:55.5765453Z org.apache.flink.util.FlinkException: Could not 
> stop with a savepoint job "a92a74de8446a80403798bb4806b73f3".
> 2020-05-19T21:06:55.5766873Z  at 
> org.apache.flink.client.cli.CliFrontend.lambda$stop$5(CliFrontend.java:485)
> 2020-05-19T21:06:55.5767980Z  at 
> org.apache.flink.client.cli.CliFrontend.runClusterAction(CliFrontend.java:854)
> 2020-05-19T21:06:55.5769014Z  at 
> org.apache.flink.client.cli.CliFrontend.stop(CliFrontend.java:477)
> 2020-05-19T21:06:55.5770052Z  at 
> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:921)
> 2020-05-19T21:06:55.5771107Z  at 
> org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:982)
> 2020-05-19T21:06:55.5772223Z  at 
> org.apache.flink.runtime.security.contexts.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:30)
> 2020-05-19T21:06:55.5773325Z  at 
> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:982)
> 2020-05-19T21:06:55.5774871Z Caused by: 
> java.util.concurrent.ExecutionException: 
> java.util.concurrent.CompletionException: 
> java.util.concurrent.CompletionException: 
> org.apache.flink.runtime.checkpoint.CheckpointException: Checkpoint 
> Coordinator is suspending.
> 2020-05-19T21:06:55.5777183Z  at 
> java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357)
> 2020-05-19T21:06:55.5778884Z  at 
> java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1928)
> 2020-05-19T21:06:55.5779920Z  at 
> org.apache.flink.client.cli.CliFrontend.lambda$stop$5(CliFrontend.java:483)
> 2020-05-19T21:06:55.5781175Z  ... 6 more
> 2020-05-19T21:06:55.5782391Z Caused by: 
> java.util.concurrent.CompletionException: 
> java.util.concurrent.CompletionException: 
> org.apache.flink.runtime.checkpoint.CheckpointException: Checkpoint 
> Coordinator is suspending.
> 2020-05-19T21:06:55.5783885Z  at 
> org.apache.flink.runtime.scheduler.SchedulerBase.lambda$stopWithSavepoint$9(SchedulerBase.java:890)
> 2020-05-19T21:06:55.5784992Z  at 
> java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:836)
> 2020-05-19T21:06:55.5786492Z  at 
> 

[GitHub] [flink] wuchong commented on a change in pull request #12436: [FLINK-17847][table sql / planner] ArrayIndexOutOfBoundsException happens in StreamExecCalc operator

2020-06-02 Thread GitBox


wuchong commented on a change in pull request #12436:
URL: https://github.com/apache/flink/pull/12436#discussion_r434317766



##
File path: 
flink-table/flink-table-planner-blink/src/main/scala/org/apache/flink/table/planner/codegen/calls/ScalarOperatorGens.scala
##
@@ -1643,19 +1644,32 @@ object ScalarOperatorGens {
 val resultTypeTerm = primitiveTypeTermForType(componentInfo)
 val defaultTerm = primitiveDefaultValue(componentInfo)
 
+if (index.literalValue.isDefined &&
+index.literalValue.get.isInstanceOf[Int] &&
+index.literalValue.get.asInstanceOf[Int] < 1) {
+  throw new ValidationException(s"Array element access needs an index 
starting at 1 but was " +

Review comment:
   Can be simplified to 
   
   ```scala
   index.literalValue match {
 case Some(v: Int) if v < 1 =>
   throw new ValidationException(
 s"Array element access needs an index starting at 1 but was $v")
 case _ => // nothing
   }
   ```

##
File path: 
flink-table/flink-table-planner-blink/src/main/scala/org/apache/flink/table/planner/codegen/calls/ScalarOperatorGens.scala
##
@@ -1643,19 +1644,32 @@ object ScalarOperatorGens {
 val resultTypeTerm = primitiveTypeTermForType(componentInfo)
 val defaultTerm = primitiveDefaultValue(componentInfo)
 
+if (index.literalValue.isDefined &&
+index.literalValue.get.isInstanceOf[Int] &&
+index.literalValue.get.asInstanceOf[Int] < 1) {
+  throw new ValidationException(s"Array element access needs an index 
starting at 1 but was " +
+s"${index.literalValue.get.asInstanceOf[Int]}.")
+}
 val idxStr = s"${index.resultTerm} - 1"
 val arrayIsNull = s"${array.resultTerm}.isNullAt($idxStr)"
 val arrayGet =
   rowFieldReadAccess(ctx, idxStr, array.resultTerm, componentInfo)
 
+/**
+ * Return null when array index out of bounds which follows Calcite's 
behaviour.
+ * @see org.apache.calcite.sql.fun.SqlStdOperatorTable

Review comment:
   `* @see [[org.apache.calcite.sql.fun.SqlStdOperatorTable.ITEM]]`
   
   Could you move this Javadoc to the above of method signature? Javadoc is 
usually not used in code. 

##
File path: 
flink-table/flink-table-planner-blink/src/main/scala/org/apache/flink/table/planner/codegen/calls/ScalarOperatorGens.scala
##
@@ -1643,19 +1644,35 @@ object ScalarOperatorGens {
 val resultTypeTerm = primitiveTypeTermForType(componentInfo)
 val defaultTerm = primitiveDefaultValue(componentInfo)
 
+if (index.literalValue.isDefined &&
+index.literalValue.get.isInstanceOf[Int] &&
+index.literalValue.get.asInstanceOf[Int] < 1) {
+  throw new ValidationException(s"Array element access needs an index 
starting at 1 but was " +
+s"${index.literalValue.get.asInstanceOf[Int]}.")
+}
 val idxStr = s"${index.resultTerm} - 1"
 val arrayIsNull = s"${array.resultTerm}.isNullAt($idxStr)"
 val arrayGet =
   rowFieldReadAccess(ctx, idxStr, array.resultTerm, componentInfo)
 
+/**
+ * Return null when array index out of bounds which follows Calcite's 
behaviour.
+ * @see org.apache.calcite.sql.fun.SqlStdOperatorTable
+ */
 val arrayAccessCode =
-  s"""
- |${array.code}
- |${index.code}
- |boolean $nullTerm = ${array.nullTerm} || ${index.nullTerm} || 
$arrayIsNull;
- |$resultTypeTerm $resultTerm = $nullTerm ? $defaultTerm : $arrayGet;
- |""".stripMargin
-
+s"""
+|${array.code}
+|${index.code}
+|$resultTypeTerm $resultTerm;
+|boolean $nullTerm;
+|if (${idxStr} < 0 || ${idxStr} >= ${array.resultTerm}.size()) {

Review comment:
   I got what's your concern @libenchao . I think we don't have divergency 
on the runtime null behavior, the only concern is the compile exception. But I 
still prefer the current way, because it's very likely a typo if `my_arr[0]` is 
used, so a hint exception is more useful than surprising result. 





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-15318) RocksDBWriteBatchPerformanceTest.benchMark fails on ppc64le

2020-06-02 Thread Robert Metzger (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15318?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17124616#comment-17124616
 ] 

Robert Metzger commented on FLINK-15318:


https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=2587=logs=f0ac5c25-1168-55a5-07ff-0e88223afed9=39a61cac-5c62-532f-d2c1-dea450a66708

> RocksDBWriteBatchPerformanceTest.benchMark fails on ppc64le
> ---
>
> Key: FLINK-15318
> URL: https://issues.apache.org/jira/browse/FLINK-15318
> Project: Flink
>  Issue Type: Bug
>  Components: Benchmarks, Runtime / State Backends
> Environment: arch: ppc64le
> os: rhel7.6, ubuntu 18.04
> jdk: 8, 11
> mvn: 3.3.9, 3.6.2
>Reporter: Siddhesh Ghadi
>Priority: Major
> Attachments: surefire-report.txt
>
>
> RocksDBWriteBatchPerformanceTest.benchMark fails due to TestTimedOut, however 
> when test-timeout is increased from 2s to 5s in 
> org/apache/flink/contrib/streaming/state/benchmark/RocksDBWriteBatchPerformanceTest.java:75,
>  it passes. Is this acceptable solution?
> Note: Tests are ran inside a container.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-18071) CoordinatorEventsExactlyOnceITCase.checkListContainsSequence fails on CI

2020-06-02 Thread Robert Metzger (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18071?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17124615#comment-17124615
 ] 

Robert Metzger commented on FLINK-18071:


https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=2587=logs=6bfdaf55-0c08-5e3f-a2d2-2a0285fd41cf=93018cfc-6498-565b-e034-4b68eb90fc80

> CoordinatorEventsExactlyOnceITCase.checkListContainsSequence fails on CI
> 
>
> Key: FLINK-18071
> URL: https://issues.apache.org/jira/browse/FLINK-18071
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination, Tests
>Affects Versions: 1.12.0
>Reporter: Robert Metzger
>Priority: Critical
>  Labels: test-stability
> Fix For: 1.12.0
>
>
> CI: 
> https://dev.azure.com/georgeryan1322/Flink/_build/results?buildId=330=logs=6e58d712-c5cc-52fb-0895-6ff7bd56c46b=f30a8e80-b2cf-535c-9952-7f521a4ae374
> {code}
> [ERROR] Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 8.795 
> s <<< FAILURE! - in 
> org.apache.flink.runtime.operators.coordination.CoordinatorEventsExactlyOnceITCase
> [ERROR] 
> test(org.apache.flink.runtime.operators.coordination.CoordinatorEventsExactlyOnceITCase)
>   Time elapsed: 4.647 s  <<< FAILURE!
> java.lang.AssertionError: List did not contain expected sequence of 200 
> elements, but was: [152, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 
> 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 
> 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 
> 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 
> 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 
> 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 
> 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 
> 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 
> 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 
> 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 
> 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 
> 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 
> 198, 199]
>   at org.junit.Assert.fail(Assert.java:88)
>   at 
> org.apache.flink.runtime.operators.coordination.CoordinatorEventsExactlyOnceITCase.failList(CoordinatorEventsExactlyOnceITCase.java:160)
>   at 
> org.apache.flink.runtime.operators.coordination.CoordinatorEventsExactlyOnceITCase.checkListContainsSequence(CoordinatorEventsExactlyOnceITCase.java:148)
>   at 
> org.apache.flink.runtime.operators.coordination.CoordinatorEventsExactlyOnceITCase.test(CoordinatorEventsExactlyOnceITCase.java:143)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #12144: [FLINK-17384][flink-dist] support read hbase conf dir from flink.conf and change HBaseConfiguration construction.

2020-06-02 Thread GitBox


flinkbot edited a comment on pull request #12144:
URL: https://github.com/apache/flink/pull/12144#issuecomment-628475376


   
   ## CI report:
   
   * d36b959fb16a91c15babda10dd884cbbdec58420 UNKNOWN
   * 84f420d9043272150872fe4a32c2eaad2400a750 Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2592)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-16694) Resuming Externalized Checkpoint end-to-end test failed on travis

2020-06-02 Thread Robert Metzger (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16694?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17124613#comment-17124613
 ] 

Robert Metzger commented on FLINK-16694:


It happened again, on the Hadoop 3.1.3 profile: 
https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=2587=logs=91bf6583-3fb2-592f-e4d4-d79d79c3230a=94459a52-42b6-5bfc-5d74-690b5d3c6de8

Most likely it's the same issue as FLINK-17404.

> Resuming Externalized Checkpoint end-to-end test failed on travis
> -
>
> Key: FLINK-16694
> URL: https://issues.apache.org/jira/browse/FLINK-16694
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Metrics, Tests
>Affects Versions: 1.9.2
>Reporter: Piotr Nowojski
>Assignee: Chesnay Schepler
>Priority: Critical
>  Labels: pull-request-available, test-stability
> Fix For: 1.11.0, 1.10.2, 1.9.4
>
>
> Running 'Resuming Externalized Checkpoint (rocks, incremental, scale down) 
> end-to-end test' failed on travis (release-1.9 branch) with the error:
> {code:java}
> The job exceeded the maximum log length, and has been terminated.
> {code}
> https://api.travis-ci.org/v3/job/664469537/log.txt
> https://travis-ci.org/github/apache/flink/builds/664469494
> This error is probably because of metrics logging and it's masking some other 
> underlying issue, potentially FLINK-16695 as they happened in the same build.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-18075) Kafka connector does not call open method of (de)serialization schema

2020-06-02 Thread Kurt Young (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18075?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kurt Young updated FLINK-18075:
---
Fix Version/s: 1.11.0

> Kafka connector does not call open method of (de)serialization schema
> -
>
> Key: FLINK-18075
> URL: https://issues.apache.org/jira/browse/FLINK-18075
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Kafka, Formats (JSON, Avro, Parquet, ORC, 
> SequenceFile)
>Affects Versions: 1.11.0, 1.12.0
>Reporter: Seth Wiesman
>Assignee: Dawid Wysakowicz
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>
> The Kafka consumer and producer do not  call the open methods of plain 
> (De)SerializationSchema interfaces. Only the Keyed and Kafka specific 
> interfaces. The updated SQL implementations such as 
> AvroRowDataSeriailzationSchema use these methods and so SQL queries using 
> avro and kafka will fail in a null pointer exception. 
> cc [~aljoscha]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-16572) CheckPubSubEmulatorTest is flaky on Azure

2020-06-02 Thread Robert Metzger (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16572?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17124611#comment-17124611
 ] 

Robert Metzger commented on FLINK-16572:


https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=2591=logs=c88eea3b-64a0-564d-0031-9fdcd7b8abee=1e2bbe5b-4657-50be-1f07-d84bfce5b1f5

> CheckPubSubEmulatorTest is flaky on Azure
> -
>
> Key: FLINK-16572
> URL: https://issues.apache.org/jira/browse/FLINK-16572
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Google Cloud PubSub, Tests
>Affects Versions: 1.11.0
>Reporter: Aljoscha Krettek
>Assignee: Richard Deurwaarder
>Priority: Critical
>  Labels: pull-request-available, test-stability
> Fix For: 1.11.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Log: 
> https://dev.azure.com/aljoschakrettek/Flink/_build/results?buildId=56=logs=1f3ed471-1849-5d3c-a34c-19792af4ad16=ce095137-3e3b-5f73-4b79-c42d3d5f8283=7842



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] godfreyhe commented on a change in pull request #12431: [FLINK-18055] [sql-client] Fix catalog/database does not exist in sql client

2020-06-02 Thread GitBox


godfreyhe commented on a change in pull request #12431:
URL: https://github.com/apache/flink/pull/12431#discussion_r434317689



##
File path: 
flink-table/flink-sql-client/src/test/java/org/apache/flink/table/client/cli/CliClientTest.java
##
@@ -166,6 +167,64 @@ public void write(int b) throws IOException {
}
}
 
+   @Test
+   public void testUseCatalog() throws Exception {
+   TestingExecutor executor = new TestingExecutorBuilder()
+   .setUseCatalogConsumer((ignored1, catalogName) 
-> {
+   if (!catalogName.equals("cat")) {
+   throw new 
SqlExecutionException("unexpected catalog name: " + catalogName);
+   }
+   })
+   .build();
+
+   InputStream inputStream = new ByteArrayInputStream("use catalog 
cat;\n".getBytes());
+   ByteArrayOutputStream outputStream = new 
ByteArrayOutputStream(256);
+   CliClient cliClient = null;
+   SessionContext sessionContext = new 
SessionContext("test-session", new Environment());
+   String sessionId = executor.openSession(sessionContext);
+
+   try (Terminal terminal = new DumbTerminal(inputStream, 
outputStream)) {

Review comment:
   `CliClient` does not inherit from `AutoCloseable`, so we can put 
`CliClient` into `try-with-resource` block





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-18068) Job scheduling stops but not exits after throwing non-fatal exception

2020-06-02 Thread Jiayi Liao (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18068?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17124610#comment-17124610
 ] 

Jiayi Liao commented on FLINK-18068:


[~azagrebin] Sorry for not checking the JM logs after the period of 
{{batchSlotTimeout}}. Your thoughts about the timeout are right, the scheduling 
will be triggered again after the period of {{batchSlotTimeout}}.

And yes, I agree with you and [~trohrmann], this should be treated as fatal 
error, but I've found other places like {{requestYarnContainerIfRequired}} in 
{{onContainersCompleted}}, which have the similar problem. I think we should 
fix this in a general way like:


{code:java}
public class FatalErrorRunnable implements Runnable {

private final Runnable runnable;

FatalErrorRunnable(Runnable runnable) {
this.runnable = runnable;
}

@Override
public void run() {
try {
runnable.run();
} catch (Throwable e) {
onFatalError(e);
}
}
}
{code}


> Job scheduling stops but not exits after throwing non-fatal exception
> -
>
> Key: FLINK-18068
> URL: https://issues.apache.org/jira/browse/FLINK-18068
> Project: Flink
>  Issue Type: Improvement
>  Components: Deployment / YARN
>Affects Versions: 1.10.1
>Reporter: Jiayi Liao
>Priority: Major
>
> The batch job will stop but still be alive with doing nothing for a long time 
> (maybe forever?) if any non fatal exception is thrown from interacting with 
> YARN. Here is the example :
> {code:java}
> java.lang.IllegalStateException: The RMClient's and YarnResourceManagers 
> internal state about the number of pending container requests for resource 
>  has diverged. Number client's pending container 
> requests 40 != Number RM's pending container requests 0.
> at 
> org.apache.flink.util.Preconditions.checkState(Preconditions.java:217) 
> ~[flink-dist_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]
> at 
> org.apache.flink.yarn.YarnResourceManager.getPendingRequestsAndCheckConsistency(YarnResourceManager.java:518)
>  ~[flink-dist_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]
> at 
> org.apache.flink.yarn.YarnResourceManager.onContainersOfResourceAllocated(YarnResourceManager.java:431)
>  ~[flink-dist_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]
> at 
> org.apache.flink.yarn.YarnResourceManager.lambda$onContainersAllocated$1(YarnResourceManager.java:395)
>  ~[flink-dist_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]
> at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRunAsync(AkkaRpcActor.java:402)
>  ~[flink-dist_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]
> at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:195)
>  ~[flink-dist_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]
> at 
> org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:74)
>  ~[flink-dist_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]
> at 
> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(AkkaRpcActor.java:152)
>  ~[flink-dist_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]
> at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:26) 
> [flink-dist_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]
> at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:21) 
> [flink-dist_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]
> at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:123) 
> [flink-ljy-1.0.jar:?]
> at 
> akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:21) 
> [flink-dist_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]
> at 
> scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:170) 
> [flink-ljy-1.0.jar:?]
> at 
> scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171) 
> [flink-ljy-1.0.jar:?]
> at 
> scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171) 
> [flink-ljy-1.0.jar:?]
> at akka.actor.Actor$class.aroundReceive(Actor.scala:517) 
> [flink-dist_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]
> at akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:225) 
> [flink-dist_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]
> at akka.actor.ActorCell.receiveMessage(ActorCell.scala:592) 
> [flink-dist_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]
> at akka.actor.ActorCell.invoke(ActorCell.scala:561) 
> [flink-dist_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]
> at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:258) 
> [flink-dist_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]
> at akka.dispatch.Mailbox.run(Mailbox.scala:225) 
> [flink-dist_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]
> at akka.dispatch.Mailbox.exec(Mailbox.scala:235) 
> 

[jira] [Updated] (FLINK-17491) Translate Training page on project website

2020-06-02 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17491?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-17491:
---
Labels: pull-request-available  (was: )

> Translate Training page on project website
> --
>
> Key: FLINK-17491
> URL: https://issues.apache.org/jira/browse/FLINK-17491
> Project: Flink
>  Issue Type: Improvement
>  Components: chinese-translation, Project Website
>Reporter: David Anderson
>Assignee: Li Ying
>Priority: Major
>  Labels: pull-request-available
>
> Translate the training page for the project website to Chinese. The file is 
> training.zh.md.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-11909) Provide default failure/timeout/backoff handling strategy for AsyncIO functions

2020-06-02 Thread alex zhu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-11909?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17124607#comment-17124607
 ] 

alex zhu commented on FLINK-11909:
--

I believe most flink users (such as me) are focusing on business or data 
handling, not a native java programmer。So introducing a built-in retry 
mechanism would be very great usability improvement。Retry for each record is 
enough, it can solve "service/database restarting" such common scenarios。

> Provide default failure/timeout/backoff handling strategy for AsyncIO 
> functions
> ---
>
> Key: FLINK-11909
> URL: https://issues.apache.org/jira/browse/FLINK-11909
> Project: Flink
>  Issue Type: Improvement
>  Components: API / DataStream
>Reporter: Rong Rong
>Assignee: Rong Rong
>Priority: Major
>
> Currently Flink AsyncIO by default fails the entire job when async function 
> invoke fails [1]. It would be nice to have some default Async IO 
> failure/timeout handling strategy, or opens up some APIs for AsyncFunction 
> timeout method to interact with the AsyncWaitOperator. For example (quote 
> [~suez1224] in [2]):
> * FAIL_OPERATOR (default & current behavior)
> * FIX_INTERVAL_RETRY (retry with configurable fixed interval up to N times)
> * EXP_BACKOFF_RETRY (retry with exponential backoff up to N times)
> Discussion also extended to introduce configuration such as: 
> * MAX_RETRY_COUNT
> * RETRY_FAILURE_POLICY
> REF:
> [1] 
> https://ci.apache.org/projects/flink/flink-docs-release-1.7/dev/stream/operators/asyncio.html#timeout-handling
> [2] 
> http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/Backoff-strategies-for-async-IO-functions-tt26580.html



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] klion26 commented on a change in pull request #12313: [FLINK-17005][docs] Translate the CREATE TABLE ... LIKE syntax documentation to Chinese

2020-06-02 Thread GitBox


klion26 commented on a change in pull request #12313:
URL: https://github.com/apache/flink/pull/12313#discussion_r434312012



##
File path: docs/dev/table/sql/create.zh.md
##
@@ -249,59 +257,60 @@ CREATE TABLE Orders_with_watermark (
 );
 {% endhighlight %}
 
-The merging logic of table features can be controlled with `like options`.
+表属性的合并逻辑可以用 `like options` 来控制。
 
-You can control the merging behavior of:
+可以控制合并的表属性如下:
 
-* CONSTRAINTS - constraints such as primary and unique keys
-* GENERATED - computed columns
-* OPTIONS - connector options that describe connector and format properties
-* PARTITIONS - partition of the tables
-* WATERMARKS - watermark declarations
+* CONSTRAINTS - 主键和唯一键约束
+* GENERATED - 计算列
+* OPTIONS - 连接器信息、格式化方式等配置项
+* PARTITIONS - 表分区信息
+* WATERMARKS - watermark 定义
 
-with three different merging strategies:
+并且有三种不同的表属性合并策略:
 
-* INCLUDING - Includes the feature of the source table, fails on duplicate 
entries, e.g. if an option with the same key exists in both tables.
-* EXCLUDING - Does not include the given feature of the source table.
-* OVERWRITING - Includes the feature of the source table, overwrites duplicate 
entries of the source table with properties of the new table, e.g. if an option 
with the same key exists in both tables, the one from the current statement 
will be used.
+* INCLUDING - 新表包含源表(source table)所有的表属性,如果和源表的表属性重复则会直接失败,例如新表和源表存在相同 key 的属性。
+* EXCLUDING - 新表不包含源表指定的任何表属性。
+* OVERWRITING - 新表包含源表的表属性,但如果出现重复项,则会用新表的表属性覆盖源表中的重复表属性,例如,两个表中都存在相同 key 
的属性,则会使用当前语句中定义的 key 的属性值。
 
-Additionally, you can use the `INCLUDING/EXCLUDING ALL` option to specify what 
should be the strategy if there was no specific strategy defined, i.e. if you 
use `EXCLUDING ALL INCLUDING WATERMARKS` only the watermarks will be included 
from the source table.
+并且你可以使用 `INCLUDING/EXCLUDING ALL` 这种声明方式来指定使用怎样的合并策略,例如使用 `EXCLUDING ALL 
INCLUDING WATERMARKS`,那么代表只有源表的 WATERMARKS 属性才会被包含进新表。
 
-Example:
+示例如下:
 {% highlight sql %}
--- A source table stored in a filesystem
+
+-- 存储在文件系统的源表
 CREATE TABLE Orders_in_file (
 user BIGINT,
 product STRING,
 order_time_string STRING,
 order_time AS to_timestamp(order_time)
-
+

Review comment:
   这个无关的修改可以去掉





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink-web] liying919 opened a new pull request #345: [FLINK-17491] Translate Training page on project website

2020-06-02 Thread GitBox


liying919 opened a new pull request #345:
URL: https://github.com/apache/flink-web/pull/345


   Translate Training page on project website.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink-web] klion26 commented on pull request #342: [FLINK-17926] Fix the build problem of docker image

2020-06-02 Thread GitBox


klion26 commented on pull request #342:
URL: https://github.com/apache/flink-web/pull/342#issuecomment-637952600


   Ahh. maybe there is someting not compatible with `ruby:2.5` docker image, I 
tried to execute the command `docker run --rm --volume="$PWD:/srv/flink-web" 
--expose=4000 -p 4000:4000 -it ruby:2.5 bash -c 'cd /srv/flink-web && 
./build.sh -p'` on my laptop, it failed with the following message, I'll try to 
find out the reason here.
   ```
   Configuration file: /srv/flink-web/_config.yml
   bundler: failed to load command: jekyll 
(/srv/flink-web/.rubydeps/ruby/2.5.0/bin/jekyll)
   LoadError: libruby-2.5.so.2.5: cannot open shared object file: No such file 
or directory - 
/srv/flink-web/.rubydeps/ruby/2.5.0/gems/posix-spawn-0.3.14/lib/posix_spawn_ext.so
 
/srv/flink-web/.rubydeps/ruby/2.5.0/gems/posix-spawn-0.3.14/lib/posix/spawn.rb:2:in
 `require'
 
/srv/flink-web/.rubydeps/ruby/2.5.0/gems/posix-spawn-0.3.14/lib/posix/spawn.rb:2:in
 `'
 
/srv/flink-web/.rubydeps/ruby/2.5.0/gems/pygments.rb-0.6.3/lib/pygments/popen.rb:2:in
 `require'
 
/srv/flink-web/.rubydeps/ruby/2.5.0/gems/pygments.rb-0.6.3/lib/pygments/popen.rb:2:in
 `'
 
/srv/flink-web/.rubydeps/ruby/2.5.0/gems/pygments.rb-0.6.3/lib/pygments.rb:1:in 
`require'
 
/srv/flink-web/.rubydeps/ruby/2.5.0/gems/pygments.rb-0.6.3/lib/pygments.rb:1:in 
`'
 /srv/flink-web/_plugins/highlightCode.rb:22:in `require'
 /srv/flink-web/_plugins/highlightCode.rb:22:in `'
 
/srv/flink-web/.rubydeps/ruby/2.5.0/gems/jekyll-3.0.5/lib/jekyll/plugin_manager.rb:75:in
 `require'
 
/srv/flink-web/.rubydeps/ruby/2.5.0/gems/jekyll-3.0.5/lib/jekyll/plugin_manager.rb:75:in
 `block (2 levels) in require_plugin_files'
 
/srv/flink-web/.rubydeps/ruby/2.5.0/gems/jekyll-3.0.5/lib/jekyll/plugin_manager.rb:74:in
 `each'
 
/srv/flink-web/.rubydeps/ruby/2.5.0/gems/jekyll-3.0.5/lib/jekyll/plugin_manager.rb:74:in
 `block in require_plugin_files'
 
/srv/flink-web/.rubydeps/ruby/2.5.0/gems/jekyll-3.0.5/lib/jekyll/plugin_manager.rb:73:in
 `each'
 
/srv/flink-web/.rubydeps/ruby/2.5.0/gems/jekyll-3.0.5/lib/jekyll/plugin_manager.rb:73:in
 `require_plugin_files'
 
/srv/flink-web/.rubydeps/ruby/2.5.0/gems/jekyll-3.0.5/lib/jekyll/plugin_manager.rb:18:in
 `conscientious_require'
 
/srv/flink-web/.rubydeps/ruby/2.5.0/gems/jekyll-3.0.5/lib/jekyll/site.rb:97:in 
`setup'
 
/srv/flink-web/.rubydeps/ruby/2.5.0/gems/jekyll-3.0.5/lib/jekyll/site.rb:49:in 
`initialize'
 
/srv/flink-web/.rubydeps/ruby/2.5.0/gems/jekyll-3.0.5/lib/jekyll/commands/build.rb:30:in
 `new'
 
/srv/flink-web/.rubydeps/ruby/2.5.0/gems/jekyll-3.0.5/lib/jekyll/commands/build.rb:30:in
 `process'
 
/srv/flink-web/.rubydeps/ruby/2.5.0/gems/jekyll-3.0.5/lib/jekyll/commands/serve.rb:26:in
 `block (2 levels) in init_with_program'
 
/srv/flink-web/.rubydeps/ruby/2.5.0/gems/mercenary-0.3.6/lib/mercenary/command.rb:220:in
 `block in execute'
 
/srv/flink-web/.rubydeps/ruby/2.5.0/gems/mercenary-0.3.6/lib/mercenary/command.rb:220:in
 `each'
 
/srv/flink-web/.rubydeps/ruby/2.5.0/gems/mercenary-0.3.6/lib/mercenary/command.rb:220:in
 `execute'
 
/srv/flink-web/.rubydeps/ruby/2.5.0/gems/mercenary-0.3.6/lib/mercenary/program.rb:42:in
 `go'
 
/srv/flink-web/.rubydeps/ruby/2.5.0/gems/mercenary-0.3.6/lib/mercenary.rb:19:in 
`program'
 /srv/flink-web/.rubydeps/ruby/2.5.0/gems/jekyll-3.0.5/bin/jekyll:17:in 
`'
 /srv/flink-web/.rubydeps/ruby/2.5.0/bin/jekyll:23:in `load'
 /srv/flink-web/.rubydeps/ruby/2.5.0/bin/jekyll:23:in `'
   ```
   @rmetzger can `./docker/run.sh` success on your side?



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12453: [FLINK-18061] [table] TableResult#collect method should return a closeable iterator to avoid resource leak

2020-06-02 Thread GitBox


flinkbot edited a comment on pull request #12453:
URL: https://github.com/apache/flink/pull/12453#issuecomment-637941328


   
   ## CI report:
   
   * c2f3319900afad80a0f9fac5a8c9aa4cfdadd4ca Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2600)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] klion26 commented on a change in pull request #12237: [FLINK-17290] [chinese-translation, Documentation / Training] Transla…

2020-06-02 Thread GitBox


klion26 commented on a change in pull request #12237:
URL: https://github.com/apache/flink/pull/12237#discussion_r433926976



##
File path: docs/training/streaming_analytics.zh.md
##
@@ -29,123 +29,104 @@ under the License.
 
 ## Event Time and Watermarks
 
-### Introduction
+
+### 概要
 
-Flink explicitly supports three different notions of time:
+Flink 明确支持以下三种时间语义:
 
-* _event time:_ the time when an event occurred, as recorded by the device 
producing (or storing) the event
+* _事件时间(event time):_ 事件产生的时间,记录的是设备生产(或者存储)事件的时间
 
-* _ingestion time:_ a timestamp recorded by Flink at the moment it ingests the 
event
+* _摄取时间(ingestion time):_ Flink 读取事件时记录的时间
 
-* _processing time:_ the time when a specific operator in your pipeline is 
processing the event
+* _处理时间(processing time):_ Flink pipeline 中具体算子处理事件的时间
 
-For reproducible results, e.g., when computing the maximum price a stock 
reached during the first
-hour of trading on a given day, you should use event time. In this way the 
result won't depend on
-when the calculation is performed. This kind of real-time application is 
sometimes performed using
-processing time, but then the results are determined by the events that happen 
to be processed
-during that hour, rather than the events that occurred then. Computing 
analytics based on processing
-time causes inconsistencies, and makes it difficult to re-analyze historic 
data or test new
-implementations.
+为了获得可重现的结果,例如在计算过去的特定一天里第一个小时股票的最高价格时,我们应该使用事件时间。这样的话,无论
+什么时间去计算都不会影响输出结果。然而有些人,在实时计算应用中使用处理时间,这样的话,输出结果就会被处理时间点所决

Review comment:
   这里 44 行需要和 43 行合并,否则 “无论” 和 “什么” 之间会有空格
   45 和 44 也需要合并,否则 “所决” 和 “定” 之间有空格
   
   `This kind of real-time application is sometimes performed using processing 
time, but then the results are determined by the events that happen to be 
processed during that hour, rather than the events that occurred then.` 
这句话的翻译直接翻译即可,比如 `然后使用 processing time 
的那些实时应用程序的处理结果,则由那一小时中所处理的数据所决定`(不一定是这个翻译,可以用你自己的语句描述)

##
File path: docs/training/streaming_analytics.zh.md
##
@@ -29,123 +29,104 @@ under the License.
 
 ## Event Time and Watermarks
 
-### Introduction
+
+### 概要
 
-Flink explicitly supports three different notions of time:
+Flink 明确支持以下三种时间语义:
 
-* _event time:_ the time when an event occurred, as recorded by the device 
producing (or storing) the event
+* _事件时间(event time):_ 事件产生的时间,记录的是设备生产(或者存储)事件的时间
 
-* _ingestion time:_ a timestamp recorded by Flink at the moment it ingests the 
event
+* _摄取时间(ingestion time):_ Flink 读取事件时记录的时间
 
-* _processing time:_ the time when a specific operator in your pipeline is 
processing the event
+* _处理时间(processing time):_ Flink pipeline 中具体算子处理事件的时间
 
-For reproducible results, e.g., when computing the maximum price a stock 
reached during the first
-hour of trading on a given day, you should use event time. In this way the 
result won't depend on
-when the calculation is performed. This kind of real-time application is 
sometimes performed using
-processing time, but then the results are determined by the events that happen 
to be processed
-during that hour, rather than the events that occurred then. Computing 
analytics based on processing
-time causes inconsistencies, and makes it difficult to re-analyze historic 
data or test new
-implementations.
+为了获得可重现的结果,例如在计算过去的特定一天里第一个小时股票的最高价格时,我们应该使用事件时间。这样的话,无论
+什么时间去计算都不会影响输出结果。然而有些人,在实时计算应用中使用处理时间,这样的话,输出结果就会被处理时间点所决
+定,而不是生产事件的时间。基于处理时间会导致多次计算的结果不一致,也可能会导致再次分析历史数据或者测试新代码变得异常困难。
 
-### Working with Event Time
+
+### 使用 Event Time
 
-By default, Flink will use processing time. To change this, you can set the 
Time Characteristic:
+Flink 在默认情况下是使用处理时间。也可以通过下面配置来告诉 Flink 选择哪种时间语义:
 
 {% highlight java %}
 final StreamExecutionEnvironment env =
 StreamExecutionEnvironment.getExecutionEnvironment();
 env.setStreamTimeCharacteristic(TimeCharacteristic.EventTime);
 {% endhighlight %}
 
-If you want to use event time, you will also need to supply a Timestamp 
Extractor and Watermark
-Generator that Flink will use to track the progress of event time. This will 
be covered in the
-section below on [Working with Watermarks]({% link
-training/streaming_analytics.zh.md %}#working-with-watermarks), but first we 
should explain what
-watermarks are.
+如果想要使用事件时间,需要额外给 Flink 提供一个时间戳的提取器和 Watermark 生成器,Flink 将使用它们来跟踪事件时间的进度。这
+将在选节[使用 Watermarks]({% link training/streaming_analytics.zh.md 
%}#Working-with-Watermarks)中介绍,但是首先我们需要解释一下
+ watermarks 是什么。
 
 ### Watermarks
 
-Let's work through a simple example that will show why watermarks are needed, 
and how they work.
+让我们通过一个简单的示例来演示为什么需要 watermarks 及其工作方式。
 
-In this example you have a stream of timestamped events that arrive somewhat 
out of order, as shown
-below. The numbers shown are timestamps that indicate when these events 
actually occurred. The first
-event to arrive happened at time 4, and it is followed by an event that 
happened earlier, at time 2,
-and so on:

[jira] [Updated] (FLINK-17686) Add document to dataGen, print, blackhole connectors

2020-06-02 Thread Danny Chen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17686?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Chen updated FLINK-17686:
---
Priority: Critical  (was: Major)

> Add document to dataGen, print, blackhole connectors
> 
>
> Key: FLINK-17686
> URL: https://issues.apache.org/jira/browse/FLINK-17686
> Project: Flink
>  Issue Type: Sub-task
>  Components: Documentation
>Affects Versions: 1.11.0
>Reporter: Jingsong Lee
>Assignee: Shengkai Fang
>Priority: Critical
> Fix For: 1.11.0
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-17831) Add documentation for the new Kafka connector

2020-06-02 Thread Danny Chen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17831?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Chen updated FLINK-17831:
---
Priority: Critical  (was: Major)

> Add documentation for the new Kafka connector
> -
>
> Key: FLINK-17831
> URL: https://issues.apache.org/jira/browse/FLINK-17831
> Project: Flink
>  Issue Type: Sub-task
>  Components: Documentation
>Reporter: Jark Wu
>Assignee: Danny Chen
>Priority: Critical
> Fix For: 1.11.0
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-17776) Add documentation for DDL in hive dialect

2020-06-02 Thread Danny Chen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17776?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Chen updated FLINK-17776:
---
Priority: Critical  (was: Major)

> Add documentation for DDL in hive dialect
> -
>
> Key: FLINK-17776
> URL: https://issues.apache.org/jira/browse/FLINK-17776
> Project: Flink
>  Issue Type: Sub-task
>  Components: Documentation
>Reporter: Jingsong Lee
>Assignee: Rui Li
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot commented on pull request #12453: [FLINK-18061] [table] TableResult#collect method should return a closeable iterator to avoid resource leak

2020-06-02 Thread GitBox


flinkbot commented on pull request #12453:
URL: https://github.com/apache/flink/pull/12453#issuecomment-637941328


   
   ## CI report:
   
   * c2f3319900afad80a0f9fac5a8c9aa4cfdadd4ca UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12422: [FLINK-17901][python] Add module interface in PyFlink

2020-06-02 Thread GitBox


flinkbot edited a comment on pull request #12422:
URL: https://github.com/apache/flink/pull/12422#issuecomment-636647115


   
   ## CI report:
   
   * 3a387da20f29d9b1f6baf1e57c6d9979d27dd81f Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2509)
 
   * a57beb047e52bacd60d9c19cbcd31bc9b1d88ec4 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2599)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] libenchao commented on a change in pull request #12436: [FLINK-17847][table sql / planner] ArrayIndexOutOfBoundsException happens in StreamExecCalc operator

2020-06-02 Thread GitBox


libenchao commented on a change in pull request #12436:
URL: https://github.com/apache/flink/pull/12436#discussion_r434297494



##
File path: 
flink-table/flink-table-planner-blink/src/main/scala/org/apache/flink/table/planner/codegen/calls/ScalarOperatorGens.scala
##
@@ -1643,19 +1644,35 @@ object ScalarOperatorGens {
 val resultTypeTerm = primitiveTypeTermForType(componentInfo)
 val defaultTerm = primitiveDefaultValue(componentInfo)
 
+if (index.literalValue.isDefined &&
+index.literalValue.get.isInstanceOf[Int] &&
+index.literalValue.get.asInstanceOf[Int] < 1) {
+  throw new ValidationException(s"Array element access needs an index 
starting at 1 but was " +
+s"${index.literalValue.get.asInstanceOf[Int]}.")
+}
 val idxStr = s"${index.resultTerm} - 1"
 val arrayIsNull = s"${array.resultTerm}.isNullAt($idxStr)"
 val arrayGet =
   rowFieldReadAccess(ctx, idxStr, array.resultTerm, componentInfo)
 
+/**
+ * Return null when array index out of bounds which follows Calcite's 
behaviour.
+ * @see org.apache.calcite.sql.fun.SqlStdOperatorTable
+ */
 val arrayAccessCode =
-  s"""
- |${array.code}
- |${index.code}
- |boolean $nullTerm = ${array.nullTerm} || ${index.nullTerm} || 
$arrayIsNull;
- |$resultTypeTerm $resultTerm = $nullTerm ? $defaultTerm : $arrayGet;
- |""".stripMargin
-
+s"""
+|${array.code}
+|${index.code}
+|$resultTypeTerm $resultTerm;
+|boolean $nullTerm;
+|if (${idxStr} < 0 || ${idxStr} >= ${array.resultTerm}.size()) {

Review comment:
   What I want to express is, the index could be a literal or a field 
reference.
   And we want to throw Exception for index < 1, there are two cases that index 
< 1:
   - literal value, it has been checked in code gen, and throws 
ValidationException.
   - field reference, currently you just return null for this case. 





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-16975) Add docs for FileSystem connector

2020-06-02 Thread Danny Chen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16975?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Chen updated FLINK-16975:
---
Priority: Critical  (was: Major)

> Add docs for FileSystem connector
> -
>
> Key: FLINK-16975
> URL: https://issues.apache.org/jira/browse/FLINK-16975
> Project: Flink
>  Issue Type: Sub-task
>  Components: Documentation
>Affects Versions: 1.11.0
>Reporter: Leonard Xu
>Assignee: Jingsong Lee
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-17599) Update documents due to FLIP-84

2020-06-02 Thread Danny Chen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17599?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Chen updated FLINK-17599:
---
Priority: Critical  (was: Major)

> Update documents due to FLIP-84
> ---
>
> Key: FLINK-17599
> URL: https://issues.apache.org/jira/browse/FLINK-17599
> Project: Flink
>  Issue Type: Sub-task
>  Components: Documentation
>Reporter: Kurt Young
>Assignee: godfrey he
>Priority: Critical
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-17832) Add documentation for the new Elasticsearch connector

2020-06-02 Thread Danny Chen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17832?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Chen updated FLINK-17832:
---
Priority: Critical  (was: Major)

> Add documentation for the new Elasticsearch connector
> -
>
> Key: FLINK-17832
> URL: https://issues.apache.org/jira/browse/FLINK-17832
> Project: Flink
>  Issue Type: Sub-task
>  Components: Documentation
>Reporter: Jark Wu
>Assignee: Shengkai Fang
>Priority: Critical
> Fix For: 1.11.0
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-18065) Add documentation for new scalar/table functions

2020-06-02 Thread Danny Chen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18065?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Chen updated FLINK-18065:
---
Priority: Critical  (was: Major)

> Add documentation for new scalar/table functions
> 
>
> Key: FLINK-18065
> URL: https://issues.apache.org/jira/browse/FLINK-18065
> Project: Flink
>  Issue Type: Sub-task
>  Components: Documentation
>Reporter: Timo Walther
>Assignee: Timo Walther
>Priority: Critical
>
> Write documentation for scalar/table functions of FLIP-65.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-17829) Add documentation for the new JDBC connector

2020-06-02 Thread Danny Chen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17829?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Chen updated FLINK-17829:
---
Priority: Critical  (was: Major)

> Add documentation for the new JDBC connector
> 
>
> Key: FLINK-17829
> URL: https://issues.apache.org/jira/browse/FLINK-17829
> Project: Flink
>  Issue Type: Sub-task
>  Components: Documentation
>Reporter: Jark Wu
>Assignee: Leonard Xu
>Priority: Critical
> Fix For: 1.11.0
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-18066) Add documentation for how to develop a new table source/sink

2020-06-02 Thread Danny Chen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18066?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Chen updated FLINK-18066:
---
Priority: Critical  (was: Major)

> Add documentation for how to develop a new table source/sink
> 
>
> Key: FLINK-18066
> URL: https://issues.apache.org/jira/browse/FLINK-18066
> Project: Flink
>  Issue Type: Sub-task
>  Components: Documentation
>Reporter: Timo Walther
>Assignee: Timo Walther
>Priority: Critical
>
> Covers how to write a custom source/sink and format using FLIP-95 interfaces.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-17406) Add documentation about dynamic table options

2020-06-02 Thread Danny Chen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17406?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Chen updated FLINK-17406:
---
Priority: Critical  (was: Major)

> Add documentation about dynamic table options
> -
>
> Key: FLINK-17406
> URL: https://issues.apache.org/jira/browse/FLINK-17406
> Project: Flink
>  Issue Type: Sub-task
>  Components: Documentation
>Reporter: Kurt Young
>Assignee: Danny Chen
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-17830) Add documentation for the new HBase connector

2020-06-02 Thread Danny Chen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17830?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Chen updated FLINK-17830:
---
Priority: Critical  (was: Major)

> Add documentation for the new HBase connector
> -
>
> Key: FLINK-17830
> URL: https://issues.apache.org/jira/browse/FLINK-17830
> Project: Flink
>  Issue Type: Sub-task
>  Components: Documentation
>Reporter: Jark Wu
>Assignee: Jark Wu
>Priority: Critical
> Fix For: 1.11.0
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-17931) Document fromValues clause

2020-06-02 Thread Danny Chen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17931?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Chen updated FLINK-17931:
---
Priority: Critical  (was: Major)

> Document fromValues clause
> --
>
> Key: FLINK-17931
> URL: https://issues.apache.org/jira/browse/FLINK-17931
> Project: Flink
>  Issue Type: Sub-task
>  Components: Documentation, Table SQL / API
>Reporter: Dawid Wysakowicz
>Assignee: Dawid Wysakowicz
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-17834) Add document for Hive streaming source

2020-06-02 Thread Danny Chen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17834?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Chen updated FLINK-17834:
---
Priority: Critical  (was: Major)

> Add document for Hive streaming source
> --
>
> Key: FLINK-17834
> URL: https://issues.apache.org/jira/browse/FLINK-17834
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Danny Chen
>Assignee: Jingsong Lee
>Priority: Critical
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-17836) Add document for Hive dim join

2020-06-02 Thread Danny Chen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17836?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Chen updated FLINK-17836:
---
Priority: Critical  (was: Major)

> Add document for Hive dim join
> --
>
> Key: FLINK-17836
> URL: https://issues.apache.org/jira/browse/FLINK-17836
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Danny Chen
>Assignee: Rui Li
>Priority: Critical
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-17835) Add document for Hive streaming sink

2020-06-02 Thread Danny Chen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17835?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Chen updated FLINK-17835:
---
Priority: Critical  (was: Major)

> Add document for Hive streaming sink
> 
>
> Key: FLINK-17835
> URL: https://issues.apache.org/jira/browse/FLINK-17835
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Danny Chen
>Assignee: Jingsong Lee
>Priority: Critical
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] Jiayi-Liao edited a comment on pull request #12444: [FLINK-18067]Change default value of yarnMinAllocationMB from zero to DEFAULT_RM_S…

2020-06-02 Thread GitBox


Jiayi-Liao edited a comment on pull request #12444:
URL: https://github.com/apache/flink/pull/12444#issuecomment-637920335


   @wangyang0918  Thanks for your review. I've addressed your comments and also 
update the description(the CI failure is caused by checkstyle error which is 
already fixed in the latest commit). Please take another look when you can 
spare time.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] leonardBang commented on a change in pull request #12436: [FLINK-17847][table sql / planner] ArrayIndexOutOfBoundsException happens in StreamExecCalc operator

2020-06-02 Thread GitBox


leonardBang commented on a change in pull request #12436:
URL: https://github.com/apache/flink/pull/12436#discussion_r434295803



##
File path: 
flink-table/flink-table-planner-blink/src/main/scala/org/apache/flink/table/planner/codegen/calls/ScalarOperatorGens.scala
##
@@ -1643,19 +1644,35 @@ object ScalarOperatorGens {
 val resultTypeTerm = primitiveTypeTermForType(componentInfo)
 val defaultTerm = primitiveDefaultValue(componentInfo)
 
+if (index.literalValue.isDefined &&
+index.literalValue.get.isInstanceOf[Int] &&
+index.literalValue.get.asInstanceOf[Int] < 1) {
+  throw new ValidationException(s"Array element access needs an index 
starting at 1 but was " +
+s"${index.literalValue.get.asInstanceOf[Int]}.")
+}
 val idxStr = s"${index.resultTerm} - 1"
 val arrayIsNull = s"${array.resultTerm}.isNullAt($idxStr)"
 val arrayGet =
   rowFieldReadAccess(ctx, idxStr, array.resultTerm, componentInfo)
 
+/**
+ * Return null when array index out of bounds which follows Calcite's 
behaviour.
+ * @see org.apache.calcite.sql.fun.SqlStdOperatorTable
+ */
 val arrayAccessCode =
-  s"""
- |${array.code}
- |${index.code}
- |boolean $nullTerm = ${array.nullTerm} || ${index.nullTerm} || 
$arrayIsNull;
- |$resultTypeTerm $resultTerm = $nullTerm ? $defaultTerm : $arrayGet;
- |""".stripMargin
-
+s"""
+|${array.code}
+|${index.code}
+|$resultTypeTerm $resultTerm;
+|boolean $nullTerm;
+|if (${idxStr} < 0 || ${idxStr} >= ${array.resultTerm}.size()) {

Review comment:
   > consider this case:
   > 
   > ```sql
   > CREATE TABLE my_source (
   >   idx int,
   >   my_arr array
   > ) WITH (...);
   > 
   > SELECT my_arr[idx] FROM my_source;
   > ```
   For this case, idx has been an integer value rather than an input reference 
in runtime IIUC.
   





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-18049) The Flink kafka consumer job will be interrupted if the upstream kafka producer change the AVRO schema

2020-06-02 Thread Jark Wu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18049?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17124584#comment-17124584
 ] 

Jark Wu commented on FLINK-18049:
-

Thanks for the discussion and looking into this problem [~openinx]. If I 
understand correctly, what you want to fix is such case:

0. there is a Kafka Avro topic whose schema is [int a, string b].
1. there is a long running Flink SQL job which consumes Kafka Avro data using 
DDL {{CREATE TABLE xxx (a INT, b STRING) WITH (... format=avro)}}.
2. the producer of this Kafka Avro topic evolves their schema with 
backward-compatible by adding a column which default null. [int a, string b, 
double c default null]. In theory, the existing consumers of this topic should 
work fine, according to the [Avro schema 
rule|http://avro.apache.org/docs/current/spec.html#Schema+Resolution].
3. However, there is an exception thrown in the running Flink SQL job.

If this is the case (and I think the added test is verifying this case), I 
think it's reasonable to improve the current implementation. 

Besides, in another case, if we are using DDL  {{CREATE TABLE xxx (a INT, b 
STRING) WITH (... format=avro)}} to read existing Kafka topic with avro schema  
[int a, string b, double c default null], an error will happen too. So I think 
this is a bug. 


> The Flink kafka consumer job will be interrupted if the upstream kafka 
> producer change the AVRO schema
> --
>
> Key: FLINK-18049
> URL: https://issues.apache.org/jira/browse/FLINK-18049
> Project: Flink
>  Issue Type: Bug
>Reporter: Zheng Hu
>Priority: Critical
>  Labels: pull-request-available
>
> We have encountered a critical case from online services.  we have the data 
> pipeline:  (producer) -> (kafka) -> (flink consumer job), and all those 
> records are encoded in AVRO format.  Once the producer changed the AVRO 
> schema , says adding an extra column to the existing schema and writing few 
> data into the Kafka. 
> Then the downstream flink job crashed with the following stacktrace: 
> {code}
> ==WARNING==  allocating large 
> array--thread_id[0x7fccd9c16800]--thread_name[Source: Custom Source 
> (1/1)]--array_size[1590681120 bytes]--array_length[1590681103 elememts]
> os_prio=0 tid=0x7fccd9c16800 nid=0x226c0 runnable 
>   at org.shaded.apache.avro.util.Utf8.setByteLength(Utf8.java:78)
>   at 
> org.shaded.apache.avro.io.BinaryDecoder.readString(BinaryDecoder.java:261)
>   at 
> org.shaded.apache.avro.io.BinaryDecoder.readString(BinaryDecoder.java:272)
>   at 
> org.shaded.apache.avro.io.ResolvingDecoder.readString(ResolvingDecoder.java:214)
>   at 
> org.shaded.apache.avro.generic.GenericDatumReader.readString(GenericDatumReader.java:412)
>   at 
> org.shaded.apache.avro.generic.GenericDatumReader.readWithoutConversion(GenericDatumReader.java:181)
>   at 
> org.shaded.apache.avro.specific.SpecificDatumReader.readField(SpecificDatumReader.java:116)
>   at 
> org.shaded.apache.avro.generic.GenericDatumReader.readRecord(GenericDatumReader.java:222)
>   at 
> org.shaded.apache.avro.generic.GenericDatumReader.readWithoutConversion(GenericDatumReader.java:175)
>   at 
> org.shaded.apache.avro.generic.GenericDatumReader.read(GenericDatumReader.java:153)
>   at 
> org.shaded.apache.avro.generic.GenericDatumReader.read(GenericDatumReader.java:145)
>   at 
> org.apache.flink.formats.avro.AvroRowDeserializationSchema.deserialize(AvroRowDeserializationSchema.java:167)
>   at 
> org.apache.flink.formats.avro.AvroRowDeserializationSchema.deserialize(AvroRowDeserializationSchema.java:78)
>   at 
> org.apache.flink.streaming.util.serialization.KeyedDeserializationSchemaWrapper.deserialize(KeyedDeserializationSchemaWrapper.java:44)
>   at 
> org.apache.flink.streaming.connectors.kafka.internal.Kafka09Fetcher.runFetchLoop(Kafka09Fetcher.java:192)
>   at 
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumerBase.run(FlinkKafkaConsumerBase.java:771)
>   at 
> org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:120)
>   at 
> org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:74)
>   at 
> org.apache.flink.streaming.runtime.tasks.SourceStreamTask.run(SourceStreamTask.java:129)
>   at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:398)
>   at org.apache.flink.runtime.taskmanager.Task.run(Task.java:736)
>   at java.lang.Thread.run(Thread.java:834)
> {code} 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] dianfu commented on a change in pull request #12422: [FLINK-17901][python] Add module interface in PyFlink

2020-06-02 Thread GitBox


dianfu commented on a change in pull request #12422:
URL: https://github.com/apache/flink/pull/12422#discussion_r434294304



##
File path: flink-python/pyflink/table/module.py
##
@@ -0,0 +1,45 @@
+
+#  Licensed to the Apache Software Foundation (ASF) under one
+#  or more contributor license agreements.  See the NOTICE file
+#  distributed with this work for additional information
+#  regarding copyright ownership.  The ASF licenses this file
+#  to you under the Apache License, Version 2.0 (the
+#  "License"); you may not use this file except in compliance
+#  with the License.  You may obtain a copy of the License at
+#
+#  http://www.apache.org/licenses/LICENSE-2.0
+#
+#  Unless required by applicable law or agreed to in writing, software
+#  distributed under the License is distributed on an "AS IS" BASIS,
+#  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+#  See the License for the specific language governing permissions and
+# limitations under the License.
+
+from pyflink.java_gateway import get_gateway
+
+__all__ = ['Module', 'HiveModule']
+
+
+class Module(object):
+"""
+Modules define a set of metadata, including functions, user defined types, 
operators, rules,
+etc.
+Metadata from modules are regarded as built-in or system metadata that 
users can take advantages
+of.
+"""
+
+def __init__(self, j_module):
+self._j_module = j_module
+
+
+class HiveModule(Module):
+"""
+Module to provide Hive built-in metadata.
+"""
+
+def __init__(self, j_hive_module=None):
+gateway = get_gateway()
+
+if j_hive_module is None:

Review comment:
   it seems that the parameter `j_hive_module` is not necessary. I noticed 
that there is such a parameter in HiveCatalog, however I guess we should also 
remove it. What do you think?





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #12453: [FLINK-18061] [table] TableResult#collect method should return a closeable iterator to avoid resource leak

2020-06-02 Thread GitBox


flinkbot commented on pull request #12453:
URL: https://github.com/apache/flink/pull/12453#issuecomment-637937506


   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit c2f3319900afad80a0f9fac5a8c9aa4cfdadd4ca (Wed Jun 03 
03:57:13 UTC 2020)
   
   **Warnings:**
* No documentation files were touched! Remember to keep the Flink docs up 
to date!
* **This pull request references an unassigned [Jira 
ticket](https://issues.apache.org/jira/browse/FLINK-18061).** According to the 
[code contribution 
guide](https://flink.apache.org/contributing/contribute-code.html), tickets 
need to be assigned before starting with the implementation work.
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-18080) Translate the "Kerberos Authentication Setup and Configuration" page into Chinese

2020-06-02 Thread Yangze Guo (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18080?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17124578#comment-17124578
 ] 

Yangze Guo commented on FLINK-18080:


Could someone kindly assign this to me?

> Translate the "Kerberos Authentication Setup and Configuration" page into 
> Chinese
> -
>
> Key: FLINK-18080
> URL: https://issues.apache.org/jira/browse/FLINK-18080
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Yangze Guo
>Priority: Major
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-18080) Translate the "Kerberos Authentication Setup and Configuration" page into Chinese

2020-06-02 Thread Yangze Guo (Jira)
Yangze Guo created FLINK-18080:
--

 Summary: Translate the "Kerberos Authentication Setup and 
Configuration" page into Chinese
 Key: FLINK-18080
 URL: https://issues.apache.org/jira/browse/FLINK-18080
 Project: Flink
  Issue Type: Sub-task
Reporter: Yangze Guo






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-18061) TableResult#collect should return closeable iterator to avoid resource leak

2020-06-02 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18061?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-18061:
---
Labels: pull-request-available  (was: )

> TableResult#collect should return closeable iterator to avoid resource leak
> ---
>
> Key: FLINK-18061
> URL: https://issues.apache.org/jira/browse/FLINK-18061
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API
>Reporter: godfrey he
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>
> as discussed in ML: 
> http://mail-archives.apache.org/mod_mbox/flink-dev/202005.mbox/%3cd4ee47e1-0214-aa2f-f5ac-c9daf708e...@apache.org%3e,
>  we should return a closeable iterator for TableResult#collect method *to 
> avoid resource leak*. The suggested change is:
> {code:java}
> public interface TableResult {
>   CloseableIterator collect();
> }
> {code}
> we use existing {{org.apache.flink.util.CloseableIterator}} instead of 
> additional api class, and {{CollectResultIterator}} can also inherit from 
> {{CloseableIterator}}.
> This change does not break current api.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (FLINK-18072) HBaseLookupFunction can not work with new internal data structure RowData

2020-06-02 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18072?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu reassigned FLINK-18072:
---

Assignee: Leonard Xu

> HBaseLookupFunction can not work with new internal data structure RowData
> -
>
> Key: FLINK-18072
> URL: https://issues.apache.org/jira/browse/FLINK-18072
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / HBase
>Affects Versions: 1.11.0
>Reporter: Leonard Xu
>Assignee: Leonard Xu
>Priority: Major
> Fix For: 1.11.0, 1.12.0
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] godfreyhe opened a new pull request #12453: [FLINK-18061] [table] TableResult#collect method should return a closeable iterator to avoid resource leak

2020-06-02 Thread GitBox


godfreyhe opened a new pull request #12453:
URL: https://github.com/apache/flink/pull/12453


   
   ## What is the purpose of the change
   
   *as discussed in ML: 
http://mail-archives.apache.org/mod_mbox/flink-dev/202005.mbox/%3cd4ee47e1-0214-aa2f-f5ac-c9daf708e...@apache.org%3e,
 we should return a closeable iterator for TableResult#collect method to avoid 
resource leak. This pr aims to change the return type of TableResult#collect 
method from Iterator to CloseableIterator*
   
   
   ## Brief change log
   
 - *change the return type of TableResult#collect method from Iterator 
to CloseableIterator*
 - *CollectResultIterator inherits from CloseableIterator.*
   
   
   ## Verifying this change
   
   *(Please pick either of the following options)*
   
   
   This change added tests and can be verified as follows:
   
 - *Extended TableITCase to verify close logic*
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (yes / **no**)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (**yes** / no)
 - The serializers: (yes / no / don't know)
 - The runtime per-record code paths (performance sensitive): (yes / **no** 
/ don't know)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn/Mesos, ZooKeeper: (yes / **no** / 
don't know)
 - The S3 file system connector: (yes / **no** / don't know)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (yes / **no**)
 - If yes, how is the feature documented? (not applicable / docs / JavaDocs 
/ **not documented**)
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12452: [FLINK-18056][fs-connector] Removing the remaining in-progress file on initialization for hadoop path-based writer

2020-06-02 Thread GitBox


flinkbot edited a comment on pull request #12452:
URL: https://github.com/apache/flink/pull/12452#issuecomment-637922636


   
   ## CI report:
   
   * 7cb10bd1269179bbe66f5b840c0f0d2494c66a55 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2598)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12422: [FLINK-17901][python] Add module interface in PyFlink

2020-06-02 Thread GitBox


flinkbot edited a comment on pull request #12422:
URL: https://github.com/apache/flink/pull/12422#issuecomment-636647115


   
   ## CI report:
   
   * 3a387da20f29d9b1f6baf1e57c6d9979d27dd81f Azure: 
[SUCCESS](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2509)
 
   * a57beb047e52bacd60d9c19cbcd31bc9b1d88ec4 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-13866) develop testing plan for many Hive versions that we support

2020-06-02 Thread Rui Li (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-13866?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17124577#comment-17124577
 ] 

Rui Li commented on FLINK-13866:


We already have the maven profiles to test different hive versions. I think we 
can close this one, if nobody objects.

> develop testing plan for many Hive versions that we support
> ---
>
> Key: FLINK-13866
> URL: https://issues.apache.org/jira/browse/FLINK-13866
> Project: Flink
>  Issue Type: Test
>  Components: Connectors / Hive
>Reporter: Bowen Li
>Assignee: Xuefu Zhang
>Priority: Major
>
> with FLINK-13841, we will start to support quite a few hive versions, let 
> alone other major versions like 1.1, 2.2, and 3.x.
> We need to come up with a testing plan to cover all these Hive versions to 
> guarantee 1) help identify and fix breaking changes ASAP, 2) minimize 
> developers' efforts in manually test and maintain compatibilities of all 
> these Hive versions, and automate as much as possible.
> Set it to 1.10.0 for now.
> cc [~xuefuz] [~lirui] [~Terry1897]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] jxeditor commented on a change in pull request #12431: [FLINK-18055] [sql-client] Fix catalog/database does not exist in sql client

2020-06-02 Thread GitBox


jxeditor commented on a change in pull request #12431:
URL: https://github.com/apache/flink/pull/12431#discussion_r434289783



##
File path: 
flink-table/flink-sql-client/src/main/java/org/apache/flink/table/client/cli/SqlCommandParser.java
##
@@ -132,11 +132,10 @@ private SqlCommandParser() {
cmd = SqlCommand.DROP_CATALOG;
} else if (operation instanceof UseCatalogOperation) {
cmd = SqlCommand.USE_CATALOG;
-   operands = new String[] { String.format("`%s`", 
((UseCatalogOperation) operation).getCatalogName()) };
+   operands = new String[] { ((UseCatalogOperation) 
operation).getCatalogName() };

Review comment:
   ```suggestion
operands = new 
String[]{StringUtils.strip(((UseCatalogOperation) operation).getCatalogName(), 
"`")};
   ```

##
File path: 
flink-table/flink-sql-client/src/main/java/org/apache/flink/table/client/cli/SqlCommandParser.java
##
@@ -132,11 +132,10 @@ private SqlCommandParser() {
cmd = SqlCommand.DROP_CATALOG;
} else if (operation instanceof UseCatalogOperation) {
cmd = SqlCommand.USE_CATALOG;
-   operands = new String[] { String.format("`%s`", 
((UseCatalogOperation) operation).getCatalogName()) };
+   operands = new String[] { ((UseCatalogOperation) 
operation).getCatalogName() };
} else if (operation instanceof UseDatabaseOperation) {
cmd = SqlCommand.USE;
-   UseDatabaseOperation op = ((UseDatabaseOperation) 
operation);
-   operands = new String[] { String.format("`%s`.`%s`", 
op.getCatalogName(), op.getDatabaseName()) };
+   operands = new String[] { ((UseDatabaseOperation) 
operation).getDatabaseName() };

Review comment:
   ```suggestion
operands = new String[]{String.format("%s.%s", 
StringUtils.strip(op.getCatalogName(), "`"), 
StringUtils.strip(op.getDatabaseName(), "`"))};
   ```





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-17733) Add documentation for real-time hive

2020-06-02 Thread Kurt Young (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17733?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kurt Young updated FLINK-17733:
---
Priority: Blocker  (was: Major)

> Add documentation for real-time hive
> 
>
> Key: FLINK-17733
> URL: https://issues.apache.org/jira/browse/FLINK-17733
> Project: Flink
>  Issue Type: Sub-task
>  Components: Documentation
>Reporter: Jingsong Lee
>Priority: Blocker
> Fix For: 1.11.0
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-18061) TableResult#collect should return closeable iterator to avoid resource leak

2020-06-02 Thread godfrey he (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18061?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

godfrey he updated FLINK-18061:
---
Description: 
as discussed in ML: 
http://mail-archives.apache.org/mod_mbox/flink-dev/202005.mbox/%3cd4ee47e1-0214-aa2f-f5ac-c9daf708e...@apache.org%3e,
 we should return a closeable iterator for TableResult#collect method *to avoid 
resource leak*. The suggested change is:

{code:java}

public interface TableResult {

  CloseableIterator collect();

}
{code}

we use existing {{org.apache.flink.util.CloseableIterator}} instead of 
additional api class, and {{CollectResultIterator}} can also inherit from 
{{CloseableIterator}}.

This change does not break current api.

  was:
as discussed in ML: 
http://mail-archives.apache.org/mod_mbox/flink-dev/202005.mbox/%3cd4ee47e1-0214-aa2f-f5ac-c9daf708e...@apache.org%3e,
 we should return a closeable iterator for TableResult#collect method *to avoid 
resource leak*. The suggested change is:

{code:java}

public interface TableResult {

  CloseableRowIterator collect();

}

public interface CloseableRowIterator extends Iterator, AutoCloseable {

} 
{code}

This change does not break current api.


> TableResult#collect should return closeable iterator to avoid resource leak
> ---
>
> Key: FLINK-18061
> URL: https://issues.apache.org/jira/browse/FLINK-18061
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API
>Reporter: godfrey he
>Priority: Blocker
> Fix For: 1.11.0
>
>
> as discussed in ML: 
> http://mail-archives.apache.org/mod_mbox/flink-dev/202005.mbox/%3cd4ee47e1-0214-aa2f-f5ac-c9daf708e...@apache.org%3e,
>  we should return a closeable iterator for TableResult#collect method *to 
> avoid resource leak*. The suggested change is:
> {code:java}
> public interface TableResult {
>   CloseableIterator collect();
> }
> {code}
> we use existing {{org.apache.flink.util.CloseableIterator}} instead of 
> additional api class, and {{CollectResultIterator}} can also inherit from 
> {{CloseableIterator}}.
> This change does not break current api.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-15261) add dedicated documentation for blink planner

2020-06-02 Thread Kurt Young (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15261?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kurt Young updated FLINK-15261:
---
Fix Version/s: (was: 1.10.2)
   (was: 1.11.0)

> add dedicated documentation for blink planner 
> --
>
> Key: FLINK-15261
> URL: https://issues.apache.org/jira/browse/FLINK-15261
> Project: Flink
>  Issue Type: Task
>  Components: Documentation, Table SQL / Planner
>Affects Versions: 1.10.0
>Reporter: Bowen Li
>Assignee: Kurt Young
>Priority: Major
>
> we are missing a dedicated page under `Table API and SQL` section to describe 
> in detail what are the advantages of blink planner, and why users should use 
> it over the legacy one.
> I'm trying to reference a blink planner page in Flink's Hive documentation, 
> and realized there's even not one yet



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-15242) Add doc to introduce ddls or dmls supported by sql cli

2020-06-02 Thread Kurt Young (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15242?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kurt Young updated FLINK-15242:
---
Priority: Critical  (was: Major)

> Add doc to introduce ddls or dmls supported by sql cli
> --
>
> Key: FLINK-15242
> URL: https://issues.apache.org/jira/browse/FLINK-15242
> Project: Flink
>  Issue Type: Sub-task
>  Components: Documentation, Table SQL / Client
>Affects Versions: 1.10.0
>Reporter: Terry Wang
>Priority: Critical
> Fix For: 1.11.0, 1.10.2
>
>
> Now in the document of sql client
> https://ci.apache.org/projects/flink/flink-docs-master/dev/table/sqlClient.html,
>  there isn't a part to introduce the ddls/dmls in a whole story. We should 
> complete it before the 1.10 release.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-17908) Vague document about Kafka config in SQL-CLI

2020-06-02 Thread Kurt Young (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17908?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kurt Young updated FLINK-17908:
---
Priority: Critical  (was: Minor)

> Vague document about Kafka config in SQL-CLI
> 
>
> Key: FLINK-17908
> URL: https://issues.apache.org/jira/browse/FLINK-17908
> Project: Flink
>  Issue Type: Improvement
>  Components: Documentation, Table SQL / API
>Affects Versions: 1.11.0
>Reporter: Shengkai Fang
>Priority: Critical
> Fix For: 1.11.0
>
>
> Currently Flink doesn't offer any default config value for Kafka and use the 
> deault config from Kafka. However, it uses the different config value when 
> describe how to use Kafka Connector in sql-client. Document of the connector 
> use value 'ealiest-offset' for 'connector.startup-mode', which is different 
> from Kafka's default behaviour. I think this vague document may mislead 
> users, especially for newbies. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-14256) [Umbrella] Introduce FileSystemTableFactory with partitioned support

2020-06-02 Thread Kurt Young (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-14256?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kurt Young updated FLINK-14256:
---
Summary: [Umbrella] Introduce FileSystemTableFactory with partitioned 
support  (was: Introduce FileSystemTableFactory with partitioned support)

> [Umbrella] Introduce FileSystemTableFactory with partitioned support
> 
>
> Key: FLINK-14256
> URL: https://issues.apache.org/jira/browse/FLINK-14256
> Project: Flink
>  Issue Type: New Feature
>  Components: Table SQL / Planner
>Reporter: Jingsong Lee
>Assignee: Jingsong Lee
>Priority: Major
> Fix For: 1.11.0
>
>
> Introduce FileSystemTableFactory to unify all file system connectors.
> More information in 
> [https://cwiki.apache.org/confluence/display/FLINK/FLIP-115%3A+Filesystem+connector+in+Table]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-17635) Add documentation about view support

2020-06-02 Thread Kurt Young (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17635?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kurt Young updated FLINK-17635:
---
Priority: Blocker  (was: Major)

> Add documentation about view support 
> -
>
> Key: FLINK-17635
> URL: https://issues.apache.org/jira/browse/FLINK-17635
> Project: Flink
>  Issue Type: Sub-task
>  Components: Documentation
>Reporter: Kurt Young
>Assignee: Caizhi Weng
>Priority: Blocker
> Fix For: 1.11.0
>
>




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-13866) develop testing plan for many Hive versions that we support

2020-06-02 Thread Kurt Young (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-13866?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kurt Young updated FLINK-13866:
---
Fix Version/s: (was: 1.11.0)

> develop testing plan for many Hive versions that we support
> ---
>
> Key: FLINK-13866
> URL: https://issues.apache.org/jira/browse/FLINK-13866
> Project: Flink
>  Issue Type: Test
>  Components: Connectors / Hive
>Reporter: Bowen Li
>Assignee: Xuefu Zhang
>Priority: Major
>
> with FLINK-13841, we will start to support quite a few hive versions, let 
> alone other major versions like 1.1, 2.2, and 3.x.
> We need to come up with a testing plan to cover all these Hive versions to 
> guarantee 1) help identify and fix breaking changes ASAP, 2) minimize 
> developers' efforts in manually test and maintain compatibilities of all 
> these Hive versions, and automate as much as possible.
> Set it to 1.10.0 for now.
> cc [~xuefuz] [~lirui] [~Terry1897]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-16448) add documentation for Hive table source and sink parallelism setting strategy

2020-06-02 Thread Kurt Young (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16448?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kurt Young updated FLINK-16448:
---
Component/s: Documentation

> add documentation for Hive table source and sink parallelism setting strategy
> -
>
> Key: FLINK-16448
> URL: https://issues.apache.org/jira/browse/FLINK-16448
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Hive, Documentation
>Reporter: Bowen Li
>Assignee: Jingsong Lee
>Priority: Major
> Fix For: 1.11.0
>
>
> per user-zh mailing list question, would be beneficial to add documentation 
> for Hive table sink parallelism setting strategy



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-16448) add documentation for Hive table source and sink parallelism setting strategy

2020-06-02 Thread Kurt Young (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16448?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kurt Young updated FLINK-16448:
---
Priority: Critical  (was: Major)

> add documentation for Hive table source and sink parallelism setting strategy
> -
>
> Key: FLINK-16448
> URL: https://issues.apache.org/jira/browse/FLINK-16448
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Hive, Documentation
>Reporter: Bowen Li
>Assignee: Jingsong Lee
>Priority: Critical
> Fix For: 1.11.0
>
>
> per user-zh mailing list question, would be beneficial to add documentation 
> for Hive table sink parallelism setting strategy



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-18079) KafkaShuffle Manual Tests

2020-06-02 Thread Yuan Mei (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18079?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yuan Mei updated FLINK-18079:
-
Description: Manual Tests and Results to demonstrate KafkaShuffle working 
as expected.  (was: Manual Tests and Results to demonstrate KafkaShuffle is 
working as expected.)

> KafkaShuffle Manual Tests
> -
>
> Key: FLINK-18079
> URL: https://issues.apache.org/jira/browse/FLINK-18079
> Project: Flink
>  Issue Type: Improvement
>  Components: API / DataStream, Runtime / Checkpointing
>Affects Versions: 1.11.0
>Reporter: Yuan Mei
>Priority: Major
> Fix For: 1.11.0
>
>
> Manual Tests and Results to demonstrate KafkaShuffle working as expected.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #12452: [FLINK-18056][fs-connector] Removing the remaining in-progress file on initialization for hadoop path-based writer

2020-06-02 Thread GitBox


flinkbot edited a comment on pull request #12452:
URL: https://github.com/apache/flink/pull/12452#issuecomment-637922636


   
   ## CI report:
   
   * 7cb10bd1269179bbe66f5b840c0f0d2494c66a55 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2598)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12451: [FLINK-17857][e2e] Fix Kubernetes and docker e2e tests on Mac OS

2020-06-02 Thread GitBox


flinkbot edited a comment on pull request #12451:
URL: https://github.com/apache/flink/pull/12451#issuecomment-637922585


   
   ## CI report:
   
   * 4c93b68fa64dd913f53d69dad4eab467e78e6507 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2597)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot edited a comment on pull request #12444: [FLINK-18067]Change default value of yarnMinAllocationMB from zero to DEFAULT_RM_S…

2020-06-02 Thread GitBox


flinkbot edited a comment on pull request #12444:
URL: https://github.com/apache/flink/pull/12444#issuecomment-637596755


   
   ## CI report:
   
   * 16ec8b694dc07564b09b60040d20b0da9c0e80c4 UNKNOWN
   * 59c44869c548ce9b1930d5b742a6f2c256c5a5b5 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2576)
 
   * c23450c9e7735ca7bad0ff97c37480de6c420d0e Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2596)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Created] (FLINK-18079) KafkaShuffle Manual Tests

2020-06-02 Thread Yuan Mei (Jira)
Yuan Mei created FLINK-18079:


 Summary: KafkaShuffle Manual Tests
 Key: FLINK-18079
 URL: https://issues.apache.org/jira/browse/FLINK-18079
 Project: Flink
  Issue Type: Improvement
  Components: API / DataStream, Runtime / Checkpointing
Affects Versions: 1.11.0
Reporter: Yuan Mei
 Fix For: 1.11.0


Manual Tests and Results to demonstrate KafkaShuffle is working as expected.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] danny0405 commented on a change in pull request #12386: [FLINK-17995][docs][connectors] Redesign Table & SQL Connectors page and add HBase connector documentation

2020-06-02 Thread GitBox


danny0405 commented on a change in pull request #12386:
URL: https://github.com/apache/flink/pull/12386#discussion_r434281449



##
File path: docs/dev/table/connectors/index.zh.md
##
@@ -0,0 +1,268 @@
+---
+title: "Table & SQL Connectors"
+nav-id: sql-connectors
+nav-parent_id: connectors-root
+nav-pos: 2
+nav-show_overview: true
+---
+
+
+
+Flink's Table API & SQL programs can be connected to other external systems 
for reading and writing both batch and streaming tables. A table source 
provides access to data which is stored in external systems (such as a 
database, key-value store, message queue, or file system). A table sink emits a 
table to an external storage system. Depending on the type of source and sink, 
they support different formats such as CSV, Avro, Parquet, or ORC.
+
+This page describes how to register table sources and table sinks in Flink 
using the natively supported connectors. After a source or sink has been 
registered, it can be accessed by Table API & SQL statements.
+
+NOTE If you want to implement your own 
*custom* table source or sink, have a look at the [user-defined sources & sinks 
page](sourceSinks.html).
+
+Attention Flink Table & SQL introduces 
a new set of connector options since 1.11.0, if you are using the legacy 
connector options, please refer to the [legacy documentation]({{ site.baseurl 
}}/dev/table/connect.html).
+
+* This will be replaced by the TOC
+{:toc}
+
+Supported Connectors
+
+
+Flink natively support various connectors. The following tables list all 
available connectors.
+
+
+
+  
+Name
+Version
+Source
+Sink
+  
+
+
+
+  Filesystem
+  
+  Bounded and Unbounded Scan, Lookup
+  Streaming Sink, Batch Sink
+
+
+  Elasticsearch
+  6.x & 7.x
+  Not supported
+  Streaming Sink, Batch Sink
+
+
+  Apache Kafka
+  0.10+
+  Unbounded Scan
+  Streaming Sink, Batch Sink
+
+
+  JDBC
+  
+  Bounded Scan, Lookup
+  Streaming Sink, Batch Sink
+
+
+  Apache 
HBase
+  1.4.x
+  Bounded Scan, Lookup
+  Streaming Sink, Batch Sink
+
+
+
+
+{% top %}
+
+How to use connectors
+
+
+Flink supports to use SQL CREATE TABLE statement to register a table. One can 
define the name of the table, the schema of the table, the connector options 
for connecting to an external system.
+
+The following code shows a full example of how to connect to Kafka for reading 
Json records.
+
+
+
+{% highlight sql %}
+CREATE TABLE MyUserTable (
+  -- declare the schema of the table
+  `user` BIGINT,
+  message STRING,
+  ts TIMESTAMP,
+  proctime AS PROCTIME(), -- use computed column to define proctime attribute
+  WATERMARK FOR ts AS ts - INTERVAL '5' SECOND  -- use WATERMARK statement to 
define rowtime attribute
+) WITH (
+  -- declare the external system to connect to
+  'connector' = 'kafka',
+  'topic' = 'topic_name',
+  'scan.startup.mode' = 'earliest-offset',
+  'properties.bootstrap.servers' = 'localhost:9092',
+  'format' = 'json'   -- declare a format for this system
+)
+{% endhighlight %}
+
+
+
+In this ways the desired connection properties are converted into normalized, 
string-based key-value pairs. So-called [table 
factories](sourceSinks.html#define-a-tablefactory) create configured table 
sources, table sinks, and corresponding formats from the key-value pairs. All 
table factories that can be found via Java's [Service Provider Interfaces 
(SPI)](https://docs.oracle.com/javase/tutorial/sound/SPI-intro.html) are taken 
into account when searching for exactly-one matching table factory.
+
+If no factory can be found or multiple factories match for the given 
properties, an exception will be thrown with additional information about 
considered factories and supported properties.
+
+{% top %}
+
+Schema Mapping
+
+
+The body clause of a SQL `CREATE TABLE` statement defines the names and types 
of columns, constraints and watermarks. Flink doesn't hold the data, thus the 
schema definition only declares how to map types from an external system to 
Flink’s representation. The mapping may not be mapped by names, it depends on 
the implementation of formats and connectors. For example, a MySQL database 
table is mapped by field names (not case sensitive), and a CSV filesystem is 
mapped by field order (field names can be arbitrary). This will be explained in 
every connectors.
+
+The following example shows a simple schema without time attributes and 
one-to-one field mapping of input/output to table columns.
+
+
+
+{% highlight sql %}
+CREATE TABLE MyTable (
+  MyField1 INT,
+  MyField2 STRING,
+  MyField3 BOOLEAN
+) WITH (
+  ...
+)
+{% endhighlight %}
+
+
+
+### Primary Key
+
+Primary key constraints tell that a column or a set of columns of a table are 
unique and they do not contain nulls. Primary key uniquely identify a row in a 
table.
+
+The primary key of 

[jira] [Closed] (FLINK-17940) It will throw NullPointerException when write data with Avro format using new property key in SQL-Client

2020-06-02 Thread Jark Wu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-17940?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jark Wu closed FLINK-17940.
---
Fix Version/s: (was: 1.12.0)
   Resolution: Duplicate

> It will throw NullPointerException when write data with Avro format using new 
> property key in SQL-Client 
> -
>
> Key: FLINK-17940
> URL: https://issues.apache.org/jira/browse/FLINK-17940
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kafka, Table SQL / Client
>Affects Versions: 1.11.0
> Environment: Docker Environment:
> zookeeper:
>  image: wurstmeister/zookeeper:3.4.6
>  ports:
>  - "2181:2181"
> kafka:
>  image: wurstmeister/kafka:2.12-2.2.1
>  ports:
>  - "9092:9092"
>  - "9094:9094"
> depends_on:
>       zookeeper
>  environment:
>  - KAFKA_ADVERTISED_LISTENERS=INSIDE://:9094,OUTSIDE://localhost:9092
>  - KAFKA_LISTENERS=INSIDE://:9094,OUTSIDE://:9092
>  - KAFKA_LISTENER_SECURITY_PROTOCOL_MAP=INSIDE:PLAINTEXT,OUTSIDE:PLAINTEXT
>  - KAFKA_INTER_BROKER_LISTENER_NAME=INSIDE
>  - KAFKA_ZOOKEEPER_CONNECT=zookeeper:2181
>  - KAFKA_CREATE_TOPICS:"order_cnt:1:1,orders:1:1,currency:1:1"
> volumes:
>  - /var/run/docker.sock:/var/run/docker.sock
>Reporter: Shengkai Fang
>Priority: Major
>
> For the following job:
> {noformat}
> create table csv( 
> user_name VARCHAR, is_new BOOLEAN, content VARCHAR
> ) with ( 
> 'connector' = 'filesystem', 
> 'path' = '/Users/ohmeatball/Work/flink-sql-etl/data-  
> generator/src/main/resources/user.csv', 
> 'format' = 'csv');
> -
> CREATE TABLE AvroTest ( 
> user_name VARCHAR, is_new BOOLEAN, content VARCHAR
> ) WITH (
> 'connector' = 'kafka', 'topic' = 'avro_from_csv',   
> 'properties.zookeeper.connect' = 'localhost:2181', 
> 'properties.bootstrap.servers' = 'localhost:9092', 'properties.group.id' = 
> 'testGroup3', 'scan.startup.mode' = 'earliest-offset', 'format' = 'avro');
> -
> insert into AvroTest select user_name, is_new, content from csv;
> {noformat}
> The exception stack is following:
>  
> {code:java}
> 2020-05-26 19:51:22,212 WARN  org.apache.flink.runtime.taskmanager.Task   
>  [] - FileSystemTableSource(user_name, is_new, content) -> Sink: 
> Sink(table=[default_catalog.default_database.AvroTest], fields=[user_name, 
> is_new, content]) (1/1) (283a383f3ed93b051f56d4b5aca7dfb9) switched from 
> RUNNING to FAILED.java.lang.RuntimeException: Failed to serialize row.at 
> org.apache.flink.formats.avro.AvroRowDataSerializationSchema.serialize(AvroRowDataSerializationSchema.java:118)
>  ~[flink-avro-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]at 
> org.apache.flink.formats.avro.AvroRowDataSerializationSchema.serialize(AvroRowDataSerializationSchema.java:63)
>  ~[flink-avro-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]at 
> org.apache.flink.streaming.connectors.kafka.internals.KeyedSerializationSchemaWrapper.serializeValue(KeyedSerializationSchemaWrapper.java:51)
>  ~[flink-sql-connector-kafka_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]at 
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer.invoke(FlinkKafkaProducer.java:775)
>  ~[flink-sql-connector-kafka_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]at 
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer.invoke(FlinkKafkaProducer.java:98)
>  ~[flink-sql-connector-kafka_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]at 
> org.apache.flink.streaming.api.functions.sink.TwoPhaseCommitSinkFunction.invoke(TwoPhaseCommitSinkFunction.java:235)
>  ~[flink-dist_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]at 
> org.apache.flink.table.runtime.operators.sink.SinkOperator.processElement(SinkOperator.java:86)
>  ~[flink-table-blink_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.pushToOperator(OperatorChain.java:717)
>  ~[flink-dist_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:692)
>  ~[flink-dist_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:672)
>  ~[flink-dist_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]at 
> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:52)
>  ~[flink-dist_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]at 
> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:30)
>  ~[flink-dist_2.11-1.11-SNAPSHOT.jar:1.11-SNAPSHOT]at 
> 

[jira] [Created] (FLINK-18078) E2E tests manually for Hive streaming dim join

2020-06-02 Thread Jingsong Lee (Jira)
Jingsong Lee created FLINK-18078:


 Summary: E2E tests manually for Hive streaming dim join
 Key: FLINK-18078
 URL: https://issues.apache.org/jira/browse/FLINK-18078
 Project: Flink
  Issue Type: Sub-task
  Components: Connectors / Hive
Reporter: Jingsong Lee
 Fix For: 1.11.0






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on pull request #12444: [FLINK-18067]Change default value of yarnMinAllocationMB from zero to DEFAULT_RM_S…

2020-06-02 Thread GitBox


flinkbot edited a comment on pull request #12444:
URL: https://github.com/apache/flink/pull/12444#issuecomment-637596755


   
   ## CI report:
   
   * 16ec8b694dc07564b09b60040d20b0da9c0e80c4 UNKNOWN
   * 59c44869c548ce9b1930d5b742a6f2c256c5a5b5 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2576)
 
   * c23450c9e7735ca7bad0ff97c37480de6c420d0e UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Created] (FLINK-18077) E2E tests manually for Hive streaming source

2020-06-02 Thread Jingsong Lee (Jira)
Jingsong Lee created FLINK-18077:


 Summary: E2E tests manually for Hive streaming source
 Key: FLINK-18077
 URL: https://issues.apache.org/jira/browse/FLINK-18077
 Project: Flink
  Issue Type: Sub-task
  Components: Connectors / Hive
Reporter: Jingsong Lee
 Fix For: 1.11.0






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot commented on pull request #12451: [FLINK-17857][e2e] Fix Kubernetes and docker e2e tests on Mac OS

2020-06-02 Thread GitBox


flinkbot commented on pull request #12451:
URL: https://github.com/apache/flink/pull/12451#issuecomment-637922585


   
   ## CI report:
   
   * 4c93b68fa64dd913f53d69dad4eab467e78e6507 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-18025) E2E tests manually for Hive streaming sink

2020-06-02 Thread Jingsong Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18025?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee updated FLINK-18025:
-
Description: 
- hive streaming sink failover
 - hive streaming sink job re-run
 - hive streaming sink without partition
 - ...

  was:
- Hive streaming source
- Hive streaming sink
- Hive dim join


> E2E tests manually for Hive streaming sink
> --
>
> Key: FLINK-18025
> URL: https://issues.apache.org/jira/browse/FLINK-18025
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / Hive
>Affects Versions: 1.11.0
>Reporter: Danny Chen
>Priority: Major
> Fix For: 1.11.0
>
>
> - hive streaming sink failover
>  - hive streaming sink job re-run
>  - hive streaming sink without partition
>  - ...



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot commented on pull request #12452: [FLINK-18056][fs-connector] Removing the remaining in-progress file on initialization for hadoop path-based writer

2020-06-02 Thread GitBox


flinkbot commented on pull request #12452:
URL: https://github.com/apache/flink/pull/12452#issuecomment-637922636


   
   ## CI report:
   
   * 7cb10bd1269179bbe66f5b840c0f0d2494c66a55 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Commented] (FLINK-18076) Sql client uses wrong class loader when parsing queries

2020-06-02 Thread Jark Wu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18076?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17124547#comment-17124547
 ] 

Jark Wu commented on FLINK-18076:
-

>From the recently bug issues, it seems that it's urgent to add end-to-end 
>tests for the new connectors and new DDLs. 

> Sql client uses wrong class loader when parsing queries
> ---
>
> Key: FLINK-18076
> URL: https://issues.apache.org/jira/browse/FLINK-18076
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Client
>Affects Versions: 1.11.0, 1.12.0
>Reporter: Dawid Wysakowicz
>Priority: Blocker
> Fix For: 1.11.0
>
>
> Sql-client when parsing queries does not use the user class loader from 
> {{ExecutionContext}}. This makes it impossible to query any sources if the 
> dependencies are added with {{-j}} flag.
> In order to reproduce it try querying e.g. KafkaDynamicSource with
> {code}
> CREATE TABLE MyUserTable (
>f0 BIGINT
> ) WITH (
>   'connector' = 'kafka',   
>   'topic' = 'topic_name', -- required: topic name from which the table is read
>  -- required: specify the Kafka server connection string
>   'properties.bootstrap.servers' = 'localhost:9092',
>   -- required for Kafka source, optional for Kafka sink, specify consumer 
> group
>   'properties.group.id' = 'testGroup',
>   -- optional: valid modes are "earliest-offset", "latest-offset", 
> "group-offsets", "specific-offsets" or "timestamp"
> 'scan.startup.mode' = 'earliest-offset',
>'format' = 'avro'
> );
> SELECT * FROM MyUserTable;
> {code}
> It give exception:
> {code}
> Exception in thread "main" org.apache.flink.table.client.SqlClientException: 
> Unexpected exception. This is a bug. Please consider filing an issue.
>   at org.apache.flink.table.client.SqlClient.main(SqlClient.java:213)
> Caused by: org.apache.flink.table.client.gateway.SqlExecutionException: 
> Invalidate SQL statement.
>   at 
> org.apache.flink.table.client.cli.SqlCommandParser.parseBySqlParser(SqlCommandParser.java:95)
>   at 
> org.apache.flink.table.client.cli.SqlCommandParser.parse(SqlCommandParser.java:79)
>   at 
> org.apache.flink.table.client.cli.CliClient.parseCommand(CliClient.java:256)
>   at org.apache.flink.table.client.cli.CliClient.open(CliClient.java:212)
>   at org.apache.flink.table.client.SqlClient.openCli(SqlClient.java:142)
>   at org.apache.flink.table.client.SqlClient.start(SqlClient.java:114)
>   at org.apache.flink.table.client.SqlClient.main(SqlClient.java:201)
> Caused by: org.apache.flink.table.api.ValidationException: Unable to create a 
> source for reading table 'default_catalog.default_database.MyUserTable'.
> Table options are:
> 'connector'='kafka'
> 'format'='avro'
> 'properties.bootstrap.servers'='localhost:9092'
> 'properties.group.id'='testGroup'
> 'scan.startup.mode'='earliest-offset'
> 'topic'='topic_name'
>   at 
> org.apache.flink.table.factories.FactoryUtil.createTableSource(FactoryUtil.java:125)
>   at 
> org.apache.flink.table.planner.plan.schema.CatalogSourceTable.buildTableScan(CatalogSourceTable.scala:135)
>   at 
> org.apache.flink.table.planner.plan.schema.CatalogSourceTable.toRel(CatalogSourceTable.scala:78)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.toRel(SqlToRelConverter.java:3492)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertIdentifier(SqlToRelConverter.java:2415)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2102)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertFrom(SqlToRelConverter.java:2051)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertSelectImpl(SqlToRelConverter.java:661)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertSelect(SqlToRelConverter.java:642)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertQueryRecursive(SqlToRelConverter.java:3345)
>   at 
> org.apache.calcite.sql2rel.SqlToRelConverter.convertQuery(SqlToRelConverter.java:568)
>   at 
> org.apache.flink.table.planner.calcite.FlinkPlannerImpl.org$apache$flink$table$planner$calcite$FlinkPlannerImpl$$rel(FlinkPlannerImpl.scala:164)
>   at 
> org.apache.flink.table.planner.calcite.FlinkPlannerImpl.rel(FlinkPlannerImpl.scala:151)
>   at 
> org.apache.flink.table.planner.operations.SqlToOperationConverter.toQueryOperation(SqlToOperationConverter.java:773)
>   at 
> org.apache.flink.table.planner.operations.SqlToOperationConverter.convertSqlQuery(SqlToOperationConverter.java:745)
>   at 
> org.apache.flink.table.planner.operations.SqlToOperationConverter.convert(SqlToOperationConverter.java:238)
>   at 
> org.apache.flink.table.planner.delegation.ParserImpl.parse(ParserImpl.java:66)
>   at 
> 

[GitHub] [flink] flinkbot edited a comment on pull request #12144: [FLINK-17384][flink-dist] support read hbase conf dir from flink.conf and change HBaseConfiguration construction.

2020-06-02 Thread GitBox


flinkbot edited a comment on pull request #12144:
URL: https://github.com/apache/flink/pull/12144#issuecomment-628475376


   
   ## CI report:
   
   * d36b959fb16a91c15babda10dd884cbbdec58420 UNKNOWN
   * 3df97a6fdb130416fa447295f4ac7eb174be9cc4 Azure: 
[FAILURE](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2557)
 
   * 84f420d9043272150872fe4a32c2eaad2400a750 Azure: 
[PENDING](https://dev.azure.com/apache-flink/98463496-1af2-4620-8eab-a2ecc1a2e6fe/_build/results?buildId=2592)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-18025) E2E tests manually for Hive streaming sink

2020-06-02 Thread Jingsong Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18025?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee updated FLINK-18025:
-
Summary: E2E tests manually for Hive streaming sink  (was: E2E tests 
manually for Hive integrate streaming)

> E2E tests manually for Hive streaming sink
> --
>
> Key: FLINK-18025
> URL: https://issues.apache.org/jira/browse/FLINK-18025
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / Hive
>Affects Versions: 1.11.0
>Reporter: Danny Chen
>Priority: Major
> Fix For: 1.11.0
>
>
> - Hive streaming source
> - Hive streaming sink
> - Hive dim join



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-18056) Hive file sink throws exception when the target in-progress file exists.

2020-06-02 Thread Yun Gao (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18056?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17124542#comment-17124542
 ] 

Yun Gao commented on FLINK-18056:
-

Hi [~sjwiesman] I'm a little confusing on the scenario since when rollover the 
part-file on checkpoint, it should always write to a new file. Without failover 
or restart the new part-file should not exist before and thus the corresponding 
in-progress file should not be created. A single sink with parallelism > 1 
should also not cause the conflict since the part-file name contains the 
subtask index. Does the job also have multiple sink writing to the same Hive 
table ? Or is it possible that the Hive table has been written by some jobs 
previously but may left some hidden files not cleared when cleaning the 
previous results ? 

> Hive file sink throws exception when the target in-progress file exists.
> 
>
> Key: FLINK-18056
> URL: https://issues.apache.org/jira/browse/FLINK-18056
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / FileSystem, Connectors / Hive
>Reporter: Yun Gao
>Assignee: Yun Gao
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>
> Currently after failover or restart, the Hive file sink will try to overwrite 
> the data since the last checkpoint, however, currently neither the 
> in-progress file is deleted nor hive uses the overwritten mode, thus an 
> exception occurs after restarting:
> {code:java}
> org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException):
>  failed to create file 
> /user/hive/warehouse/datalake/dt=2020-06-01/hr=22/.part-0-10.inprogress for 
> DFSClient_NONMAPREDUCE_-1017064593_62 for client 100.96.206.42 because 
> current leaseholder is trying to recreate file.
> {code}
> The full stack of the exception is
> {code:java}
> org.apache.flink.connectors.hive.FlinkHiveException: 
> org.apache.flink.table.catalog.exceptions.CatalogException: Failed to create 
> Hive RecordWriter
> at 
> org.apache.flink.connectors.hive.write.HiveWriterFactory.createRecordWriter(HiveWriterFactory.java:159)
> at 
> org.apache.flink.connectors.hive.write.HiveBulkWriterFactory.create(HiveBulkWriterFactory.java:47)
> at 
> org.apache.flink.formats.hadoop.bulk.HadoopPathBasedPartFileWriter$HadoopPathBasedBucketWriter.openNewInProgressFile(HadoopPathBasedPartFileWriter.java:234)
> at 
> org.apache.flink.formats.hadoop.bulk.HadoopPathBasedPartFileWriter$HadoopPathBasedBucketWriter.openNewInProgressFile(HadoopPathBasedPartFileWriter.java:207)
> at 
> org.apache.flink.streaming.api.functions.sink.filesystem.Bucket.rollPartFile(Bucket.java:209)
> at 
> org.apache.flink.streaming.api.functions.sink.filesystem.Bucket.write(Bucket.java:200)
> at 
> org.apache.flink.streaming.api.functions.sink.filesystem.Buckets.onElement(Buckets.java:284)
> at 
> org.apache.flink.streaming.api.functions.sink.filesystem.StreamingFileSinkHelper.onElement(StreamingFileSinkHelper.java:104)
> at 
> org.apache.flink.table.filesystem.stream.StreamingFileWriter.processElement(StreamingFileWriter.java:118)
> at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.pushToOperator(OperatorChain.java:717)
> at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:692)
> at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:672)
> at 
> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:52)
> at 
> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:30)
> at StreamExecCalc$16.processElement(Unknown Source)
> at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.pushToOperator(OperatorChain.java:717)
> at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:692)
> at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:672)
> at 
> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:52)
> at 
> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:30)
> at 
> org.apache.flink.table.runtime.operators.wmassigners.WatermarkAssignerOperator.processElement(WatermarkAssignerOperator.java:123)
> at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.pushToOperator(OperatorChain.java:717)
> at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:692)
> at 
> 

[GitHub] [flink] Jiayi-Liao commented on pull request #12444: [FLINK-18067]Change default value of yarnMinAllocationMB from zero to DEFAULT_RM_S…

2020-06-02 Thread GitBox


Jiayi-Liao commented on pull request #12444:
URL: https://github.com/apache/flink/pull/12444#issuecomment-637920335


   @wangyang0918  Thanks for your review. I've addressed your comments and also 
update the description. Please take another look when you can spare time.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] flinkbot commented on pull request #12452: [FLINK-18056][fs-connector] Removing the remaining in-progress file on initialization for hadoop path-based writer

2020-06-02 Thread GitBox


flinkbot commented on pull request #12452:
URL: https://github.com/apache/flink/pull/12452#issuecomment-637919303


   Thanks a lot for your contribution to the Apache Flink project. I'm the 
@flinkbot. I help the community
   to review your pull request. We will use this comment to track the progress 
of the review.
   
   
   ## Automated Checks
   Last check on commit 7cb10bd1269179bbe66f5b840c0f0d2494c66a55 (Wed Jun 03 
02:43:50 UTC 2020)
   
   **Warnings:**
* No documentation files were touched! Remember to keep the Flink docs up 
to date!
   
   
   Mention the bot in a comment to re-run the automated checks.
   ## Review Progress
   
   * ❓ 1. The [description] looks good.
   * ❓ 2. There is [consensus] that the contribution should go into to Flink.
   * ❓ 3. Needs [attention] from.
   * ❓ 4. The change fits into the overall [architecture].
   * ❓ 5. Overall code [quality] is good.
   
   Please see the [Pull Request Review 
Guide](https://flink.apache.org/contributing/reviewing-prs.html) for a full 
explanation of the review process.
The Bot is tracking the review progress through labels. Labels are applied 
according to the order of the review items. For consensus, approval by a Flink 
committer of PMC member is required Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot approve description` to approve one or more aspects (aspects: 
`description`, `consensus`, `architecture` and `quality`)
- `@flinkbot approve all` to approve all aspects
- `@flinkbot approve-until architecture` to approve everything until 
`architecture`
- `@flinkbot attention @username1 [@username2 ..]` to require somebody's 
attention
- `@flinkbot disapprove architecture` to remove an approval you gave earlier
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[GitHub] [flink] Jiayi-Liao commented on a change in pull request #12444: [FLINK-18067]Change default value of yarnMinAllocationMB from zero to DEFAULT_RM_S…

2020-06-02 Thread GitBox


Jiayi-Liao commented on a change in pull request #12444:
URL: https://github.com/apache/flink/pull/12444#discussion_r434277415



##
File path: 
flink-yarn/src/main/java/org/apache/flink/yarn/YarnClusterDescriptor.java
##
@@ -508,7 +508,9 @@ public void killCluster(ApplicationId applicationId) throws 
FlinkException {
throw new YarnDeploymentException("Could not retrieve 
information about free cluster resources.", e);
}
 
-   final int yarnMinAllocationMB = 
yarnConfiguration.getInt(YarnConfiguration.RM_SCHEDULER_MINIMUM_ALLOCATION_MB, 
0);
+   final int yarnMinAllocationMB = yarnConfiguration.getInt(

Review comment:
   I put in `validateClusterResources` because I thought it's a part of the 
validation, but I'm also okay with your way.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org




[jira] [Updated] (FLINK-18056) Hive file sink throws exception when the target in-progress file exists.

2020-06-02 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18056?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-18056:
---
Labels: pull-request-available  (was: )

> Hive file sink throws exception when the target in-progress file exists.
> 
>
> Key: FLINK-18056
> URL: https://issues.apache.org/jira/browse/FLINK-18056
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / FileSystem, Connectors / Hive
>Reporter: Yun Gao
>Assignee: Yun Gao
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>
> Currently after failover or restart, the Hive file sink will try to overwrite 
> the data since the last checkpoint, however, currently neither the 
> in-progress file is deleted nor hive uses the overwritten mode, thus an 
> exception occurs after restarting:
> {code:java}
> org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException):
>  failed to create file 
> /user/hive/warehouse/datalake/dt=2020-06-01/hr=22/.part-0-10.inprogress for 
> DFSClient_NONMAPREDUCE_-1017064593_62 for client 100.96.206.42 because 
> current leaseholder is trying to recreate file.
> {code}
> The full stack of the exception is
> {code:java}
> org.apache.flink.connectors.hive.FlinkHiveException: 
> org.apache.flink.table.catalog.exceptions.CatalogException: Failed to create 
> Hive RecordWriter
> at 
> org.apache.flink.connectors.hive.write.HiveWriterFactory.createRecordWriter(HiveWriterFactory.java:159)
> at 
> org.apache.flink.connectors.hive.write.HiveBulkWriterFactory.create(HiveBulkWriterFactory.java:47)
> at 
> org.apache.flink.formats.hadoop.bulk.HadoopPathBasedPartFileWriter$HadoopPathBasedBucketWriter.openNewInProgressFile(HadoopPathBasedPartFileWriter.java:234)
> at 
> org.apache.flink.formats.hadoop.bulk.HadoopPathBasedPartFileWriter$HadoopPathBasedBucketWriter.openNewInProgressFile(HadoopPathBasedPartFileWriter.java:207)
> at 
> org.apache.flink.streaming.api.functions.sink.filesystem.Bucket.rollPartFile(Bucket.java:209)
> at 
> org.apache.flink.streaming.api.functions.sink.filesystem.Bucket.write(Bucket.java:200)
> at 
> org.apache.flink.streaming.api.functions.sink.filesystem.Buckets.onElement(Buckets.java:284)
> at 
> org.apache.flink.streaming.api.functions.sink.filesystem.StreamingFileSinkHelper.onElement(StreamingFileSinkHelper.java:104)
> at 
> org.apache.flink.table.filesystem.stream.StreamingFileWriter.processElement(StreamingFileWriter.java:118)
> at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.pushToOperator(OperatorChain.java:717)
> at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:692)
> at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:672)
> at 
> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:52)
> at 
> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:30)
> at StreamExecCalc$16.processElement(Unknown Source)
> at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.pushToOperator(OperatorChain.java:717)
> at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:692)
> at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:672)
> at 
> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:52)
> at 
> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:30)
> at 
> org.apache.flink.table.runtime.operators.wmassigners.WatermarkAssignerOperator.processElement(WatermarkAssignerOperator.java:123)
> at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.pushToOperator(OperatorChain.java:717)
> at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:692)
> at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:672)
> at 
> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:52)
> at 
> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:30)
> at StreamExecCalc$2.processElement(Unknown Source)
> at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.pushToOperator(OperatorChain.java:717)
> at 
> org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:692)
> at 
> 

[GitHub] [flink] dianfu commented on a change in pull request #12422: [FLINK-17901][python] Add module interface in PyFlink

2020-06-02 Thread GitBox


dianfu commented on a change in pull request #12422:
URL: https://github.com/apache/flink/pull/12422#discussion_r434276134



##
File path: flink-python/pyflink/table/tests/test_table_environment_api.py
##
@@ -628,6 +628,15 @@ def test_table_environment_with_blink_planner(self):
 
 self.assert_equals(results, ['2,hi,hello\n', '3,hello,hello\n'])
 
+def test_unload_and_load_module(self):

Review comment:
   Move it to class TableEnvironmentTest?

##
File path: flink-python/pyflink/table/module.py
##
@@ -0,0 +1,77 @@
+
+#  Licensed to the Apache Software Foundation (ASF) under one
+#  or more contributor license agreements.  See the NOTICE file
+#  distributed with this work for additional information
+#  regarding copyright ownership.  The ASF licenses this file
+#  to you under the Apache License, Version 2.0 (the
+#  "License"); you may not use this file except in compliance
+#  with the License.  You may obtain a copy of the License at
+#
+#  http://www.apache.org/licenses/LICENSE-2.0
+#
+#  Unless required by applicable law or agreed to in writing, software
+#  distributed under the License is distributed on an "AS IS" BASIS,
+#  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+#  See the License for the specific language governing permissions and
+# limitations under the License.
+
+from pyflink.java_gateway import get_gateway
+from pyflink.table.function_definition import FunctionDefinition
+
+__all__ = ['Module', 'HiveModule']
+
+
+class Module(object):
+"""
+Modules define a set of metadata, including functions, user defined types, 
operators, rules,
+etc.
+Metadata from modules are regarded as built-in or system metadata that 
users can take advantages
+of.
+"""
+
+def __init__(self, j_module):
+self._j_module = j_module
+
+@staticmethod
+def _get(j_module):
+if j_module.getClass().getName() == 
'org.apache.flink.table.module.hive.HiveModule':
+return HiveModule(j_hive_module=j_module)
+else:
+return Module(j_module)
+
+def list_functions(self):

Review comment:
   This interfaces `list_functions` and `get_function_definition` are 
designed to used by:
   - Internally by the ModuleManager
   - Used to create custom Module
   
   So I think there is no need to expose them to Python users and we should 
remove them(and also the Python FunctionDefinition class).

##
File path: flink-python/pyflink/table/table_environment.py
##
@@ -148,6 +147,29 @@ def get_catalog(self, catalog_name):
 else:
 return None
 
+def load_module(self, module_name, module):
+"""
+Loads a :class:`~pyflink.table.Module` under a unique name. Modules 
will be kept
+in the loaded order.
+ValidationException is thrown when there is already a module with the 
same name.
+
+:param module_name: Name of the :class:`~pyflink.table.Module`.
+:type module_name: str
+:param module: The module instance.
+:type module: pyflink.table.Module

Review comment:
   Add the version information
   ```
   .. versionadded:: 1.12.0
   ```

##
File path: flink-python/pyflink/table/table_environment.py
##
@@ -148,6 +147,29 @@ def get_catalog(self, catalog_name):
 else:
 return None
 
+def load_module(self, module_name, module):
+"""
+Loads a :class:`~pyflink.table.Module` under a unique name. Modules 
will be kept
+in the loaded order.
+ValidationException is thrown when there is already a module with the 
same name.
+
+:param module_name: Name of the :class:`~pyflink.table.Module`.
+:type module_name: str
+:param module: The module instance.
+:type module: pyflink.table.Module
+"""
+self._j_tenv.loadModule(module_name, module._j_module)
+
+def unload_module(self, module_name):
+"""
+Unloads a :class:`~pyflink.table.Module` with given name.
+ValidationException is thrown when there is no module with the given 
name.
+
+:param module_name: Name of the :class:`~pyflink.table.Module`.
+:type module_name: str
+"""
+self._j_tenv.unloadModule(module_name)

Review comment:
   ditto

##
File path: flink-python/pyflink/table/module.py
##
@@ -0,0 +1,77 @@
+
+#  Licensed to the Apache Software Foundation (ASF) under one
+#  or more contributor license agreements.  See the NOTICE file
+#  distributed with this work for additional information
+#  regarding copyright ownership.  The ASF licenses this file
+#  to you under the Apache License, Version 2.0 (the
+#  

[jira] [Updated] (FLINK-18021) Manually tests for 1.11 SQL features

2020-06-02 Thread Kurt Young (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18021?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kurt Young updated FLINK-18021:
---
Summary: Manually tests for 1.11 SQL features  (was: Complement tests for 
1.11 SQL)

> Manually tests for 1.11 SQL features
> 
>
> Key: FLINK-18021
> URL: https://issues.apache.org/jira/browse/FLINK-18021
> Project: Flink
>  Issue Type: Task
>  Components: Table SQL / API
>Affects Versions: 1.11.0
>Reporter: Danny Chen
>Priority: Major
> Fix For: 1.11.0
>
>
> This is an umbrella issue to collect all kinds of tests (e2e and ITCases) 
> that need to cover for 1.11 release.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-18021) Manually tests for 1.11 SQL features

2020-06-02 Thread Kurt Young (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18021?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kurt Young updated FLINK-18021:
---
Priority: Critical  (was: Major)

> Manually tests for 1.11 SQL features
> 
>
> Key: FLINK-18021
> URL: https://issues.apache.org/jira/browse/FLINK-18021
> Project: Flink
>  Issue Type: Task
>  Components: Table SQL / API
>Affects Versions: 1.11.0
>Reporter: Danny Chen
>Priority: Critical
> Fix For: 1.11.0
>
>
> This is an umbrella issue to collect all kinds of tests (e2e and ITCases) 
> that need to cover for 1.11 release.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-18021) Manually tests for 1.11 SQL features

2020-06-02 Thread Kurt Young (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-18021?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kurt Young updated FLINK-18021:
---
Priority: Blocker  (was: Critical)

> Manually tests for 1.11 SQL features
> 
>
> Key: FLINK-18021
> URL: https://issues.apache.org/jira/browse/FLINK-18021
> Project: Flink
>  Issue Type: Task
>  Components: Table SQL / API
>Affects Versions: 1.11.0
>Reporter: Danny Chen
>Priority: Blocker
> Fix For: 1.11.0
>
>
> This is an umbrella issue to collect all kinds of tests (e2e and ITCases) 
> that need to cover for 1.11 release.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


  1   2   3   4   5   6   7   >