[GitHub] [flink] PatrickRen commented on a diff in pull request #20489: [FLINK-28854][connector/jdbc] Migrate JDBC lookup table to the new LookupFunction and caching interface
PatrickRen commented on code in PR #20489: URL: https://github.com/apache/flink/pull/20489#discussion_r940927210 ## flink-connectors/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/table/JdbcDynamicTableFactory.java: ## @@ -181,6 +179,27 @@ private JdbcDmlOptions getJdbcDmlOptions( .build(); } +@Nullable +private LookupCache getLookupCache(ReadableConfig tableOptions) { +LookupCache cache = null; +// Legacy cache options +if (tableOptions.get(LOOKUP_CACHE_MAX_ROWS) > 0 +&& tableOptions.get(LOOKUP_CACHE_TTL).compareTo(Duration.ZERO) > 0) { +cache = +DefaultLookupCache.newBuilder() + .maximumSize(tableOptions.get(LOOKUP_CACHE_MAX_ROWS)) + .expireAfterWrite(tableOptions.get(LOOKUP_CACHE_TTL)) + .cacheMissingKey(tableOptions.get(LOOKUP_CACHE_MISSING_KEY)) +.build(); +} Review Comment: Actually old keys are not fully compatible with new options considering we introduces LookupOptions#CACHE_TYPE and maybe full caching in the future. I'll keep the current logic here for now. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] xintongsong commented on pull request #20456: [FLINK-28785][network] Hybrid shuffle consumer thread and upstream thread may have deadlock
xintongsong commented on PR #20456: URL: https://github.com/apache/flink/pull/20456#issuecomment-1208929966 @flinkbot run azure -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-18572) Flink web UI doesn't display tolerable-failed-checkpoints
[ https://issues.apache.org/jira/browse/FLINK-18572?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17577144#comment-17577144 ] Zhanghao Chen commented on FLINK-18572: --- Verified that this is already included in Web UI in Flink 1.16, we can close this ticket so that we won't confuse others. > Flink web UI doesn't display tolerable-failed-checkpoints > - > > Key: FLINK-18572 > URL: https://issues.apache.org/jira/browse/FLINK-18572 > Project: Flink > Issue Type: Improvement > Components: Runtime / REST, Runtime / Web Frontend >Affects Versions: 1.11.0 >Reporter: Steven Zhen Wu >Priority: Not a Priority > Labels: auto-deprioritized-major, auto-deprioritized-minor > Attachments: image-2020-07-12-10-14-49-990.png > > > might be helpful to display the number of tolerable-failed-checkpoints in web > UI. > > !image-2020-07-12-10-14-49-990.png! -- This message was sent by Atlassian Jira (v8.20.10#820010)
[GitHub] [flink] lindong28 commented on pull request #20275: [FLINK-28606][Runtime/Checkpointing] Preserve consistency of OperatorEvent from OperatorCoordinator to subtasks
lindong28 commented on PR #20275: URL: https://github.com/apache/flink/pull/20275#issuecomment-1208916714 @flinkbot run azure -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] Sxnan commented on pull request #20491: [FLINK-28857][docs] Add Document for DataStream Cache API
Sxnan commented on PR #20491: URL: https://github.com/apache/flink/pull/20491#issuecomment-1208908109 @gaoyunhaii Thanks for your comment! The PR has been updated accordingly. Could you please take another look? -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-25592) Improvement of parser, optimizer and execution for Flink Batch SQL
[ https://issues.apache.org/jira/browse/FLINK-25592?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jark Wu updated FLINK-25592: Fix Version/s: (was: 1.16.0) > Improvement of parser, optimizer and execution for Flink Batch SQL > -- > > Key: FLINK-25592 > URL: https://issues.apache.org/jira/browse/FLINK-25592 > Project: Flink > Issue Type: Improvement > Components: Table SQL / API, Table SQL / Planner, Table SQL / Runtime >Reporter: Jing Zhang >Priority: Major > > This is a parent JIRA to track improvements on Flink Batch SQL, including > parser, optimizer and execution. > For example, > 1. using Hive dialect and default dialect, some sql query would be translated > into different plans > 2. specify hash/sort aggregate strategy and hash/sort merge join strategy in > sql hint > 3. take parquet metadata into consideration in optimization > 4. and so on > Please note, some improvements are not limited to batch sql. Maybe streaming > sql job could also benefits from some improvements in this JIRA. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (FLINK-27660) Table API support create function using customed jar
[ https://issues.apache.org/jira/browse/FLINK-27660?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jark Wu updated FLINK-27660: Fix Version/s: (was: 1.16.0) > Table API support create function using customed jar > > > Key: FLINK-27660 > URL: https://issues.apache.org/jira/browse/FLINK-27660 > Project: Flink > Issue Type: Sub-task > Components: Table SQL / API >Affects Versions: 1.16.0 >Reporter: dalongliu >Assignee: Peter Huang >Priority: Major > Labels: pull-request-available > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Assigned] (FLINK-28871) Make DPP also works if batch shuffle mode is not ALL_BLOCKING
[ https://issues.apache.org/jira/browse/FLINK-28871?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] godfrey he reassigned FLINK-28871: -- Assignee: Yun Gao > Make DPP also works if batch shuffle mode is not ALL_BLOCKING > - > > Key: FLINK-28871 > URL: https://issues.apache.org/jira/browse/FLINK-28871 > Project: Flink > Issue Type: Sub-task > Components: Table SQL / Planner >Affects Versions: 1.16.0 >Reporter: Yun Gao >Assignee: Yun Gao >Priority: Major > Labels: pull-request-available > Fix For: 1.16.0 > > > Currently dpp only works when all edges is blocking. Otherwise if the dynamic > filtering data collector is located in the same region with the fact source, > the fact source would not be started after the data collector task. > To fix this issue, we'll force the collector task's output edges to be > blocking. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[GitHub] [flink] yunfengzhou-hub commented on pull request #20275: [FLINK-28606][Runtime/Checkpointing] Preserve consistency of OperatorEvent from OperatorCoordinator to subtasks
yunfengzhou-hub commented on PR #20275: URL: https://github.com/apache/flink/pull/20275#issuecomment-1208903178 @flinkbot run azure -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] Myasuka commented on a diff in pull request #20405: [FLINK-28010][state] Use deleteRange to optimize the clear method of RocksDBMapState.
Myasuka commented on code in PR #20405: URL: https://github.com/apache/flink/pull/20405#discussion_r940884644 ## flink-state-backends/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/AbstractRocksDBState.java: ## @@ -223,4 +223,27 @@ public StateIncrementalVisitor getStateIncrementalVisitor( throw new UnsupportedOperationException( "Global state entry iterator is unsupported for RocksDb backend"); } + +/** + * Similar to decimal addition, add 1 to the last digit to calculate the upper bound. + * + * @param prefix the starting prefix for seek. + * @return end prefix for seek. + */ +protected final byte[] calculateUpperBound(byte[] prefix) { +byte[] upperBound = new byte[prefix.length]; +System.arraycopy(prefix, 0, upperBound, 0, prefix.length); +boolean overFlow = true; +for (int i = prefix.length - 1; i >= 0; i--) { +int unsignedValue = prefix[i] & 0xff; +int result = unsignedValue + 1; +upperBound[i] = (byte) (result & 0xff); +if (result >> 8 == 0) { +overFlow = false; +break; +} +} +Preconditions.checkArgument(!overFlow, "The upper boundary overflows."); Review Comment: Though RocksDB uses `unsigned` bytes, it will be really strange to treat `-128` with a single byte as not overflow on the java side. As I said before, we will not meet the case to write `-128` within [clipDBWithKeyGroupRange](https://github.com/apache/flink/blob/master/flink-state-backends/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/RocksDBIncrementalCheckpointUtils.java#L128). During the restoring, we cannot change the max parallelism. If we can make the condition `currentKeyGroupRange.getEndKeyGroup() > targetKeyGroupRange.getEndKeyGroup()` pass, the max parallelism is larger than 128, which means the `prefixKeyGroupBytes` is 2. Thus, we will get `0x0080` for the prefix byte array of `127 + 1`. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] godfreyhe commented on a diff in pull request #20497: [FLINK-28871][table-planner] Force the output edges of dynamic filtering data collector to be BLOCKING
godfreyhe commented on code in PR #20497: URL: https://github.com/apache/flink/pull/20497#discussion_r940877755 ## flink-table/flink-table-planner/src/main/java/org/apache/flink/table/planner/plan/nodes/exec/processor/DynamicFilteringDependencyProcessor.java: ## @@ -84,4 +100,108 @@ protected void visitNode(ExecNode node) { return execGraph; } + +private ExecNodeGraph enforceDimSideBlockingExchange( Review Comment: make sense -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Created] (FLINK-28882) ENCODE return error
Luning Wang created FLINK-28882: --- Summary: ENCODE return error Key: FLINK-28882 URL: https://issues.apache.org/jira/browse/FLINK-28882 Project: Flink Issue Type: Improvement Components: Table SQL / API, Table SQL / Planner Affects Versions: 1.15.0 Reporter: Luning Wang Run the following in SQL Client, it will return 'k' rather than 'kyuubi' but it returns 'kyuubi' in the 1.14 version. {code:java} select encode('kyuubi', 'UTF-8') {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[GitHub] [flink] dianfu commented on a diff in pull request #20499: [FLINK-28862][python][format/parquet] Support ParquetRowDataWriter
dianfu commented on code in PR #20499: URL: https://github.com/apache/flink/pull/20499#discussion_r94089 ## flink-python/pyflink/datastream/__init__.py: ## @@ -225,6 +225,9 @@ - :class:`formats.parquet.ParquetColumnarRowInputFormat`: A :class:`connectors.file_system.BulkFormat` to read columnar parquet files into Row data in a batch-processing fashion. +- :class:`formats.parquet.ParquetRowDataWriter`: + Convenient builder to create a :class:`BulkWriterFactory` that writes Rows with a defined Review Comment: ```suggestion Convenient builder to create a :class:`connector.file_system.BulkWriterFactory` that writes Rows with a defined ``` ## flink-python/pyflink/datastream/__init__.py: ## @@ -225,6 +225,9 @@ - :class:`formats.parquet.ParquetColumnarRowInputFormat`: A :class:`connectors.file_system.BulkFormat` to read columnar parquet files into Row data in a batch-processing fashion. +- :class:`formats.parquet.ParquetRowDataWriter`: + Convenient builder to create a :class:`BulkWriterFactory` that writes Rows with a defined + :class:`RowType` into Parquet files in a batch fashion. Review Comment: ```suggestion :class:`pyflink.table.types.RowType` into Parquet files in a batch fashion. ``` ## flink-python/pyflink/datastream/formats/parquet.py: ## @@ -155,3 +158,54 @@ def for_generic_record(schema: 'AvroSchema') -> 'BulkWriterFactory': jvm = get_gateway().jvm JAvroParquetWriters = jvm.org.apache.flink.formats.parquet.avro.AvroParquetWriters return BulkWriterFactory(JAvroParquetWriters.forGenericRecord(schema._j_schema)) + + +class ParquetRowDataWriter(object): Review Comment: Could we group the classes in this file a bit to make it more readable? e.g. in the following order: AvroParquetReaders, AvroParquetWriters, ParquetColumnarRowInputFormat, ParquetBulkWriter. ## flink-python/pyflink/datastream/formats/parquet.py: ## @@ -155,3 +158,54 @@ def for_generic_record(schema: 'AvroSchema') -> 'BulkWriterFactory': jvm = get_gateway().jvm JAvroParquetWriters = jvm.org.apache.flink.formats.parquet.avro.AvroParquetWriters return BulkWriterFactory(JAvroParquetWriters.forGenericRecord(schema._j_schema)) + + +class ParquetRowDataWriter(object): +""" +Convenient builder to create a :class:`BulkWriterFactory` that writes Rows with a defined +:class:`RowType` into Parquet files in a batch fashion. + +.. versionadded:: 1.16.0 +""" + +@staticmethod +def for_row_type(row_type: RowType, hadoop_config: Optional[Configuration] = None, + utc_timestamp: bool = False) -> 'BulkWriterFactory': +""" +Create a :class:`RowDataBulkWriterFactory` that writes Rows records with a defined Review Comment: ```suggestion Create a :class:`pyflink.datastream.connectors.file_system.RowDataBulkWriterFactory` that writes Rows records with a defined ``` Need to make sure it could generate a valid link in the documentation. ## flink-python/pyflink/datastream/formats/parquet.py: ## @@ -155,3 +158,54 @@ def for_generic_record(schema: 'AvroSchema') -> 'BulkWriterFactory': jvm = get_gateway().jvm JAvroParquetWriters = jvm.org.apache.flink.formats.parquet.avro.AvroParquetWriters return BulkWriterFactory(JAvroParquetWriters.forGenericRecord(schema._j_schema)) + + +class ParquetRowDataWriter(object): +""" +Convenient builder to create a :class:`BulkWriterFactory` that writes Rows with a defined +:class:`RowType` into Parquet files in a batch fashion. + +.. versionadded:: 1.16.0 +""" + +@staticmethod +def for_row_type(row_type: RowType, hadoop_config: Optional[Configuration] = None, + utc_timestamp: bool = False) -> 'BulkWriterFactory': +""" +Create a :class:`RowDataBulkWriterFactory` that writes Rows records with a defined +:class:`RowType` into Parquet files in a batch fashion. + +Example: +:: + +>>> row_type = DataTypes.ROW([ +... DataTypes.FIELD('string', DataTypes.STRING()), +... DataTypes.FIELD('int_array', DataTypes.ARRAY(DataTypes.INT())) +... ]) +>>> row_type_info = Types.ROW_NAMED( +... ['string', 'int_array'], +... [Types.STRING(), Types.LIST(Types.INT())] +... ) +>>> sink = FileSink.for_bulk_format( +... OUTPUT_DIR, ParquetRowDataWriter.for_row_type( +... row_type, +... hadoop_config=Configuration(), +... utc_timestamp=True, +... ) +... ).build() Review Comment: ```suggestion ... ) ``` ##
[GitHub] [flink] Myasuka commented on a diff in pull request #20421: [FLINK-28732][state] Deprecate ambiguous StateTTLConfig#cleanFullSnapshot API
Myasuka commented on code in PR #20421: URL: https://github.com/apache/flink/pull/20421#discussion_r940868846 ## docs/content/docs/dev/datastream/fault-tolerance/state.md: ## @@ -474,9 +474,10 @@ ttl_config = StateTtlConfig \ For more fine-grained control over some special cleanup in background, you can configure it separately as described below. Currently, heap state backend relies on incremental cleanup and RocksDB backend uses compaction filter for background cleanup. -# Cleanup in full snapshot +# Cleanup in full scan snapshot -Additionally, you can activate the cleanup at the moment of taking the full state snapshot which +Additionally, you can activate the cleanup at the moment of taking the full scan state snapshot (including the canonical savepoint, +or the full/incremental checkpoint of hashmap state-backend, or the full checkpoint of RocksDB state-backend), which Review Comment: Got it, I will refactor these descriptions. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Created] (FLINK-28881) PubSubConsumingTest.testStoppingConnectorWhenDeserializationSchemaIndicatesEndOfStream test failure
Leonard Xu created FLINK-28881: -- Summary: PubSubConsumingTest.testStoppingConnectorWhenDeserializationSchemaIndicatesEndOfStream test failure Key: FLINK-28881 URL: https://issues.apache.org/jira/browse/FLINK-28881 Project: Flink Issue Type: Bug Components: Connectors / Google Cloud PubSub Affects Versions: 1.16.0 Reporter: Leonard Xu CI link: https://dev.azure.com/leonardBang/Azure_CI/_build/results?buildId=713=logs=3796201e-ea88-5776-0ea8-9ccca648a70c=8ca54b76-085e-5cf1-8060-2c500a258258 -- This message was sent by Atlassian Jira (v8.20.10#820010)
[GitHub] [flink] zoltar9264 commented on pull request #20152: [FLINK-27155][changelog] Reduce multiple reads to the same Changelog …
zoltar9264 commented on PR #20152: URL: https://github.com/apache/flink/pull/20152#issuecomment-1208878901 Thanks @rkhachatryan ! I'm very grateful for your help over the past few weeks, I've learned a lot in it, and I'm sorry for the trouble I've caused you due to my inexperience. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-27691) RankHarnessTest. testUpdateRankWithRowNumberSortKeyDropsToNotLast test failed with result mismatch
[ https://issues.apache.org/jira/browse/FLINK-27691?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Leonard Xu updated FLINK-27691: --- Priority: Critical (was: Major) > RankHarnessTest. testUpdateRankWithRowNumberSortKeyDropsToNotLast test failed > with result mismatch > -- > > Key: FLINK-27691 > URL: https://issues.apache.org/jira/browse/FLINK-27691 > Project: Flink > Issue Type: Bug > Components: Table SQL / Runtime >Affects Versions: 1.16.0 >Reporter: Huang Xingbo >Assignee: lincoln lee >Priority: Critical > Labels: test-stability > > {code:java} > 2022-05-19T04:34:04.2677138Z May 19 04:34:04 [ERROR] > RankHarnessTest.testUpdateRankWithRowNumberSortKeyDropsToNotLast > 2022-05-19T04:34:04.2689553Z May 19 04:34:04 [ERROR] Run 1: [result > mismatch] > 2022-05-19T04:34:04.2690614Z May 19 04:34:04 expected: [+I(a,1,100,1), > 2022-05-19T04:34:04.2691128Z May 19 04:34:04 +I(b,1,90,2), > 2022-05-19T04:34:04.2691552Z May 19 04:34:04 +I(c,1,90,3), > 2022-05-19T04:34:04.2692235Z May 19 04:34:04 +I(d,1,80,4), > 2022-05-19T04:34:04.2692634Z May 19 04:34:04 +I(e,1,80,5), > 2022-05-19T04:34:04.2693060Z May 19 04:34:04 +I(f,1,70,6), > 2022-05-19T04:34:04.2693468Z May 19 04:34:04 +U(b,1,80,5), > 2022-05-19T04:34:04.2693874Z May 19 04:34:04 +U(c,1,90,2), > 2022-05-19T04:34:04.2694282Z May 19 04:34:04 +U(d,1,80,3), > 2022-05-19T04:34:04.2694670Z May 19 04:34:04 +U(e,1,80,4), > 2022-05-19T04:34:04.2696097Z May 19 04:34:04 -U(b,1,90,2), > 2022-05-19T04:34:04.2696718Z May 19 04:34:04 -U(c,1,90,3), > 2022-05-19T04:34:04.2697298Z May 19 04:34:04 -U(d,1,80,4), > 2022-05-19T04:34:04.2698102Z May 19 04:34:04 -U(e,1,80,5)] > 2022-05-19T04:34:04.2698758Z May 19 04:34:04 but was: [+I(a,1,100,1), > 2022-05-19T04:34:04.2699189Z May 19 04:34:04 +I(b,1,90,1), > 2022-05-19T04:34:04.2699607Z May 19 04:34:04 +I(c,1,90,2), > 2022-05-19T04:34:04.2700017Z May 19 04:34:04 +I(d,1,80,3), > 2022-05-19T04:34:04.2712164Z May 19 04:34:04 +I(e,1,80,4), > 2022-05-19T04:34:04.2712777Z May 19 04:34:04 +I(f,1,70,5), > 2022-05-19T04:34:04.2713191Z May 19 04:34:04 +U(b,1,80,4), > 2022-05-19T04:34:04.2713621Z May 19 04:34:04 +U(c,1,90,1), > 2022-05-19T04:34:04.2714029Z May 19 04:34:04 +U(d,1,80,2), > 2022-05-19T04:34:04.2714435Z May 19 04:34:04 +U(e,1,80,3), > 2022-05-19T04:34:04.2715272Z May 19 04:34:04 -U(b,1,90,1), > 2022-05-19T04:34:04.2715847Z May 19 04:34:04 -U(c,1,90,2), > 2022-05-19T04:34:04.2716420Z May 19 04:34:04 -U(d,1,80,3), > 2022-05-19T04:34:04.2716990Z May 19 04:34:04 -U(e,1,80,4)] > {code} > https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=35815=logs=a9db68b9-a7e0-54b6-0f98-010e0aff39e2=cdd32e0b-6047-565b-c58f-14054472f1be=10445 -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (FLINK-27691) RankHarnessTest. testUpdateRankWithRowNumberSortKeyDropsToNotLast test failed with result mismatch
[ https://issues.apache.org/jira/browse/FLINK-27691?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17577122#comment-17577122 ] Leonard Xu commented on FLINK-27691: https://dev.azure.com/leonardBang/Azure_CI/_build/results?buildId=713=ms.vss-test-web.build-test-results-tab=20042=111227=debug > RankHarnessTest. testUpdateRankWithRowNumberSortKeyDropsToNotLast test failed > with result mismatch > -- > > Key: FLINK-27691 > URL: https://issues.apache.org/jira/browse/FLINK-27691 > Project: Flink > Issue Type: Bug > Components: Table SQL / Runtime >Affects Versions: 1.16.0 >Reporter: Huang Xingbo >Assignee: lincoln lee >Priority: Major > Labels: test-stability > > {code:java} > 2022-05-19T04:34:04.2677138Z May 19 04:34:04 [ERROR] > RankHarnessTest.testUpdateRankWithRowNumberSortKeyDropsToNotLast > 2022-05-19T04:34:04.2689553Z May 19 04:34:04 [ERROR] Run 1: [result > mismatch] > 2022-05-19T04:34:04.2690614Z May 19 04:34:04 expected: [+I(a,1,100,1), > 2022-05-19T04:34:04.2691128Z May 19 04:34:04 +I(b,1,90,2), > 2022-05-19T04:34:04.2691552Z May 19 04:34:04 +I(c,1,90,3), > 2022-05-19T04:34:04.2692235Z May 19 04:34:04 +I(d,1,80,4), > 2022-05-19T04:34:04.2692634Z May 19 04:34:04 +I(e,1,80,5), > 2022-05-19T04:34:04.2693060Z May 19 04:34:04 +I(f,1,70,6), > 2022-05-19T04:34:04.2693468Z May 19 04:34:04 +U(b,1,80,5), > 2022-05-19T04:34:04.2693874Z May 19 04:34:04 +U(c,1,90,2), > 2022-05-19T04:34:04.2694282Z May 19 04:34:04 +U(d,1,80,3), > 2022-05-19T04:34:04.2694670Z May 19 04:34:04 +U(e,1,80,4), > 2022-05-19T04:34:04.2696097Z May 19 04:34:04 -U(b,1,90,2), > 2022-05-19T04:34:04.2696718Z May 19 04:34:04 -U(c,1,90,3), > 2022-05-19T04:34:04.2697298Z May 19 04:34:04 -U(d,1,80,4), > 2022-05-19T04:34:04.2698102Z May 19 04:34:04 -U(e,1,80,5)] > 2022-05-19T04:34:04.2698758Z May 19 04:34:04 but was: [+I(a,1,100,1), > 2022-05-19T04:34:04.2699189Z May 19 04:34:04 +I(b,1,90,1), > 2022-05-19T04:34:04.2699607Z May 19 04:34:04 +I(c,1,90,2), > 2022-05-19T04:34:04.2700017Z May 19 04:34:04 +I(d,1,80,3), > 2022-05-19T04:34:04.2712164Z May 19 04:34:04 +I(e,1,80,4), > 2022-05-19T04:34:04.2712777Z May 19 04:34:04 +I(f,1,70,5), > 2022-05-19T04:34:04.2713191Z May 19 04:34:04 +U(b,1,80,4), > 2022-05-19T04:34:04.2713621Z May 19 04:34:04 +U(c,1,90,1), > 2022-05-19T04:34:04.2714029Z May 19 04:34:04 +U(d,1,80,2), > 2022-05-19T04:34:04.2714435Z May 19 04:34:04 +U(e,1,80,3), > 2022-05-19T04:34:04.2715272Z May 19 04:34:04 -U(b,1,90,1), > 2022-05-19T04:34:04.2715847Z May 19 04:34:04 -U(c,1,90,2), > 2022-05-19T04:34:04.2716420Z May 19 04:34:04 -U(d,1,80,3), > 2022-05-19T04:34:04.2716990Z May 19 04:34:04 -U(e,1,80,4)] > {code} > https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=35815=logs=a9db68b9-a7e0-54b6-0f98-010e0aff39e2=cdd32e0b-6047-565b-c58f-14054472f1be=10445 -- This message was sent by Atlassian Jira (v8.20.10#820010)
[GitHub] [flink] JesseAtSZ commented on pull request #20091: [FLINK-27570][runtime] Fix checkFailureCounter in CheckpointFailureManager for finalize failure
JesseAtSZ commented on PR #20091: URL: https://github.com/apache/flink/pull/20091#issuecomment-1208875024 @rkhachatryan -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Comment Edited] (FLINK-28880) Fix CEP doc with wrong result of strict contiguity of looping patterns
[ https://issues.apache.org/jira/browse/FLINK-28880?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17577117#comment-17577117 ] Huang Xingbo edited comment on FLINK-28880 at 8/9/22 3:39 AM: -- Merged into master via a08b050eb9fbea319275771fd9e95bbb025e2737 Merged into release-1.15 via 6eced8aa39c9e16b9918bd8b05fed1a1e17b6fe7 Merged into release-1.14 via c5108f6ab078ffbf1cbc483ad469197988e5553b was (Author: hxbks2ks): Merged into master via a08b050eb9fbea319275771fd9e95bbb025e2737 > Fix CEP doc with wrong result of strict contiguity of looping patterns > -- > > Key: FLINK-28880 > URL: https://issues.apache.org/jira/browse/FLINK-28880 > Project: Flink > Issue Type: Bug > Components: Library / CEP >Affects Versions: 1.15.1 >Reporter: Juntao Hu >Priority: Minor > Labels: pull-request-available > Fix For: 1.16.0, 1.15.2, 1.14.6 > > > https://nightlies.apache.org/flink/flink-docs-release-1.15/docs/libs/cep/#contiguity-within-looping-patterns > The result of strict contiguity should be {a b1 c}, {a b2 c}, {a b3 c}, since > b is *followed by* c. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (FLINK-28880) Fix CEP doc with wrong result of strict contiguity of looping patterns
[ https://issues.apache.org/jira/browse/FLINK-28880?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Huang Xingbo updated FLINK-28880: - Fix Version/s: 1.15.2 1.14.6 > Fix CEP doc with wrong result of strict contiguity of looping patterns > -- > > Key: FLINK-28880 > URL: https://issues.apache.org/jira/browse/FLINK-28880 > Project: Flink > Issue Type: Bug > Components: Library / CEP >Affects Versions: 1.15.1 >Reporter: Juntao Hu >Priority: Minor > Labels: pull-request-available > Fix For: 1.16.0, 1.15.2, 1.14.6 > > > https://nightlies.apache.org/flink/flink-docs-release-1.15/docs/libs/cep/#contiguity-within-looping-patterns > The result of strict contiguity should be {a b1 c}, {a b2 c}, {a b3 c}, since > b is *followed by* c. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Resolved] (FLINK-28880) Fix CEP doc with wrong result of strict contiguity of looping patterns
[ https://issues.apache.org/jira/browse/FLINK-28880?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Huang Xingbo resolved FLINK-28880. -- Resolution: Fixed Merged into master via a08b050eb9fbea319275771fd9e95bbb025e2737 > Fix CEP doc with wrong result of strict contiguity of looping patterns > -- > > Key: FLINK-28880 > URL: https://issues.apache.org/jira/browse/FLINK-28880 > Project: Flink > Issue Type: Bug > Components: Library / CEP >Affects Versions: 1.15.1 >Reporter: Juntao Hu >Priority: Minor > Labels: pull-request-available > Fix For: 1.16.0 > > > https://nightlies.apache.org/flink/flink-docs-release-1.15/docs/libs/cep/#contiguity-within-looping-patterns > The result of strict contiguity should be {a b1 c}, {a b2 c}, {a b3 c}, since > b is *followed by* c. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[GitHub] [flink] HuangXingBo closed pull request #20508: [FLINK-28880][docs][cep] Fix wrong result of strict contiguity of looping patterns
HuangXingBo closed pull request #20508: [FLINK-28880][docs][cep] Fix wrong result of strict contiguity of looping patterns URL: https://github.com/apache/flink/pull/20508 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] leonardBang commented on a diff in pull request #20495: [FLINK-28868][connector/hbase] Migrate HBase table connector to the new LookupFunction interface
leonardBang commented on code in PR #20495: URL: https://github.com/apache/flink/pull/20495#discussion_r940855756 ## flink-connectors/flink-connector-hbase-1.4/src/main/java/org/apache/flink/connector/hbase1/source/HBaseDynamicTableSource.java: ## @@ -38,23 +39,19 @@ public HBaseDynamicTableSource( String tableName, HBaseTableSchema hbaseSchema, String nullStringLiteral, -HBaseLookupOptions lookupOptions) { -super(conf, tableName, hbaseSchema, nullStringLiteral, lookupOptions); +int maxRetryTimes, +@Nullable LookupCache cache) { +super(conf, tableName, hbaseSchema, nullStringLiteral, maxRetryTimes, cache); } @Override public DynamicTableSource copy() { return new HBaseDynamicTableSource( -conf, tableName, hbaseSchema, nullStringLiteral, lookupOptions); +conf, tableName, hbaseSchema, nullStringLiteral, maxRetryTimes, cache); } @Override Review Comment: Add missed `equals()` and `hashcode()` methods? ## flink-connectors/flink-connector-hbase-1.4/src/test/java/org/apache/flink/connector/hbase1/HBaseDynamicTableFactoryTest.java: ## @@ -141,6 +143,32 @@ public void testTableSourceFactory() { hbaseSchema.getQualifierDataTypes("f4")); } +@Test +public void testCacheOptions() { Review Comment: ```suggestion public void testLookupOptions() { ``` `lookup.max-retries` is not related to cachiung. ## flink-connectors/flink-connector-hbase-2.2/src/test/java/org/apache/flink/connector/hbase2/HBaseDynamicTableFactoryTest.java: ## @@ -144,6 +146,32 @@ public void testTableSourceFactory() { hbaseSchema.getQualifierDataTypes("f4")); } +@Test +public void testCacheOptions() { Review Comment: ```suggestion public void testLookupOptions() { ``` ## flink-connectors/flink-connector-hbase-1.4/src/main/java/org/apache/flink/connector/hbase1/HBase1DynamicTableFactory.java: ## @@ -78,13 +81,31 @@ public DynamicTableSource createDynamicTableSource(Context context) { String nullStringLiteral = tableOptions.get(NULL_STRING_LITERAL); HBaseTableSchema hbaseSchema = HBaseTableSchema.fromDataType(context.getPhysicalRowDataType()); +LookupCache cache = null; + +// Backward compatible to legacy caching options +if (tableOptions.get(LOOKUP_CACHE_MAX_ROWS) > 0 Review Comment: Could we use fallbackKeys way? ## flink-connectors/flink-connector-hbase-base/src/main/java/org/apache/flink/connector/hbase/table/HBaseConnectorOptions.java: ## @@ -93,6 +94,8 @@ public class HBaseConnectorOptions { .defaultValue(false) .withDescription("whether to set async lookup."); +/** @deprecated Please use {@link LookupOptions#PARTIAL_CACHE_MAX_ROWS} instead. */ +@Deprecated public static final ConfigOption LOOKUP_CACHE_MAX_ROWS = Review Comment: add fallbackKeys ## flink-connectors/flink-connector-hbase-base/src/main/java/org/apache/flink/connector/hbase/source/HBaseRowDataLookupFunction.java: ## @@ -68,54 +65,37 @@ public class HBaseRowDataLookupFunction extends TableFunction { private transient HTable table; private transient HBaseSerde serde; -private final long cacheMaxSize; -private final long cacheExpireMs; private final int maxRetryTimes; -private transient Cache cache; public HBaseRowDataLookupFunction( Configuration configuration, String hTableName, HBaseTableSchema hbaseTableSchema, String nullStringLiteral, -HBaseLookupOptions lookupOptions) { +int maxRetryTimes) { this.serializedConfig = HBaseConfigurationUtil.serializeConfiguration(configuration); this.hTableName = hTableName; this.hbaseTableSchema = hbaseTableSchema; this.nullStringLiteral = nullStringLiteral; -this.cacheMaxSize = lookupOptions.getCacheMaxSize(); -this.cacheExpireMs = lookupOptions.getCacheExpireMs(); -this.maxRetryTimes = lookupOptions.getMaxRetryTimes(); +this.maxRetryTimes = maxRetryTimes; } /** * The invoke entry point of lookup function. * - * @param rowKey the lookup key. Currently only support single rowkey. + * @param keyRow - A {@link RowData} that wraps lookup keys. Currently only support single + * rowkey. */ -public void eval(Object rowKey) throws IOException { -if (cache != null) { -RowData cacheRowData = cache.getIfPresent(rowKey); -if (cacheRowData != null) { -collect(cacheRowData); -return; -} -} +@Override +public Collection lookup(RowData keyRow)
[GitHub] [flink] flinkbot commented on pull request #20508: [FLINK-28880][docs][cep] Fix wrong result of strict contiguity of looping patterns
flinkbot commented on PR #20508: URL: https://github.com/apache/flink/pull/20508#issuecomment-1208871012 ## CI report: * f24ce7e938ce1740d9e47d3d620697a9b9d24bcd UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-28803) Add Transformer and Estimator for KBinsDiscretizer
[ https://issues.apache.org/jira/browse/FLINK-28803?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated FLINK-28803: --- Labels: pull-request-available (was: ) > Add Transformer and Estimator for KBinsDiscretizer > -- > > Key: FLINK-28803 > URL: https://issues.apache.org/jira/browse/FLINK-28803 > Project: Flink > Issue Type: New Feature > Components: Library / Machine Learning >Affects Versions: ml-2.2.0 >Reporter: Zhipeng Zhang >Assignee: Zhipeng Zhang >Priority: Major > Labels: pull-request-available > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Comment Edited] (FLINK-28838) Avoid to notify the elementQueue consumer when the fetch result is empty
[ https://issues.apache.org/jira/browse/FLINK-28838?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17577116#comment-17577116 ] Qingsheng Ren edited comment on FLINK-28838 at 8/9/22 3:30 AM: --- Thanks for the ticket [~aitozi]! Yeah we can definitely make some improvement as not all source implementations works as expected. I think your first proposal make sense to me. We can drop empty records earlier before putting into elementsQueue. I have some concerns about the second one (adding SleepTask) as we can hardly decide the length of sleep considering source implementations differ a lot. For example KafkaConsumer itself has ability to block the thread if no data is available for polling so it doesn't need the SleepTask at all. I prefer to leave it to split reader implementation itself as the doc of {{SplitReader#fetch}} is quite clear that it could be a blocking call. WDYT? BTW which source has this issue? We can check its implementation too. was (Author: renqs): Thanks for the ticket [~aitozi]! Yeah we can definitely make some improvement as not all source implementations works as expected. I think your first proposal make sense to me. We can drop empty records earlier before putting into elementsQueue. I have some concerns about the second one (adding SleepTask) as we can hardly decide the length of sleep considering source implementations vary a lot. For example KafkaConsumer itself has ability to block the thread if no data is available for polling so it doesn't need the SleepTask at all. I prefer to leave it to split reader implementation itself as the doc of {{SplitReader#fetch}} is quite clear that it could be a blocking call. WDYT? BTW which source has this issue? We can check its implementation too. > Avoid to notify the elementQueue consumer when the fetch result is empty > > > Key: FLINK-28838 > URL: https://issues.apache.org/jira/browse/FLINK-28838 > Project: Flink > Issue Type: Improvement > Components: Connectors / Common >Affects Versions: 1.15.0, 1.15.1 >Reporter: Aitozi >Priority: Major > Fix For: 1.16.0 > > Attachments: 20220805165441.jpg > > > When using the new source api, I found that if the source has no data, it > still brings high cpu usage. > The reason behind this is that it will always return the > {{RecordsWithSplitIds}} from the {{splitReader.fetch}} in FetchTask and it > will be added to the elementQueue. It will make the consumer be notified to > wake up frequently. > This causes the thread to keep busy to run and wake up, which leads to the > high sys and user cpu usage. > I think not all the SplitReader#fetch will block until there is data, if it > returns immediately when there is no data, then this problem will happen -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (FLINK-28838) Avoid to notify the elementQueue consumer when the fetch result is empty
[ https://issues.apache.org/jira/browse/FLINK-28838?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17577116#comment-17577116 ] Qingsheng Ren commented on FLINK-28838: --- Thanks for the ticket [~aitozi]! Yeah we can definitely make some improvement as not all source implementations works as expected. I think your first proposal make sense to me. We can drop empty records earlier before putting into elementsQueue. I have some concerns about the second one (adding SleepTask) as we can hardly decide the length of sleep considering source implementations vary a lot. For example KafkaConsumer itself has ability to block the thread if no data is available for polling so it doesn't need the SleepTask at all. I prefer to leave it to split reader implementation itself as the doc of {{SplitReader#fetch}} is quite clear that it could be a blocking call. WDYT? BTW which source has this issue? We can check its implementation too. > Avoid to notify the elementQueue consumer when the fetch result is empty > > > Key: FLINK-28838 > URL: https://issues.apache.org/jira/browse/FLINK-28838 > Project: Flink > Issue Type: Improvement > Components: Connectors / Common >Affects Versions: 1.15.0, 1.15.1 >Reporter: Aitozi >Priority: Major > Fix For: 1.16.0 > > Attachments: 20220805165441.jpg > > > When using the new source api, I found that if the source has no data, it > still brings high cpu usage. > The reason behind this is that it will always return the > {{RecordsWithSplitIds}} from the {{splitReader.fetch}} in FetchTask and it > will be added to the elementQueue. It will make the consumer be notified to > wake up frequently. > This causes the thread to keep busy to run and wake up, which leads to the > high sys and user cpu usage. > I think not all the SplitReader#fetch will block until there is data, if it > returns immediately when there is no data, then this problem will happen -- This message was sent by Atlassian Jira (v8.20.10#820010)
[GitHub] [flink-ml] zhipeng93 opened a new pull request, #139: [FLINK-28803] Add Transformer and Estimator for KBinsDiscretizer
zhipeng93 opened a new pull request, #139: URL: https://github.com/apache/flink-ml/pull/139 ## What is the purpose of the change - Add Transformer and Estimator for KBinsDiscretizer[1] in Flink ML. ## Brief change log - Added Transformer and Estimator for LinearSVC. - Added java test/example for Transformer and Estimator for KBinsDiscretizer. - Added python source/test/example for KBinsDiscretizer. - Comparing with Sklearn [1], we made the following changes: - We removed `encode` parameter since we can achieve the same goal with a onehot-encoder following KBinsDiscretizer. - We removed `randomState` since it is non-trivial to support reproduciable sampling in a distributed setting, while in sklearn (single-node) setting, it is much easier. If the users would like this feature, we can add it later. ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): (no) - The public API, i.e., is any changed class annotated with @Public(Evolving): (no) - Does this pull request introduce a new feature? (yes) - If yes, how is the feature documented? (Java doc) [1] https://scikit-learn.org/stable/modules/generated/sklearn.preprocessing.KBinsDiscretizer.html -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Resolved] (FLINK-28699) Native rocksdb full snapshot in non-incremental checkpointing
[ https://issues.apache.org/jira/browse/FLINK-28699?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Yun Tang resolved FLINK-28699. -- Fix Version/s: 1.16.0 Resolution: Fixed merged in master: d9a067e5e1c8672930b0ea7d76400a1d3020a1e2 > Native rocksdb full snapshot in non-incremental checkpointing > - > > Key: FLINK-28699 > URL: https://issues.apache.org/jira/browse/FLINK-28699 > Project: Flink > Issue Type: Improvement > Components: Runtime / State Backends >Affects Versions: 1.14.5, 1.15.1 >Reporter: Lihe Ma >Assignee: Lihe Ma >Priority: Major > Labels: pull-request-available > Fix For: 1.16.0 > > > When rocksdb statebackend is used and state.backend.incremental enabled, > flink will figure out newly created sst files generated by rocksdb during > checkpoint, and read all the states from rocksdb and write to files during > savepoint [1]. > When state.backend.incremental disabled, flink will read all the states from > rocksdb and generate state files in checkpoint and savepoint [2]. This makes > sense in savepoint, cause user can take a savepoint with rocksdb statebackend > and then restore it using another statebackend, but in checkpoint, > deserialisation and serialisation of state results in performance loss. > If the native rocksdb snapshot is introduced in full snapshot, theoretically > better performance can be achieved. At the same time, savepoint remains the > same as before. > > # > https://github.com/apache/flink/blob/master/flink-state-backends/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/snapshot/RocksIncrementalSnapshotStrategy.java > # > https://github.com/apache/flink/blob/master/flink-state-backends/flink-statebackend-rocksdb/src/main/java/org/apache/flink/contrib/streaming/state/snapshot/RocksFullSnapshotStrategy.java -- This message was sent by Atlassian Jira (v8.20.10#820010)
[GitHub] [flink] fsk119 commented on a diff in pull request #20493: [FLINK-28632][sql-gateway][hive] Allow to GetColumns/GetPrimaryKeys/GetTableTypes in the HiveServer2 Endpoint.
fsk119 commented on code in PR #20493: URL: https://github.com/apache/flink/pull/20493#discussion_r940202528 ## flink-connectors/flink-connector-hive/src/main/java/org/apache/flink/table/endpoint/hive/util/OperationExecutorFactory.java: ## @@ -294,4 +408,64 @@ private static List getSupportedHiveType() { INTERVAL_YEAR_MONTH_TYPE, INTERVAL_DAY_TIME_TYPE)); } + +/** + * The column size for this type. For numeric data this is the maximum precision. For character + * data this is the length in characters. For datetime types this is the length in characters of + * the String representation (assuming the maximum allowed precision of the fractional seconds + * component). For binary data this is the length in bytes. Null is returned for for data types + * where the column size is not applicable. + */ +// TODO +private static Integer getColumnSize(Type hiveColumnType, LogicalType flinkColumnType) { +if (hiveColumnType.isNumericType()) { +// Exactly precision for DECIMAL_TYPE and maximum precision for others. +return hiveColumnType == Type.DECIMAL_TYPE +? ((DecimalType) flinkColumnType).getPrecision() +: hiveColumnType.getMaxPrecision(); +} +switch (hiveColumnType) { +case STRING_TYPE: +case BINARY_TYPE: +return Integer.MAX_VALUE; +case CHAR_TYPE: +case VARCHAR_TYPE: +return TypeInfoUtils.getCharacterLengthForType( +getPrimitiveTypeInfo(hiveColumnType.getName())); +case DATE_TYPE: +return 10; +case TIMESTAMP_TYPE: +return 29; +// case TIMESTAMPLOCALTZ_TYPE: +// return 31; +// 还是用flinkColumnType来实现? +default: +return null; +} +} + +/** + * The number of fractional digits for this type. Null is returned for data types where this is + * not applicable. + */ +private static Integer getDecimalDigits(Type hiveColumnType, LogicalType flinkColumnType) { Review Comment: ditto ## flink-connectors/flink-connector-hive/src/main/java/org/apache/flink/table/endpoint/hive/util/OperationExecutorFactory.java: ## @@ -294,4 +408,64 @@ private static List getSupportedHiveType() { INTERVAL_YEAR_MONTH_TYPE, INTERVAL_DAY_TIME_TYPE)); } + +/** + * The column size for this type. For numeric data this is the maximum precision. For character + * data this is the length in characters. For datetime types this is the length in characters of + * the String representation (assuming the maximum allowed precision of the fractional seconds + * component). For binary data this is the length in bytes. Null is returned for for data types + * where the column size is not applicable. + */ +// TODO +private static Integer getColumnSize(Type hiveColumnType, LogicalType flinkColumnType) { Review Comment: I think it's better to use LogicalType only because the two inputs are equal. Add `@Nullable` for the return type. ## flink-connectors/flink-connector-hive/src/main/java/org/apache/flink/table/endpoint/hive/util/OperationExecutorFactory.java: ## @@ -203,10 +229,98 @@ private static ResultSet executeGetTables( .collect(Collectors.toList())); } +private static ResultSet executeGetColumns( +SqlGatewayService service, +SessionHandle sessionHandle, +@Nullable String catalogName, +@Nullable String schemaName, +@Nullable String tableName, +@Nullable String columnName) { +String specifiedCatalogName = +isNullOrEmpty(catalogName) ? service.getCurrentCatalog(sessionHandle) : catalogName; +Set schemaNames = +filter(service.listDatabases(sessionHandle, specifiedCatalogName), schemaName); +Set tableKinds = new HashSet<>(Arrays.asList(TableKind.TABLE, TableKind.VIEW)); +List rowData = new ArrayList<>(); + +for (String schema : schemaNames) { +Set tableInfos = +filter( +service.listTables( +sessionHandle, specifiedCatalogName, schema, tableKinds), +candidates -> candidates.getIdentifier().getObjectName(), +tableName); + +for (TableInfo tableInfo : tableInfos) { +ResolvedCatalogBaseTable table = +service.getTable(sessionHandle, tableInfo.getIdentifier()); +List columns = table.getResolvedSchema().getColumns(); + +Set requiredColumnNames = +
[GitHub] [flink] bgeng777 commented on a diff in pull request #20508: [FLINK-28880][docs][cep] Fix wrong result of strict contiguity of looping patterns
bgeng777 commented on code in PR #20508: URL: https://github.com/apache/flink/pull/20508#discussion_r940855754 ## docs/content/docs/libs/cep.md: ## @@ -754,7 +754,7 @@ The contiguity will be applied between elements accepted into such a pattern. To illustrate the above with an example, a pattern sequence `"a b+ c"` (`"a"` followed by any(non-deterministic relaxed) sequence of one or more `"b"`'s followed by a `"c"`) with input `"a", "b1", "d1", "b2", "d2", "b3" "c"` will have the following results: - 1. **Strict Contiguity**: `{a b3 c}` -- the `"d1"` after `"b1"` causes `"b1"` to be discarded, the same happens for `"b2"` because of `"d2"`. + 1. **Strict Contiguity**: `{a b1 c}`, `{a b2 c}`, `{a b3 c}` - there's no adjacent `"b"`s. Review Comment: there's -> there are -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] Myasuka closed pull request #20399: [FLINK-28699][StateBackend] Native rocksdb full snapshot in non-incremental checkpointing
Myasuka closed pull request #20399: [FLINK-28699][StateBackend] Native rocksdb full snapshot in non-incremental checkpointing URL: https://github.com/apache/flink/pull/20399 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] reswqa commented on a diff in pull request #20333: [FLINK-28623][network] Optimize the use of off heap memory by blocking and hybrid shuffle reader
reswqa commented on code in PR #20333: URL: https://github.com/apache/flink/pull/20333#discussion_r940855012 ## flink-runtime/src/main/java/org/apache/flink/runtime/io/network/partition/SortMergeResultPartitionReadScheduler.java: ## @@ -70,6 +71,17 @@ class SortMergeResultPartitionReadScheduler implements Runnable, BufferRecycler */ private static final Duration DEFAULT_BUFFER_REQUEST_TIMEOUT = Duration.ofMinutes(5); +/** Used to read buffers from file channel. */ Review Comment: fixed. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Closed] (FLINK-28568) Implements a new lookup join operator (sync mode only) with state to eliminate the non determinism
[ https://issues.apache.org/jira/browse/FLINK-28568?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] godfrey he closed FLINK-28568. -- Resolution: Fixed Fixed in master: 3f18cafa0581613ef9900da0478b3501617dc64f > Implements a new lookup join operator (sync mode only) with state to > eliminate the non determinism > -- > > Key: FLINK-28568 > URL: https://issues.apache.org/jira/browse/FLINK-28568 > Project: Flink > Issue Type: Sub-task > Components: Table SQL / Runtime >Reporter: lincoln lee >Assignee: lincoln lee >Priority: Major > Labels: pull-request-available > Fix For: 1.16.0 > > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Assigned] (FLINK-28568) Implements a new lookup join operator (sync mode only) with state to eliminate the non determinism
[ https://issues.apache.org/jira/browse/FLINK-28568?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] godfrey he reassigned FLINK-28568: -- Assignee: lincoln lee > Implements a new lookup join operator (sync mode only) with state to > eliminate the non determinism > -- > > Key: FLINK-28568 > URL: https://issues.apache.org/jira/browse/FLINK-28568 > Project: Flink > Issue Type: Sub-task > Components: Table SQL / Runtime >Reporter: lincoln lee >Assignee: lincoln lee >Priority: Major > Labels: pull-request-available > Fix For: 1.16.0 > > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[GitHub] [flink] bgeng777 commented on pull request #20508: [FLINK-28880][docs][cep] Fix wrong result of strict contiguity of looping patterns
bgeng777 commented on PR #20508: URL: https://github.com/apache/flink/pull/20508#issuecomment-1208865870 LGTM -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] godfreyhe closed pull request #20324: [FLINK-28568][table-runtime] Implements a new lookup join operator (sync mode only) with state to eliminate the non determinism
godfreyhe closed pull request #20324: [FLINK-28568][table-runtime] Implements a new lookup join operator (sync mode only) with state to eliminate the non determinism URL: https://github.com/apache/flink/pull/20324 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] leonardBang commented on a diff in pull request #20489: [FLINK-28854][connector/jdbc] Migrate JDBC lookup table to the new LookupFunction and caching interface
leonardBang commented on code in PR #20489: URL: https://github.com/apache/flink/pull/20489#discussion_r940843642 ## flink-connectors/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/table/JdbcDynamicTableFactory.java: ## @@ -181,6 +179,27 @@ private JdbcDmlOptions getJdbcDmlOptions( .build(); } +@Nullable +private LookupCache getLookupCache(ReadableConfig tableOptions) { +LookupCache cache = null; +// Legacy cache options +if (tableOptions.get(LOOKUP_CACHE_MAX_ROWS) > 0 +&& tableOptions.get(LOOKUP_CACHE_TTL).compareTo(Duration.ZERO) > 0) { +cache = +DefaultLookupCache.newBuilder() + .maximumSize(tableOptions.get(LOOKUP_CACHE_MAX_ROWS)) + .expireAfterWrite(tableOptions.get(LOOKUP_CACHE_TTL)) + .cacheMissingKey(tableOptions.get(LOOKUP_CACHE_MISSING_KEY)) +.build(); +} Review Comment: After we add fallbackKeys, we can use new introducing options ## flink-connectors/flink-connector-jdbc/src/test/java/org/apache/flink/connector/jdbc/table/JdbcRowDataLookupFunctionTest.java: ## @@ -45,30 +43,29 @@ /** Test suite for {@link JdbcRowDataLookupFunction}. */ public class JdbcRowDataLookupFunctionTest extends JdbcLookupTestBase { -private static String[] fieldNames = new String[] {"id1", "id2", "comment1", "comment2"}; -private static DataType[] fieldDataTypes = +private static final String[] fieldNames = new String[] {"id1", "id2", "comment1", "comment2"}; +private static final DataType[] fieldDataTypes = new DataType[] { DataTypes.INT(), DataTypes.STRING(), DataTypes.STRING(), DataTypes.STRING() }; -private static String[] lookupKeys = new String[] {"id1", "id2"}; - -@Test -public void testEval() throws Exception { +private static final String[] lookupKeys = new String[] {"id1", "id2"}; -JdbcLookupOptions lookupOptions = JdbcLookupOptions.builder().build(); -JdbcRowDataLookupFunction lookupFunction = buildRowDataLookupFunction(lookupOptions); +@ParameterizedTest(name = "withFailure = {0}") +@ValueSource(booleans = {false, true}) +public void testLookup(boolean withFailure) throws Exception { +JdbcRowDataLookupFunction lookupFunction = buildRowDataLookupFunction(withFailure); ListOutputCollector collector = new ListOutputCollector(); lookupFunction.setCollector(collector); lookupFunction.open(null); lookupFunction.eval(1, StringData.fromString("1")); - -// close connection -lookupFunction.getDbConnection().close(); - +if (withFailure) { +// Close connection here, and this will be recovered by retry +lookupFunction.getDbConnection().close(); Review Comment: check null before call `close()` ## flink-connectors/flink-connector-jdbc/src/main/java/org/apache/flink/connector/jdbc/table/JdbcConnectorOptions.java: ## @@ -116,6 +117,8 @@ public class JdbcConnectorOptions { // Lookup options // - +/** @deprecated please use {@link LookupOptions#PARTIAL_CACHE_MAX_ROWS} instead. */ +@Deprecated public static final ConfigOption LOOKUP_CACHE_MAX_ROWS = ConfigOptions.key("lookup.cache.max-rows") Review Comment: Could we add `lookup.cache.max-rows` to LookupOptions#PARTIAL_CACHE_MAX_ROWS 's fallbackKeys ? -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] godfreyhe commented on pull request #20324: [FLINK-28568][table-runtime] Implements a new lookup join operator (sync mode only) with state to eliminate the non determinism
godfreyhe commented on PR #20324: URL: https://github.com/apache/flink/pull/20324#issuecomment-1208864940 > An irrelevant failure case of es sink https://issues.apache.org/jira/browse/FLINK-28877 I will merge it -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-28880) Fix CEP doc with wrong result of strict contiguity of looping patterns
[ https://issues.apache.org/jira/browse/FLINK-28880?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated FLINK-28880: --- Labels: pull-request-available (was: ) > Fix CEP doc with wrong result of strict contiguity of looping patterns > -- > > Key: FLINK-28880 > URL: https://issues.apache.org/jira/browse/FLINK-28880 > Project: Flink > Issue Type: Bug > Components: Library / CEP >Affects Versions: 1.15.1 >Reporter: Juntao Hu >Priority: Minor > Labels: pull-request-available > Fix For: 1.16.0 > > > https://nightlies.apache.org/flink/flink-docs-release-1.15/docs/libs/cep/#contiguity-within-looping-patterns > The result of strict contiguity should be {a b1 c}, {a b2 c}, {a b3 c}, since > b is *followed by* c. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[GitHub] [flink] Vancior opened a new pull request, #20508: [FLINK-28880][docs][cep] Fix wrong result of strict contiguity of looping patterns
Vancior opened a new pull request, #20508: URL: https://github.com/apache/flink/pull/20508 ## What is the purpose of the change This PR fixes CEP doc. ## Verifying this change This change is a trivial rework / code cleanup without any test coverage. ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): (no) - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: (no) - The serializers: (no) - The runtime per-record code paths (performance sensitive): (no) - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn, ZooKeeper: (no) - The S3 file system connector: (no) ## Documentation - Does this pull request introduce a new feature? (no) - If yes, how is the feature documented? (no applicable) -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] flinkbot commented on pull request #20507: [FLINK-28139][docs] Add documentation for speculative execution
flinkbot commented on PR #20507: URL: https://github.com/apache/flink/pull/20507#issuecomment-1208863205 ## CI report: * 9d67f8ac91de12838b909c38e995455983c5d32f UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Resolved] (FLINK-28843) Fail to find incremental handle when restoring from changelog checkpoint in claim mode
[ https://issues.apache.org/jira/browse/FLINK-28843?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Yun Tang resolved FLINK-28843. -- Fix Version/s: 1.16.0 Resolution: Fixed merged in master: 7f708d0ba42f727b3f8c3d77cef2108206cad2de > Fail to find incremental handle when restoring from changelog checkpoint in > claim mode > -- > > Key: FLINK-28843 > URL: https://issues.apache.org/jira/browse/FLINK-28843 > Project: Flink > Issue Type: Bug > Components: Runtime / State Backends >Affects Versions: 1.15.0, 1.15.1 >Reporter: Lihe Ma >Assignee: Lihe Ma >Priority: Critical > Labels: pull-request-available > Fix For: 1.16.0 > > > # When native checkpoint is enabled and incremental checkpointing is enabled > in rocksdb statebackend,if state data is greater than > state.storage.fs.memory-threshold,it will be stored in a data file > (FileStateHandle,RelativeFileStateHandle, etc) rather than stored with > ByteStreamStateHandle in checkpoint metadata, like base-path1/chk-1/file1. > # Then restore the job from base-path1/chk-1 in claim mode,using changelog > statebackend,and the checkpoint path is set to base-path2, then new > checkpoint will be saved in base-path2/chk-2, previous checkpoint file > (base-path1/chk-1/file1) is needed. > # Then restore the job from base-path2/chk-2 in changelog statebackend, > flink will try to read base-path2/chk-2/file1, rather than the actual file > location base-path1/chk-1/file1, which leads to FileNotFoundException and job > failed. > > How to reproduce? > # Set state.storage.fs.memory-threshold to a small value, like '20b'. > # {{run > org.apache.flink.test.checkpointing.ChangelogPeriodicMaterializationSwitchStateBackendITCase#testSwitchFromDisablingToEnablingInClaimMode}} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[GitHub] [flink] Myasuka merged pull request #20484: [FLINK-28843][StateBackend] Fix restore from incremental checkpoint with changelog checkpoint in claim mode
Myasuka merged PR #20484: URL: https://github.com/apache/flink/pull/20484 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Created] (FLINK-28880) Fix CEP doc with wrong result of strict contiguity of looping patterns
Juntao Hu created FLINK-28880: - Summary: Fix CEP doc with wrong result of strict contiguity of looping patterns Key: FLINK-28880 URL: https://issues.apache.org/jira/browse/FLINK-28880 Project: Flink Issue Type: Bug Components: Library / CEP Affects Versions: 1.15.1 Reporter: Juntao Hu Fix For: 1.16.0 https://nightlies.apache.org/flink/flink-docs-release-1.15/docs/libs/cep/#contiguity-within-looping-patterns The result of strict contiguity should be {a b1 c}, {a b2 c}, {a b3 c}, since b is *followed by* c. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (FLINK-28139) Add documentation for speculative execution
[ https://issues.apache.org/jira/browse/FLINK-28139?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated FLINK-28139: --- Labels: pull-request-available (was: ) > Add documentation for speculative execution > --- > > Key: FLINK-28139 > URL: https://issues.apache.org/jira/browse/FLINK-28139 > Project: Flink > Issue Type: Sub-task > Components: Runtime / Coordination >Reporter: Zhu Zhu >Assignee: Zhu Zhu >Priority: Major > Labels: pull-request-available > Fix For: 1.16.0 > > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[GitHub] [flink] zhuzhurk opened a new pull request, #20507: [FLINK-28139][docs] Add documentation for speculative execution
zhuzhurk opened a new pull request, #20507: URL: https://github.com/apache/flink/pull/20507 ## What is the purpose of the change This PR adds documentation for speculative execution. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] HuangXingBo commented on a diff in pull request #20505: [FLINK-28876][python][format/orc] Support writing RowData into Orc files
HuangXingBo commented on code in PR #20505: URL: https://github.com/apache/flink/pull/20505#discussion_r940834537 ## flink-python/pyflink/datastream/formats/orc.py: ## @@ -0,0 +1,104 @@ + +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +from typing import Optional + +from pyflink.common import Configuration +from pyflink.datastream.connectors.file_system import BulkWriterFactory, RowDataBulkWriterFactory +from pyflink.datastream.utils import create_hadoop_configuration +from pyflink.java_gateway import get_gateway +from pyflink.table.types import _to_java_data_type, RowType +from pyflink.util.java_utils import to_jarray + + +class OrcBulkWriters(object): +""" +Convenient builder to create a :class:`BulkWriterFactory` that writes Row records with a defined +:class:`RowType` into Orc files in a batch fashion. + +.. versionadded:: 1.16.0 +""" + +@staticmethod +def for_row_data_vectorization(row_type: RowType, + writer_properties: Optional[Configuration] = None, + hadoop_config: Optional[Configuration] = None) \ +-> BulkWriterFactory: +""" +Create a :class:`RowDataBulkWriterFactory` that writes Row records with a defined +:class:`RowType` into Orc files in a batch fashion. + +Example: +:: + +>>> row_type = DataTypes.ROW([ +... DataTypes.FIELD('string', DataTypes.STRING()), +... DataTypes.FIELD('int_array', DataTypes.ARRAY(DataTypes.INT())) +... ]) +>>> row_type_info = Types.ROW_NAMED( +... ['string', 'int_array'], +... [Types.STRING(), Types.LIST(Types.INT())] +... ) +>>> sink = FileSink.for_bulk_format( +... OUTPUT_DIR, OrcBulkWriters.for_row_data_vectorization( +... row_type=row_type, +... writer_properties=Configuration(), +... hadoop_config=Configuration(), +... ) +... ).build() +>>> ds.map(lambda e: e, output_type=row_type_info).sink_to(sink) + +Note that in the above example, an identity map to indicate its :class:`RowTypeInfo` is +necessary before ``sink_to`` when ``ds`` is a source stream producing **RowData** records, +because :class:`RowDataBulkWriterFactory` assumes the input record type is :class:`Row`. +""" +if not isinstance(row_type, RowType): +raise TypeError('row_type must be an instance of RowType') + +j_data_type = _to_java_data_type(row_type) +jvm = get_gateway().jvm +j_row_type = j_data_type.getLogicalType() +orc_types = to_jarray( +jvm.org.apache.flink.table.types.logical.LogicalType, +[i for i in j_row_type.getChildren()] +) +type_description = jvm.org.apache.flink.orc \ +.OrcSplitReaderUtil.logicalTypeToOrcType(j_row_type) +if writer_properties is None: +writer_properties = Configuration() +if hadoop_config is None: +hadoop_config = Configuration() + +return RowDataBulkWriterFactory( +jvm.org.apache.flink.orc.writer.OrcBulkWriterFactory( +jvm.org.apache.flink.orc.vector.RowDataVectorizer( +type_description.toString(), +orc_types +), +OrcBulkWriters._create_properties(writer_properties), +create_hadoop_configuration(hadoop_config) +), +row_type +) + +@staticmethod +def _create_properties(conf: Configuration): Review Comment: this method can be moved to `datastream.utils`? ## docs/content/docs/connectors/datastream/filesystem.md: ## @@ -792,6 +792,36 @@ class PersonVectorizer(schema: String) extends Vectorizer[Person](schema) { {{< /tab >}} {{< /tabs >}} +For PyFlink users,
[GitHub] [flink] godfreyhe commented on a diff in pull request #20501: Catalog get statistics by partitions
godfreyhe commented on code in PR #20501: URL: https://github.com/apache/flink/pull/20501#discussion_r940832517 ## flink-connectors/flink-connector-hive/src/main/java/org/apache/flink/table/catalog/hive/util/HiveStatsUtil.java: ## @@ -90,6 +102,257 @@ public static Map createCatalogColumnSt return colStats; } +public static Map createCatalogPartitionColumnStats( +HiveMetastoreClientWrapper client, +HiveShim hiveShim, +Table hiveTable, +String partitionName, +List partitionColsSchema, +String defaultPartitionName) { +Map partitionColumnStats = new HashMap<>(); +List partitionCols = new ArrayList<>(partitionColsSchema.size()); +List partitionColsType = new ArrayList<>(partitionColsSchema.size()); +for (FieldSchema fieldSchema : partitionColsSchema) { +partitionCols.add(fieldSchema.getName()); +partitionColsType.add( +HiveTypeUtil.toFlinkType( + TypeInfoUtils.getTypeInfoFromTypeString(fieldSchema.getType())) +.getLogicalType()); +} + +// the partition column and values for the partition column +Map partitionColValues = new HashMap<>(); +CatalogPartitionSpec partitionSpec = +HivePartitionUtils.createPartitionSpec(partitionName, defaultPartitionName); +for (int i = 0; i < partitionCols.size(); i++) { +String partitionCol = partitionCols.get(i); +String partitionStrVal = partitionSpec.getPartitionSpec().get(partitionCols.get(i)); +if (partitionStrVal == null) { +partitionColValues.put(partitionCol, null); +} else { +partitionColValues.put( +partitionCol, +HivePartitionUtils.restorePartitionValueFromType( +hiveShim, +partitionStrVal, +partitionColsType.get(i), +defaultPartitionName)); +} +} + +// calculate statistic for each partition column +for (int i = 0; i < partitionCols.size(); i++) { +Object partitionValue = partitionColValues.get(partitionCols.get(i)); +LogicalType logicalType = partitionColsType.get(i); +CatalogColumnStatisticsDataBase catalogColumnStatistics = +getColumnStatistics( +client, +hiveTable, +logicalType, +partitionValue, +i, +defaultPartitionName); +if (catalogColumnStatistics != null) { +partitionColumnStats.put(partitionCols.get(i), catalogColumnStatistics); +} +} + +return partitionColumnStats; +} + +/** + * Get statistics for specific partition column. + * + * @param logicalType the specific partition column's logical type + * @param partitionValue the partition value for the specific partition column + * @param partitionColIndex the index of the specific partition + * @param defaultPartitionName the default partition name for null value + */ +private static CatalogColumnStatisticsDataBase getColumnStatistics( +HiveMetastoreClientWrapper client, +Table hiveTable, +LogicalType logicalType, +Object partitionValue, +int partitionColIndex, +String defaultPartitionName) { +switch (logicalType.getTypeRoot()) { +case CHAR: +case VARCHAR: +{ +Long maxLength = null; +Double avgLength = null; +Long nullCount = 0L; +if (partitionValue == null) { +nullCount = +getNullCount( +client, hiveTable, partitionColIndex, defaultPartitionName); +} else { +long valLength = ((String) partitionValue).length(); +maxLength = valLength; +avgLength = (double) valLength; +} +return new CatalogColumnStatisticsDataString( +maxLength, avgLength, 1L, nullCount); +} +case BOOLEAN: +{ +long trueCount = 0L; +long falseCount = 0L; +Long nullCount = 0L; +if (partitionValue == null) { +nullCount = +getNullCount( +client, hiveTable, partitionColIndex,
[GitHub] [flink] gaoyunhaii commented on a diff in pull request #20491: [FLINK-28857][docs] Add Document for DataStream Cache API
gaoyunhaii commented on code in PR #20491: URL: https://github.com/apache/flink/pull/20491#discussion_r940831093 ## docs/content/docs/dev/datastream/operators/overview.md: ## @@ -575,6 +575,44 @@ This feature is not yet supported in Python {{< /tab >}} {{< /tabs>}} +### Cache + DataStream CachedDataStream + +Cache the intermediate result of the transformation. Currently, only supported with batch execution Review Comment: It is only supported that... with the batch execution mode? -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Closed] (FLINK-28753) Improve FilterIntoJoinRule which could push some predicate to another side
[ https://issues.apache.org/jira/browse/FLINK-28753?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] godfrey he closed FLINK-28753. -- Resolution: Fixed Fixed in master: 0e6e4198ad84227c20e2c61c2dd8b0616324aa31 > Improve FilterIntoJoinRule which could push some predicate to another side > -- > > Key: FLINK-28753 > URL: https://issues.apache.org/jira/browse/FLINK-28753 > Project: Flink > Issue Type: Improvement > Components: Table SQL / Planner >Reporter: godfrey he >Assignee: godfrey he >Priority: Major > Labels: pull-request-available > Fix For: 1.16.0 > > > for sql: SELECT * FROM MyTable1 join MyTable2 ON a1 = a2 AND a1 = 2 > {{a1 = 2}} can be pushed into both left side and right side. but currently > only left side will be pushed by FilterIntoJoinRule. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[GitHub] [flink] libenchao commented on pull request #20140: [Flink 16024][Connector][JDBC] Support FilterPushdown
libenchao commented on PR #20140: URL: https://github.com/apache/flink/pull/20140#issuecomment-1208843267 > I believe JdbcFilterPushdownVisitor needs to produce strings or some equivalent data, so that it can be used by JdbcDynamicTableSource::getScanRuntimeProvider, how should I make use of PrepareStatement here? Maybe I am missing something? @qingwei91 Currently `JdbcRowDataInputFormat` already uses `PreparedStatement`, and 'scan.partition' is implemented using it, hence we can do it. `JdbcFilterPushdownVisitor` does not necessarily need to return `String`, it can return anything we need. We may need to rethink about the design, especially for the following points: 1. What functions should we support, e.g. `IN`, `BETWEEN`? 2. Are these functions supported in all JDBC dialects, and do they use the same operator? If not, how should we abstract it out to make it dialect specific? -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] Grypse commented on pull request #20467: [FLINK-28825][k8s] Add K8S pod scheduler into Kubernetes options
Grypse commented on PR #20467: URL: https://github.com/apache/flink/pull/20467#issuecomment-1208842402 The scheduler can be setting with schedulerName in podTemplate simply,I think there is no need to add options to kubernetes config. `podTemplate: apiVersion: v1 kind: Pod metadata: name: task-manager-pod-template namespace: flink-application-system spec: schedulerName: default-scheduler ` -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] godfreyhe closed pull request #20432: [FLINK-28753][table-planner] Improve FilterIntoJoinRule which could push some predicates to another side
godfreyhe closed pull request #20432: [FLINK-28753][table-planner] Improve FilterIntoJoinRule which could push some predicates to another side URL: https://github.com/apache/flink/pull/20432 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Closed] (FLINK-28183) flink-python is lacking several test dependencies
[ https://issues.apache.org/jira/browse/FLINK-28183?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Dian Fu closed FLINK-28183. --- Resolution: Fixed Merged to master via 5506930cc79eb131c66c5df0320045ad53437dce > flink-python is lacking several test dependencies > - > > Key: FLINK-28183 > URL: https://issues.apache.org/jira/browse/FLINK-28183 > Project: Flink > Issue Type: Technical Debt > Components: API / Python, Build System >Affects Versions: 1.16.0 >Reporter: Chesnay Schepler >Assignee: Chesnay Schepler >Priority: Critical > Labels: pull-request-available > Fix For: 1.16.0 > > > The pyflink_gateway_server searches the output directories of various modules > to construct a test classpath. > Half of these are not declared as actual test dependencies in maven. Because > of that there are no guarantees that these modules are actually built before > flink-python. > Additionally there seem to be no safeguards in place to verify that these > jars actually exist. > Considering that this is only required for testing most of this logic should > also be moved into maven, copying these dependencies to some directory under > flink-python/target, to make this de-facto build logic more discoverable. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Comment Edited] (FLINK-24095) Elasticsearch7DynamicSinkITCase.testWritingDocuments fails due to socket timeout
[ https://issues.apache.org/jira/browse/FLINK-24095?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17577090#comment-17577090 ] Huang Xingbo edited comment on FLINK-24095 at 8/9/22 2:24 AM: -- 1.16 instance: https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=39617=logs=4eda0b4a-bd0d-521a-0916-8285b9be9bb5=2ff6d5fa-53a6-53ac-bff7-fa524ea361a9 was (Author: hxbks2ks): https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=39617=logs=4eda0b4a-bd0d-521a-0916-8285b9be9bb5=2ff6d5fa-53a6-53ac-bff7-fa524ea361a9 > Elasticsearch7DynamicSinkITCase.testWritingDocuments fails due to socket > timeout > > > Key: FLINK-24095 > URL: https://issues.apache.org/jira/browse/FLINK-24095 > Project: Flink > Issue Type: Bug > Components: Connectors / ElasticSearch >Affects Versions: 1.14.0, 1.15.0 >Reporter: Xintong Song >Priority: Major > Labels: test-stability > Fix For: 1.16.0 > > > https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=23250=logs=d44f43ce-542c-597d-bf94-b0718c71e5e8=ed165f3f-d0f6-524b-5279-86f8ee7d0e2d=12781 > {code} > Aug 31 23:06:22 Caused by: java.net.SocketTimeoutException: 30,000 > milliseconds timeout on connection http-outgoing-3 [ACTIVE] > Aug 31 23:06:22 at > org.elasticsearch.client.RestClient.extractAndWrapCause(RestClient.java:808) > Aug 31 23:06:22 at > org.elasticsearch.client.RestClient.performRequest(RestClient.java:248) > Aug 31 23:06:22 at > org.elasticsearch.client.RestClient.performRequest(RestClient.java:235) > Aug 31 23:06:22 at > org.elasticsearch.client.RestHighLevelClient.internalPerformRequest(RestHighLevelClient.java:1514) > Aug 31 23:06:22 at > org.elasticsearch.client.RestHighLevelClient.performRequest(RestHighLevelClient.java:1499) > Aug 31 23:06:22 at > org.elasticsearch.client.RestHighLevelClient.ping(RestHighLevelClient.java:720) > Aug 31 23:06:22 at > org.apache.flink.streaming.connectors.elasticsearch7.Elasticsearch7ApiCallBridge.verifyClientConnection(Elasticsearch7ApiCallBridge.java:138) > Aug 31 23:06:22 at > org.apache.flink.streaming.connectors.elasticsearch7.Elasticsearch7ApiCallBridge.verifyClientConnection(Elasticsearch7ApiCallBridge.java:46) > Aug 31 23:06:22 at > org.apache.flink.streaming.connectors.elasticsearch.ElasticsearchSinkBase.open(ElasticsearchSinkBase.java:318) > Aug 31 23:06:22 at > org.apache.flink.api.common.functions.util.FunctionUtils.openFunction(FunctionUtils.java:34) > Aug 31 23:06:22 at > org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator.open(AbstractUdfStreamOperator.java:100) > Aug 31 23:06:22 at > org.apache.flink.streaming.api.operators.StreamSink.open(StreamSink.java:46) > Aug 31 23:06:22 at > org.apache.flink.streaming.runtime.tasks.RegularOperatorChain.initializeStateAndOpenOperators(RegularOperatorChain.java:110) > Aug 31 23:06:22 at > org.apache.flink.streaming.runtime.tasks.StreamTask.restoreGates(StreamTask.java:691) > Aug 31 23:06:22 at > org.apache.flink.streaming.runtime.tasks.StreamTaskActionExecutor$1.call(StreamTaskActionExecutor.java:55) > Aug 31 23:06:22 at > org.apache.flink.streaming.runtime.tasks.StreamTask.restoreInternal(StreamTask.java:667) > Aug 31 23:06:22 at > org.apache.flink.streaming.runtime.tasks.StreamTask.restore(StreamTask.java:639) > Aug 31 23:06:22 at > org.apache.flink.runtime.taskmanager.Task.runWithSystemExitMonitoring(Task.java:958) > Aug 31 23:06:22 at > org.apache.flink.runtime.taskmanager.Task.restoreAndInvoke(Task.java:927) > Aug 31 23:06:22 at > org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:766) > Aug 31 23:06:22 at > org.apache.flink.runtime.taskmanager.Task.run(Task.java:575) > Aug 31 23:06:22 at java.lang.Thread.run(Thread.java:748) > Aug 31 23:06:22 Caused by: java.net.SocketTimeoutException: 30,000 > milliseconds timeout on connection http-outgoing-3 [ACTIVE] > Aug 31 23:06:22 at > org.apache.http.nio.protocol.HttpAsyncRequestExecutor.timeout(HttpAsyncRequestExecutor.java:387) > Aug 31 23:06:22 at > org.apache.http.impl.nio.client.InternalIODispatch.onTimeout(InternalIODispatch.java:92) > Aug 31 23:06:22 at > org.apache.http.impl.nio.client.InternalIODispatch.onTimeout(InternalIODispatch.java:39) > Aug 31 23:06:22 at > org.apache.http.impl.nio.reactor.AbstractIODispatch.timeout(AbstractIODispatch.java:175) > Aug 31 23:06:22 at > org.apache.http.impl.nio.reactor.BaseIOReactor.sessionTimedOut(BaseIOReactor.java:261) > Aug 31 23:06:22 at > org.apache.http.impl.nio.reactor.AbstractIOReactor.timeoutCheck(AbstractIOReactor.java:502) > Aug 31
[jira] [Updated] (FLINK-24095) Elasticsearch7DynamicSinkITCase.testWritingDocuments fails due to socket timeout
[ https://issues.apache.org/jira/browse/FLINK-24095?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Huang Xingbo updated FLINK-24095: - Affects Version/s: 1.16.0 > Elasticsearch7DynamicSinkITCase.testWritingDocuments fails due to socket > timeout > > > Key: FLINK-24095 > URL: https://issues.apache.org/jira/browse/FLINK-24095 > Project: Flink > Issue Type: Bug > Components: Connectors / ElasticSearch >Affects Versions: 1.14.0, 1.15.0, 1.16.0 >Reporter: Xintong Song >Priority: Major > Labels: test-stability > Fix For: 1.16.0 > > > https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=23250=logs=d44f43ce-542c-597d-bf94-b0718c71e5e8=ed165f3f-d0f6-524b-5279-86f8ee7d0e2d=12781 > {code} > Aug 31 23:06:22 Caused by: java.net.SocketTimeoutException: 30,000 > milliseconds timeout on connection http-outgoing-3 [ACTIVE] > Aug 31 23:06:22 at > org.elasticsearch.client.RestClient.extractAndWrapCause(RestClient.java:808) > Aug 31 23:06:22 at > org.elasticsearch.client.RestClient.performRequest(RestClient.java:248) > Aug 31 23:06:22 at > org.elasticsearch.client.RestClient.performRequest(RestClient.java:235) > Aug 31 23:06:22 at > org.elasticsearch.client.RestHighLevelClient.internalPerformRequest(RestHighLevelClient.java:1514) > Aug 31 23:06:22 at > org.elasticsearch.client.RestHighLevelClient.performRequest(RestHighLevelClient.java:1499) > Aug 31 23:06:22 at > org.elasticsearch.client.RestHighLevelClient.ping(RestHighLevelClient.java:720) > Aug 31 23:06:22 at > org.apache.flink.streaming.connectors.elasticsearch7.Elasticsearch7ApiCallBridge.verifyClientConnection(Elasticsearch7ApiCallBridge.java:138) > Aug 31 23:06:22 at > org.apache.flink.streaming.connectors.elasticsearch7.Elasticsearch7ApiCallBridge.verifyClientConnection(Elasticsearch7ApiCallBridge.java:46) > Aug 31 23:06:22 at > org.apache.flink.streaming.connectors.elasticsearch.ElasticsearchSinkBase.open(ElasticsearchSinkBase.java:318) > Aug 31 23:06:22 at > org.apache.flink.api.common.functions.util.FunctionUtils.openFunction(FunctionUtils.java:34) > Aug 31 23:06:22 at > org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator.open(AbstractUdfStreamOperator.java:100) > Aug 31 23:06:22 at > org.apache.flink.streaming.api.operators.StreamSink.open(StreamSink.java:46) > Aug 31 23:06:22 at > org.apache.flink.streaming.runtime.tasks.RegularOperatorChain.initializeStateAndOpenOperators(RegularOperatorChain.java:110) > Aug 31 23:06:22 at > org.apache.flink.streaming.runtime.tasks.StreamTask.restoreGates(StreamTask.java:691) > Aug 31 23:06:22 at > org.apache.flink.streaming.runtime.tasks.StreamTaskActionExecutor$1.call(StreamTaskActionExecutor.java:55) > Aug 31 23:06:22 at > org.apache.flink.streaming.runtime.tasks.StreamTask.restoreInternal(StreamTask.java:667) > Aug 31 23:06:22 at > org.apache.flink.streaming.runtime.tasks.StreamTask.restore(StreamTask.java:639) > Aug 31 23:06:22 at > org.apache.flink.runtime.taskmanager.Task.runWithSystemExitMonitoring(Task.java:958) > Aug 31 23:06:22 at > org.apache.flink.runtime.taskmanager.Task.restoreAndInvoke(Task.java:927) > Aug 31 23:06:22 at > org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:766) > Aug 31 23:06:22 at > org.apache.flink.runtime.taskmanager.Task.run(Task.java:575) > Aug 31 23:06:22 at java.lang.Thread.run(Thread.java:748) > Aug 31 23:06:22 Caused by: java.net.SocketTimeoutException: 30,000 > milliseconds timeout on connection http-outgoing-3 [ACTIVE] > Aug 31 23:06:22 at > org.apache.http.nio.protocol.HttpAsyncRequestExecutor.timeout(HttpAsyncRequestExecutor.java:387) > Aug 31 23:06:22 at > org.apache.http.impl.nio.client.InternalIODispatch.onTimeout(InternalIODispatch.java:92) > Aug 31 23:06:22 at > org.apache.http.impl.nio.client.InternalIODispatch.onTimeout(InternalIODispatch.java:39) > Aug 31 23:06:22 at > org.apache.http.impl.nio.reactor.AbstractIODispatch.timeout(AbstractIODispatch.java:175) > Aug 31 23:06:22 at > org.apache.http.impl.nio.reactor.BaseIOReactor.sessionTimedOut(BaseIOReactor.java:261) > Aug 31 23:06:22 at > org.apache.http.impl.nio.reactor.AbstractIOReactor.timeoutCheck(AbstractIOReactor.java:502) > Aug 31 23:06:22 at > org.apache.http.impl.nio.reactor.BaseIOReactor.validate(BaseIOReactor.java:211) > Aug 31 23:06:22 at > org.apache.http.impl.nio.reactor.AbstractIOReactor.execute(AbstractIOReactor.java:280) > Aug 31 23:06:22 at > org.apache.http.impl.nio.reactor.BaseIOReactor.execute(BaseIOReactor.java:104) > Aug 31 23:06:22 at >
[jira] [Commented] (FLINK-24095) Elasticsearch7DynamicSinkITCase.testWritingDocuments fails due to socket timeout
[ https://issues.apache.org/jira/browse/FLINK-24095?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17577090#comment-17577090 ] Huang Xingbo commented on FLINK-24095: -- https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=39617=logs=4eda0b4a-bd0d-521a-0916-8285b9be9bb5=2ff6d5fa-53a6-53ac-bff7-fa524ea361a9 > Elasticsearch7DynamicSinkITCase.testWritingDocuments fails due to socket > timeout > > > Key: FLINK-24095 > URL: https://issues.apache.org/jira/browse/FLINK-24095 > Project: Flink > Issue Type: Bug > Components: Connectors / ElasticSearch >Affects Versions: 1.14.0, 1.15.0 >Reporter: Xintong Song >Priority: Major > Labels: test-stability > Fix For: 1.16.0 > > > https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=23250=logs=d44f43ce-542c-597d-bf94-b0718c71e5e8=ed165f3f-d0f6-524b-5279-86f8ee7d0e2d=12781 > {code} > Aug 31 23:06:22 Caused by: java.net.SocketTimeoutException: 30,000 > milliseconds timeout on connection http-outgoing-3 [ACTIVE] > Aug 31 23:06:22 at > org.elasticsearch.client.RestClient.extractAndWrapCause(RestClient.java:808) > Aug 31 23:06:22 at > org.elasticsearch.client.RestClient.performRequest(RestClient.java:248) > Aug 31 23:06:22 at > org.elasticsearch.client.RestClient.performRequest(RestClient.java:235) > Aug 31 23:06:22 at > org.elasticsearch.client.RestHighLevelClient.internalPerformRequest(RestHighLevelClient.java:1514) > Aug 31 23:06:22 at > org.elasticsearch.client.RestHighLevelClient.performRequest(RestHighLevelClient.java:1499) > Aug 31 23:06:22 at > org.elasticsearch.client.RestHighLevelClient.ping(RestHighLevelClient.java:720) > Aug 31 23:06:22 at > org.apache.flink.streaming.connectors.elasticsearch7.Elasticsearch7ApiCallBridge.verifyClientConnection(Elasticsearch7ApiCallBridge.java:138) > Aug 31 23:06:22 at > org.apache.flink.streaming.connectors.elasticsearch7.Elasticsearch7ApiCallBridge.verifyClientConnection(Elasticsearch7ApiCallBridge.java:46) > Aug 31 23:06:22 at > org.apache.flink.streaming.connectors.elasticsearch.ElasticsearchSinkBase.open(ElasticsearchSinkBase.java:318) > Aug 31 23:06:22 at > org.apache.flink.api.common.functions.util.FunctionUtils.openFunction(FunctionUtils.java:34) > Aug 31 23:06:22 at > org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator.open(AbstractUdfStreamOperator.java:100) > Aug 31 23:06:22 at > org.apache.flink.streaming.api.operators.StreamSink.open(StreamSink.java:46) > Aug 31 23:06:22 at > org.apache.flink.streaming.runtime.tasks.RegularOperatorChain.initializeStateAndOpenOperators(RegularOperatorChain.java:110) > Aug 31 23:06:22 at > org.apache.flink.streaming.runtime.tasks.StreamTask.restoreGates(StreamTask.java:691) > Aug 31 23:06:22 at > org.apache.flink.streaming.runtime.tasks.StreamTaskActionExecutor$1.call(StreamTaskActionExecutor.java:55) > Aug 31 23:06:22 at > org.apache.flink.streaming.runtime.tasks.StreamTask.restoreInternal(StreamTask.java:667) > Aug 31 23:06:22 at > org.apache.flink.streaming.runtime.tasks.StreamTask.restore(StreamTask.java:639) > Aug 31 23:06:22 at > org.apache.flink.runtime.taskmanager.Task.runWithSystemExitMonitoring(Task.java:958) > Aug 31 23:06:22 at > org.apache.flink.runtime.taskmanager.Task.restoreAndInvoke(Task.java:927) > Aug 31 23:06:22 at > org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:766) > Aug 31 23:06:22 at > org.apache.flink.runtime.taskmanager.Task.run(Task.java:575) > Aug 31 23:06:22 at java.lang.Thread.run(Thread.java:748) > Aug 31 23:06:22 Caused by: java.net.SocketTimeoutException: 30,000 > milliseconds timeout on connection http-outgoing-3 [ACTIVE] > Aug 31 23:06:22 at > org.apache.http.nio.protocol.HttpAsyncRequestExecutor.timeout(HttpAsyncRequestExecutor.java:387) > Aug 31 23:06:22 at > org.apache.http.impl.nio.client.InternalIODispatch.onTimeout(InternalIODispatch.java:92) > Aug 31 23:06:22 at > org.apache.http.impl.nio.client.InternalIODispatch.onTimeout(InternalIODispatch.java:39) > Aug 31 23:06:22 at > org.apache.http.impl.nio.reactor.AbstractIODispatch.timeout(AbstractIODispatch.java:175) > Aug 31 23:06:22 at > org.apache.http.impl.nio.reactor.BaseIOReactor.sessionTimedOut(BaseIOReactor.java:261) > Aug 31 23:06:22 at > org.apache.http.impl.nio.reactor.AbstractIOReactor.timeoutCheck(AbstractIOReactor.java:502) > Aug 31 23:06:22 at > org.apache.http.impl.nio.reactor.BaseIOReactor.validate(BaseIOReactor.java:211) > Aug 31 23:06:22 at > org.apache.http.impl.nio.reactor.AbstractIOReactor.execute(AbstractIOReactor.java:280) > Aug 31 23:06:22 at
[jira] [Closed] (FLINK-28877) Elasticsearch7DynamicSinkITCase.testWritingDocumentsNoPrimaryKey case failed
[ https://issues.apache.org/jira/browse/FLINK-28877?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Huang Xingbo closed FLINK-28877. Resolution: Duplicate > Elasticsearch7DynamicSinkITCase.testWritingDocumentsNoPrimaryKey case failed > > > Key: FLINK-28877 > URL: https://issues.apache.org/jira/browse/FLINK-28877 > Project: Flink > Issue Type: Bug >Reporter: lincoln lee >Priority: Major > Fix For: 1.16.0 > > > {code} > Aug 08 16:00:39 Caused by: java.lang.RuntimeException: An error occurred in > ElasticsearchSink. > Aug 08 16:00:39 at > org.apache.flink.streaming.connectors.elasticsearch.ElasticsearchSinkBase.checkErrorAndRethrow(ElasticsearchSinkBase.java:426) > > Aug 08 16:00:39 at > org.apache.flink.streaming.connectors.elasticsearch.ElasticsearchSinkBase.close(ElasticsearchSinkBase.java:365) > > Aug 08 16:00:39 at > org.apache.flink.api.common.functions.util.FunctionUtils.closeFunction(FunctionUtils.java:41) > > Aug 08 16:00:39 at > org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator.close(AbstractUdfStreamOperator.java:114) > > Aug 08 16:00:39 at > org.apache.flink.streaming.runtime.tasks.StreamOperatorWrapper.close(StreamOperatorWrapper.java:163) > > Aug 08 16:00:39 at > org.apache.flink.streaming.runtime.tasks.RegularOperatorChain.closeAllOperators(RegularOperatorChain.java:125) > > Aug 08 16:00:39 at > org.apache.flink.streaming.runtime.tasks.StreamTask.closeAllOperators(StreamTask.java:1022) > > Aug 08 16:00:39 at > org.apache.flink.streaming.runtime.tasks.StreamTask.afterInvoke(StreamTask.java:900) > > Aug 08 16:00:39 at > org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:783) > > Aug 08 16:00:39 at > org.apache.flink.runtime.taskmanager.Task.runWithSystemExitMonitoring(Task.java:935) > > Aug 08 16:00:39 at > org.apache.flink.runtime.taskmanager.Task.restoreAndInvoke(Task.java:914) > Aug 08 16:00:39 at > org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:728) > Aug 08 16:00:39 at > org.apache.flink.runtime.taskmanager.Task.run(Task.java:550) > Aug 08 16:00:39 at java.lang.Thread.run(Thread.java:748) > Aug 08 16:00:39 Caused by: java.net.SocketTimeoutException: 30,000 > milliseconds timeout on connection http-outgoing-0 [ACTIVE] > Aug 08 16:00:39 at > org.apache.http.nio.protocol.HttpAsyncRequestExecutor.timeout(HttpAsyncRequestExecutor.java:387) > > Aug 08 16:00:39 at > org.apache.http.impl.nio.client.InternalIODispatch.onTimeout(InternalIODispatch.java:92) > > Aug 08 16:00:39 at > org.apache.http.impl.nio.client.InternalIODispatch.onTimeout(InternalIODispatch.java:39) > > Aug 08 16:00:39 at > org.apache.http.impl.nio.reactor.AbstractIODispatch.timeout(AbstractIODispatch.java:175) > > Aug 08 16:00:39 at > org.apache.http.impl.nio.reactor.BaseIOReactor.sessionTimedOut(BaseIOReactor.java:261) > > Aug 08 16:00:39 at > org.apache.http.impl.nio.reactor.AbstractIOReactor.timeoutCheck(AbstractIOReactor.java:502) > > Aug 08 16:00:39 at > org.apache.http.impl.nio.reactor.BaseIOReactor.validate(BaseIOReactor.java:211) > > Aug 08 16:00:39 at > org.apache.http.impl.nio.reactor.AbstractIOReactor.execute(AbstractIOReactor.java:280) > > Aug 08 16:00:39 at > org.apache.http.impl.nio.reactor.BaseIOReactor.execute(BaseIOReactor.java:104) > > Aug 08 16:00:39 at > org.apache.http.impl.nio.reactor.AbstractMultiworkerIOReactor$Worker.run(AbstractMultiworkerIOReactor.java:591) > {code} > > https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=39617=logs=4eda0b4a-bd0d-521a-0916-8285b9be9bb5=2ff6d5fa-53a6-53ac-bff7-fa524ea361a9 -- This message was sent by Atlassian Jira (v8.20.10#820010)
[GitHub] [flink] dianfu closed pull request #20470: [FLINK-28183][python] Model python test dependencies in Maven
dianfu closed pull request #20470: [FLINK-28183][python] Model python test dependencies in Maven URL: https://github.com/apache/flink/pull/20470 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Assigned] (FLINK-28860) CacheITCase.testBatchProduceCacheStreamConsume failed
[ https://issues.apache.org/jira/browse/FLINK-28860?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Huang Xingbo reassigned FLINK-28860: Assignee: Xuannan Su > CacheITCase.testBatchProduceCacheStreamConsume failed > - > > Key: FLINK-28860 > URL: https://issues.apache.org/jira/browse/FLINK-28860 > Project: Flink > Issue Type: Bug > Components: API / DataStream >Affects Versions: 1.16.0 >Reporter: Huang Xingbo >Assignee: Xuannan Su >Priority: Critical > Labels: test-stability > Fix For: 1.16.0 > > > {code:java} > 2022-08-08T03:27:22.1988575Z Aug 08 03:27:22 [ERROR] > org.apache.flink.test.streaming.runtime.CacheITCase.testBatchProduceCacheStreamConsume(Path) > Time elapsed: 0.593 s <<< ERROR! > 2022-08-08T03:27:22.1989338Z Aug 08 03:27:22 java.lang.RuntimeException: > Producing cache IntermediateResult is not supported in streaming mode > 2022-08-08T03:27:22.1990401Z Aug 08 03:27:22 at > org.apache.flink.streaming.runtime.translators.CacheTransformationTranslator.translateForStreamingInternal(CacheTransformationTranslator.java:75) > 2022-08-08T03:27:22.1991511Z Aug 08 03:27:22 at > org.apache.flink.streaming.runtime.translators.CacheTransformationTranslator.translateForStreamingInternal(CacheTransformationTranslator.java:42) > 2022-08-08T03:27:22.1993671Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.graph.SimpleTransformationTranslator.translateForStreaming(SimpleTransformationTranslator.java:62) > 2022-08-08T03:27:22.1994900Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.graph.StreamGraphGenerator.translate(StreamGraphGenerator.java:830) > 2022-08-08T03:27:22.1995748Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.graph.StreamGraphGenerator.transform(StreamGraphGenerator.java:560) > 2022-08-08T03:27:22.1996932Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.graph.StreamGraphGenerator.getParentInputIds(StreamGraphGenerator.java:851) > 2022-08-08T03:27:22.1998562Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.graph.StreamGraphGenerator.translate(StreamGraphGenerator.java:809) > 2022-08-08T03:27:22.1999581Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.graph.StreamGraphGenerator.transform(StreamGraphGenerator.java:560) > 2022-08-08T03:27:22.2000376Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.graph.StreamGraphGenerator.generate(StreamGraphGenerator.java:319) > 2022-08-08T03:27:22.2001359Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.getStreamGraph(StreamExecutionEnvironment.java:2250) > 2022-08-08T03:27:22.2002767Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.getStreamGraph(StreamExecutionEnvironment.java:2241) > 2022-08-08T03:27:22.2004121Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.getStreamGraph(StreamExecutionEnvironment.java:2227) > 2022-08-08T03:27:22.2005059Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.executeAsync(StreamExecutionEnvironment.java:2178) > 2022-08-08T03:27:22.2005939Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.datastream.DataStream.executeAndCollectWithClient(DataStream.java:1469) > 2022-08-08T03:27:22.2006735Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.datastream.DataStream.executeAndCollect(DataStream.java:1334) > 2022-08-08T03:27:22.2007500Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.datastream.DataStream.executeAndCollect(DataStream.java:1320) > 2022-08-08T03:27:22.2008315Z Aug 08 03:27:22 at > org.apache.flink.test.streaming.runtime.CacheITCase.testBatchProduceCacheStreamConsume(CacheITCase.java:190) > {code} > https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=39518=logs=a57e0635-3fad-5b08-57c7-a4142d7d6fa9=2ef0effc-1da1-50e5-c2bd-aab434b1c5b7 -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (FLINK-28860) CacheITCase.testBatchProduceCacheStreamConsume failed
[ https://issues.apache.org/jira/browse/FLINK-28860?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17577084#comment-17577084 ] Xuannan Su commented on FLINK-28860: [~hxbks2ks]I will take a look at it. Could you assign the ticket to me? > CacheITCase.testBatchProduceCacheStreamConsume failed > - > > Key: FLINK-28860 > URL: https://issues.apache.org/jira/browse/FLINK-28860 > Project: Flink > Issue Type: Bug > Components: API / DataStream >Affects Versions: 1.16.0 >Reporter: Huang Xingbo >Priority: Critical > Labels: test-stability > Fix For: 1.16.0 > > > {code:java} > 2022-08-08T03:27:22.1988575Z Aug 08 03:27:22 [ERROR] > org.apache.flink.test.streaming.runtime.CacheITCase.testBatchProduceCacheStreamConsume(Path) > Time elapsed: 0.593 s <<< ERROR! > 2022-08-08T03:27:22.1989338Z Aug 08 03:27:22 java.lang.RuntimeException: > Producing cache IntermediateResult is not supported in streaming mode > 2022-08-08T03:27:22.1990401Z Aug 08 03:27:22 at > org.apache.flink.streaming.runtime.translators.CacheTransformationTranslator.translateForStreamingInternal(CacheTransformationTranslator.java:75) > 2022-08-08T03:27:22.1991511Z Aug 08 03:27:22 at > org.apache.flink.streaming.runtime.translators.CacheTransformationTranslator.translateForStreamingInternal(CacheTransformationTranslator.java:42) > 2022-08-08T03:27:22.1993671Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.graph.SimpleTransformationTranslator.translateForStreaming(SimpleTransformationTranslator.java:62) > 2022-08-08T03:27:22.1994900Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.graph.StreamGraphGenerator.translate(StreamGraphGenerator.java:830) > 2022-08-08T03:27:22.1995748Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.graph.StreamGraphGenerator.transform(StreamGraphGenerator.java:560) > 2022-08-08T03:27:22.1996932Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.graph.StreamGraphGenerator.getParentInputIds(StreamGraphGenerator.java:851) > 2022-08-08T03:27:22.1998562Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.graph.StreamGraphGenerator.translate(StreamGraphGenerator.java:809) > 2022-08-08T03:27:22.1999581Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.graph.StreamGraphGenerator.transform(StreamGraphGenerator.java:560) > 2022-08-08T03:27:22.2000376Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.graph.StreamGraphGenerator.generate(StreamGraphGenerator.java:319) > 2022-08-08T03:27:22.2001359Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.getStreamGraph(StreamExecutionEnvironment.java:2250) > 2022-08-08T03:27:22.2002767Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.getStreamGraph(StreamExecutionEnvironment.java:2241) > 2022-08-08T03:27:22.2004121Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.getStreamGraph(StreamExecutionEnvironment.java:2227) > 2022-08-08T03:27:22.2005059Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.executeAsync(StreamExecutionEnvironment.java:2178) > 2022-08-08T03:27:22.2005939Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.datastream.DataStream.executeAndCollectWithClient(DataStream.java:1469) > 2022-08-08T03:27:22.2006735Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.datastream.DataStream.executeAndCollect(DataStream.java:1334) > 2022-08-08T03:27:22.2007500Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.datastream.DataStream.executeAndCollect(DataStream.java:1320) > 2022-08-08T03:27:22.2008315Z Aug 08 03:27:22 at > org.apache.flink.test.streaming.runtime.CacheITCase.testBatchProduceCacheStreamConsume(CacheITCase.java:190) > {code} > https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=39518=logs=a57e0635-3fad-5b08-57c7-a4142d7d6fa9=2ef0effc-1da1-50e5-c2bd-aab434b1c5b7 -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (FLINK-28529) ChangelogPeriodicMaterializationSwitchStateBackendITCase.testSwitchFromDisablingToEnablingInClaimMode failed with CheckpointException: Checkpoint expired before comple
[ https://issues.apache.org/jira/browse/FLINK-28529?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17577082#comment-17577082 ] Huang Xingbo commented on FLINK-28529: -- https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=39675=logs=b0a398c0-685b-599c-eb57-c8c2a771138e=747432ad-a576-5911-1e2a-68c6bedc248a > ChangelogPeriodicMaterializationSwitchStateBackendITCase.testSwitchFromDisablingToEnablingInClaimMode > failed with CheckpointException: Checkpoint expired before completing > --- > > Key: FLINK-28529 > URL: https://issues.apache.org/jira/browse/FLINK-28529 > Project: Flink > Issue Type: Bug > Components: Runtime / Checkpointing, Runtime / State Backends >Affects Versions: 1.16.0 >Reporter: Huang Xingbo >Assignee: Yanfei Lei >Priority: Critical > Labels: pull-request-available, test-stability > Fix For: 1.16.0 > > > {code:java} > 2022-07-12T04:30:49.9912088Z Jul 12 04:30:49 [ERROR] > ChangelogPeriodicMaterializationSwitchStateBackendITCase.testSwitchFromDisablingToEnablingInClaimMode > Time elapsed: 617.048 s <<< ERROR! > 2022-07-12T04:30:49.9913108Z Jul 12 04:30:49 > java.util.concurrent.ExecutionException: > org.apache.flink.runtime.checkpoint.CheckpointException: Checkpoint expired > before completing. > 2022-07-12T04:30:49.9913880Z Jul 12 04:30:49 at > java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357) > 2022-07-12T04:30:49.9914606Z Jul 12 04:30:49 at > java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1908) > 2022-07-12T04:30:49.9915572Z Jul 12 04:30:49 at > org.apache.flink.test.checkpointing.ChangelogPeriodicMaterializationSwitchStateBackendITCase.testSwitchFromDisablingToEnablingInClaimMode(ChangelogPeriodicMaterializationSwitchStateBackendITCase.java:125) > 2022-07-12T04:30:49.9916483Z Jul 12 04:30:49 at > sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > 2022-07-12T04:30:49.9917377Z Jul 12 04:30:49 at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > 2022-07-12T04:30:49.9918121Z Jul 12 04:30:49 at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > 2022-07-12T04:30:49.9918788Z Jul 12 04:30:49 at > java.lang.reflect.Method.invoke(Method.java:498) > 2022-07-12T04:30:49.9919456Z Jul 12 04:30:49 at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) > 2022-07-12T04:30:49.9920193Z Jul 12 04:30:49 at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) > 2022-07-12T04:30:49.9920923Z Jul 12 04:30:49 at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) > 2022-07-12T04:30:49.9921630Z Jul 12 04:30:49 at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) > 2022-07-12T04:30:49.9922326Z Jul 12 04:30:49 at > org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26) > 2022-07-12T04:30:49.9923023Z Jul 12 04:30:49 at > org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27) > 2022-07-12T04:30:49.9923708Z Jul 12 04:30:49 at > org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:54) > 2022-07-12T04:30:49.9924449Z Jul 12 04:30:49 at > org.apache.flink.util.TestNameProvider$1.evaluate(TestNameProvider.java:45) > 2022-07-12T04:30:49.9925124Z Jul 12 04:30:49 at > org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:61) > 2022-07-12T04:30:49.9925912Z Jul 12 04:30:49 at > org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) > 2022-07-12T04:30:49.9926742Z Jul 12 04:30:49 at > org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) > 2022-07-12T04:30:49.9928142Z Jul 12 04:30:49 at > org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) > 2022-07-12T04:30:49.9928715Z Jul 12 04:30:49 at > org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) > 2022-07-12T04:30:49.9929311Z Jul 12 04:30:49 at > org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) > 2022-07-12T04:30:49.9929863Z Jul 12 04:30:49 at > org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) > 2022-07-12T04:30:49.9930376Z Jul 12 04:30:49 at > org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) > 2022-07-12T04:30:49.9930911Z Jul 12 04:30:49 at > org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) > 2022-07-12T04:30:49.9931441Z Jul 12 04:30:49 at > org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) > 2022-07-12T04:30:49.9931975Z Jul 12 04:30:49 at >
[jira] [Commented] (FLINK-28855) ThriftObjectConversions compile failed
[ https://issues.apache.org/jira/browse/FLINK-28855?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17577080#comment-17577080 ] Huang Xingbo commented on FLINK-28855: -- https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=39675=logs=b1fcf054-9138-5463-c73c-a49979b9ac2a=9291ac46-dd95-5135-b799-3839e65a8691=7350 > ThriftObjectConversions compile failed > -- > > Key: FLINK-28855 > URL: https://issues.apache.org/jira/browse/FLINK-28855 > Project: Flink > Issue Type: Bug > Components: Connectors / Hive, Table SQL / Gateway >Affects Versions: 1.16.0 >Reporter: Huang Xingbo >Assignee: yuzelin >Priority: Critical > Labels: test-stability > Fix For: 1.16.0 > > > {code:java} > 2022-08-08T00:32:45.5104326Z [ERROR] > /home/vsts/work/1/s/flink-connectors/flink-connector-hive/src/main/java/org/apache/flink/table/endpoint/hive/util/ThriftObjectConversions.java:[615,31] > cannot find symbol > 2022-08-08T00:32:45.5105191Z symbol: variable INDEX_TABLE > 2022-08-08T00:32:45.5107273Z location: class > org.apache.hadoop.hive.metastore.TableType > {code} > https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=39514=logs=87489130-75dc-54e4-1f45-80c30aa367a3=4632ba9d-f1f2-5ad2-13fc-828d0e28bac4 -- This message was sent by Atlassian Jira (v8.20.10#820010)
[GitHub] [flink] xintongsong commented on pull request #20456: [FLINK-28785][network] Hybrid shuffle consumer thread and upstream thread may have deadlock
xintongsong commented on PR #20456: URL: https://github.com/apache/flink/pull/20456#issuecomment-1208832328 @flinkbot run azure -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Closed] (FLINK-28164) Introduce utilities API for REST endpint
[ https://issues.apache.org/jira/browse/FLINK-28164?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xintong Song closed FLINK-28164. Resolution: Done master (1.16): 41bde65874c24ab036a016d17eb5f1ec09d6f819 > Introduce utilities API for REST endpint > > > Key: FLINK-28164 > URL: https://issues.apache.org/jira/browse/FLINK-28164 > Project: Flink > Issue Type: Sub-task > Components: Table SQL / Gateway >Affects Versions: 1.16.0 >Reporter: Shengkai Fang >Assignee: Wencong Liu >Priority: Major > > It includes heartbeat, get_info, api_versions API in the REST endpoint. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Closed] (FLINK-28163) Introduce the statement related API for REST endpoint
[ https://issues.apache.org/jira/browse/FLINK-28163?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xintong Song closed FLINK-28163. Fix Version/s: 1.16.0 Resolution: Done master (1.16): 47970b0435a3ae968c90922721d44fff8edb01a2 > Introduce the statement related API for REST endpoint > - > > Key: FLINK-28163 > URL: https://issues.apache.org/jira/browse/FLINK-28163 > Project: Flink > Issue Type: Sub-task > Components: Table SQL / Gateway >Affects Versions: 1.16.0 >Reporter: Shengkai Fang >Assignee: Wencong Liu >Priority: Major > Fix For: 1.16.0 > > > It includes executeStatement, fetchResults API in the FLIP-91. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (FLINK-28164) Introduce utilities API for REST endpint
[ https://issues.apache.org/jira/browse/FLINK-28164?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xintong Song updated FLINK-28164: - Fix Version/s: 1.16.0 > Introduce utilities API for REST endpint > > > Key: FLINK-28164 > URL: https://issues.apache.org/jira/browse/FLINK-28164 > Project: Flink > Issue Type: Sub-task > Components: Table SQL / Gateway >Affects Versions: 1.16.0 >Reporter: Shengkai Fang >Assignee: Wencong Liu >Priority: Major > Fix For: 1.16.0 > > > It includes heartbeat, get_info, api_versions API in the REST endpoint. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Closed] (FLINK-28162) Introduce the operation related API for REST endpoint
[ https://issues.apache.org/jira/browse/FLINK-28162?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xintong Song closed FLINK-28162. Fix Version/s: 1.16.0 Resolution: Done master (1.16): 91eccea38c88abf38504094046aa99dbb1f9c06e > Introduce the operation related API for REST endpoint > - > > Key: FLINK-28162 > URL: https://issues.apache.org/jira/browse/FLINK-28162 > Project: Flink > Issue Type: Sub-task > Components: Table SQL / Gateway >Reporter: Shengkai Fang >Assignee: Wencong Liu >Priority: Major > Labels: pull-request-available > Fix For: 1.16.0 > > > It includes getOperationStatus, cancelOperation, closeOperation. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Closed] (FLINK-28631) Allow to GetFunctions in the HiveServer2 Endpoint
[ https://issues.apache.org/jira/browse/FLINK-28631?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Shengkai Fang closed FLINK-28631. - Resolution: Implemented Merged into master: def2f5438090779fa23862027de9b1c4db36b21f > Allow to GetFunctions in the HiveServer2 Endpoint > - > > Key: FLINK-28631 > URL: https://issues.apache.org/jira/browse/FLINK-28631 > Project: Flink > Issue Type: Sub-task >Reporter: Shengkai Fang >Assignee: Shengkai Fang >Priority: Major > Labels: pull-request-available > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Closed] (FLINK-28799) Hybrid shuffle can't schedule graph contains blocking edge
[ https://issues.apache.org/jira/browse/FLINK-28799?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xintong Song closed FLINK-28799. Resolution: Fixed master (1.16): 4deaf6edc152d06488f8738386b6a8b7544fe5e9 > Hybrid shuffle can't schedule graph contains blocking edge > -- > > Key: FLINK-28799 > URL: https://issues.apache.org/jira/browse/FLINK-28799 > Project: Flink > Issue Type: Bug > Components: Runtime / Coordination >Affects Versions: 1.16.0 >Reporter: Weijie Guo >Assignee: Weijie Guo >Priority: Critical > Labels: pull-request-available > Fix For: 1.16.0 > > > Based on TPC-DS test, we found that hybrid shuffle can't schedule graph > contains blocking edge. The reason is that some batch operators will forcibly > set the exchange mode to blocking, which breaks ALL_ EDGE_HYBRID‘s constraint > makes the scheduling deadlock. > We should think of a better way to support the scheduling the graph of all > kinds of edges, including hybrid edge. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[GitHub] [flink] xintongsong closed pull request #20487: [FLINK-28799] Hybrid shuffle should supports schedule the graph contains blocking edge
xintongsong closed pull request #20487: [FLINK-28799] Hybrid shuffle should supports schedule the graph contains blocking edge URL: https://github.com/apache/flink/pull/20487 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] lincoln-lil commented on pull request #20324: [FLINK-28568][table-runtime] Implements a new lookup join operator (sync mode only) with state to eliminate the non determinism
lincoln-lil commented on PR #20324: URL: https://github.com/apache/flink/pull/20324#issuecomment-1208829686 An irrelevant failure case of es sink https://issues.apache.org/jira/browse/FLINK-28877 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] xintongsong closed pull request #20451: [FLINK-28162][sql-gateway]Introduce the operation related API for REST endpoint
xintongsong closed pull request #20451: [FLINK-28162][sql-gateway]Introduce the operation related API for REST endpoint URL: https://github.com/apache/flink/pull/20451 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] fsk119 closed pull request #20479: [FLINK-28631][sql-gateway][hive] Support to GetFunctions in the HiveServer2Endpoint
fsk119 closed pull request #20479: [FLINK-28631][sql-gateway][hive] Support to GetFunctions in the HiveServer2Endpoint URL: https://github.com/apache/flink/pull/20479 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Created] (FLINK-28879) New File Sink s3 end-to-end test failed with Output hash mismatch
Huang Xingbo created FLINK-28879: Summary: New File Sink s3 end-to-end test failed with Output hash mismatch Key: FLINK-28879 URL: https://issues.apache.org/jira/browse/FLINK-28879 Project: Flink Issue Type: Bug Components: API / DataStream, Connectors / FileSystem, Tests Affects Versions: 1.16.0 Reporter: Huang Xingbo {code:java} 2022-08-09T00:50:02.8229585Z Aug 09 00:50:02 FAIL File Streaming Sink: Output hash mismatch. Got 6037b01ca0ffc61a95c12cb475c661a8, expected 6727342fdd3aae2129e61fc8f433fb6f. 2022-08-09T00:50:02.8230700Z Aug 09 00:50:02 head hexdump of actual: 2022-08-09T00:50:02.8477319Z Aug 09 00:50:02 000 E r r o r e x e c u t i n g 2022-08-09T00:50:02.8478206Z Aug 09 00:50:02 010 a w s c o m m a n d : s 3 2022-08-09T00:50:02.8479475Z Aug 09 00:50:02 020 c p - - q u i e t s 3 : / / 2022-08-09T00:50:02.8480205Z Aug 09 00:50:02 030 f l i n k - i n t e g r a t i o 2022-08-09T00:50:02.8480924Z Aug 09 00:50:02 040 n - t e s t s / t e m p / t e s 2022-08-09T00:50:02.8481612Z Aug 09 00:50:02 050 t _ f i l e _ s i n k - 1 d 3 d 2022-08-09T00:50:02.8483048Z Aug 09 00:50:02 060 4 0 0 8 - b 0 b f - 4 2 6 5 - b 2022-08-09T00:50:02.8483618Z Aug 09 00:50:02 070 e 0 e - 3 b 9 f 7 8 2 c 5 5 2 d 2022-08-09T00:50:02.8484222Z Aug 09 00:50:02 080 / h o s t d i r / / t e m p - 2022-08-09T00:50:02.8484831Z Aug 09 00:50:02 090 t e s t - d i r e c t o r y - 2 2022-08-09T00:50:02.8485719Z Aug 09 00:50:02 0a0 3 9 3 7 7 8 2 6 8 0 / t e m p / 2022-08-09T00:50:02.8486427Z Aug 09 00:50:02 0b0 t e s t _ f i l e _ s i n k - 1 2022-08-09T00:50:02.8487134Z Aug 09 00:50:02 0c0 d 3 d 4 0 0 8 - b 0 b f - 4 2 6 2022-08-09T00:50:02.8487826Z Aug 09 00:50:02 0d0 5 - b e 0 e - 3 b 9 f 7 8 2 c 5 2022-08-09T00:50:02.8488511Z Aug 09 00:50:02 0e0 5 2 d - - e x c l u d e ' * 2022-08-09T00:50:02.8489202Z Aug 09 00:50:02 0f0 ' - - i n c l u d e ' * / p 2022-08-09T00:50:02.8489891Z Aug 09 00:50:02 100 a r t - [ ! / ] * ' - - r e c 2022-08-09T00:50:02.8490385Z Aug 09 00:50:02 110 u r s i v e \n 2022-08-09T00:50:02.8490822Z Aug 09 00:50:02 117 2022-08-09T00:50:02.8502212Z Aug 09 00:50:02 Stopping job timeout watchdog (with pid=141134) 2022-08-09T00:50:06.8430959Z rm: cannot remove '/home/vsts/work/1/s/flink-dist/target/flink-1.16-SNAPSHOT-bin/flink-1.16-SNAPSHOT/lib/flink-shaded-netty-tcnative-static-*.jar': No such file or directory 2022-08-09T00:50:06.9278248Z Aug 09 00:50:06 5ccfeb22307c2a88625a38b9537acc79001d1b29094ef40fd70692ce11407502 2022-08-09T00:50:06.9618147Z Aug 09 00:50:06 5ccfeb22307c2a88625a38b9537acc79001d1b29094ef40fd70692ce11407502 2022-08-09T00:50:06.9645077Z Aug 09 00:50:06 [FAIL] Test script contains errors. 2022-08-09T00:50:06.9666227Z Aug 09 00:50:06 Checking of logs skipped. 2022-08-09T00:50:06.9671891Z Aug 09 00:50:06 2022-08-09T00:50:06.9673050Z Aug 09 00:50:06 [FAIL] 'New File Sink s3 end-to-end test' failed after 3 minutes and 42 seconds! Test exited with exit code 1 {code} https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=39667=logs=af184cdd-c6d8-5084-0b69-7e9c67b35f7a=160c9ae5-96fd-516e-1c91-deb81f59292a=4136 -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (FLINK-28860) CacheITCase.testBatchProduceCacheStreamConsume failed
[ https://issues.apache.org/jira/browse/FLINK-28860?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17577068#comment-17577068 ] godfrey he commented on FLINK-28860: https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=39618=logs=a57e0635-3fad-5b08-57c7-a4142d7d6fa9=2ef0effc-1da1-50e5-c2bd-aab434b1c5b7 > CacheITCase.testBatchProduceCacheStreamConsume failed > - > > Key: FLINK-28860 > URL: https://issues.apache.org/jira/browse/FLINK-28860 > Project: Flink > Issue Type: Bug > Components: API / DataStream >Affects Versions: 1.16.0 >Reporter: Huang Xingbo >Priority: Critical > Labels: test-stability > Fix For: 1.16.0 > > > {code:java} > 2022-08-08T03:27:22.1988575Z Aug 08 03:27:22 [ERROR] > org.apache.flink.test.streaming.runtime.CacheITCase.testBatchProduceCacheStreamConsume(Path) > Time elapsed: 0.593 s <<< ERROR! > 2022-08-08T03:27:22.1989338Z Aug 08 03:27:22 java.lang.RuntimeException: > Producing cache IntermediateResult is not supported in streaming mode > 2022-08-08T03:27:22.1990401Z Aug 08 03:27:22 at > org.apache.flink.streaming.runtime.translators.CacheTransformationTranslator.translateForStreamingInternal(CacheTransformationTranslator.java:75) > 2022-08-08T03:27:22.1991511Z Aug 08 03:27:22 at > org.apache.flink.streaming.runtime.translators.CacheTransformationTranslator.translateForStreamingInternal(CacheTransformationTranslator.java:42) > 2022-08-08T03:27:22.1993671Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.graph.SimpleTransformationTranslator.translateForStreaming(SimpleTransformationTranslator.java:62) > 2022-08-08T03:27:22.1994900Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.graph.StreamGraphGenerator.translate(StreamGraphGenerator.java:830) > 2022-08-08T03:27:22.1995748Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.graph.StreamGraphGenerator.transform(StreamGraphGenerator.java:560) > 2022-08-08T03:27:22.1996932Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.graph.StreamGraphGenerator.getParentInputIds(StreamGraphGenerator.java:851) > 2022-08-08T03:27:22.1998562Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.graph.StreamGraphGenerator.translate(StreamGraphGenerator.java:809) > 2022-08-08T03:27:22.1999581Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.graph.StreamGraphGenerator.transform(StreamGraphGenerator.java:560) > 2022-08-08T03:27:22.2000376Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.graph.StreamGraphGenerator.generate(StreamGraphGenerator.java:319) > 2022-08-08T03:27:22.2001359Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.getStreamGraph(StreamExecutionEnvironment.java:2250) > 2022-08-08T03:27:22.2002767Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.getStreamGraph(StreamExecutionEnvironment.java:2241) > 2022-08-08T03:27:22.2004121Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.getStreamGraph(StreamExecutionEnvironment.java:2227) > 2022-08-08T03:27:22.2005059Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.executeAsync(StreamExecutionEnvironment.java:2178) > 2022-08-08T03:27:22.2005939Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.datastream.DataStream.executeAndCollectWithClient(DataStream.java:1469) > 2022-08-08T03:27:22.2006735Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.datastream.DataStream.executeAndCollect(DataStream.java:1334) > 2022-08-08T03:27:22.2007500Z Aug 08 03:27:22 at > org.apache.flink.streaming.api.datastream.DataStream.executeAndCollect(DataStream.java:1320) > 2022-08-08T03:27:22.2008315Z Aug 08 03:27:22 at > org.apache.flink.test.streaming.runtime.CacheITCase.testBatchProduceCacheStreamConsume(CacheITCase.java:190) > {code} > https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=39518=logs=a57e0635-3fad-5b08-57c7-a4142d7d6fa9=2ef0effc-1da1-50e5-c2bd-aab434b1c5b7 -- This message was sent by Atlassian Jira (v8.20.10#820010)
[GitHub] [flink] godfreyhe commented on pull request #20432: [FLINK-28753][table-planner] Improve FilterIntoJoinRule which could push some predicates to another side
godfreyhe commented on PR #20432: URL: https://github.com/apache/flink/pull/20432#issuecomment-1208825729 The failure is due to https://issues.apache.org/jira/browse/FLINK-28860 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Closed] (FLINK-28380) Produce one intermediate dataset for multiple consumers consuming the same data
[ https://issues.apache.org/jira/browse/FLINK-28380?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Yingjie Cao closed FLINK-28380. --- Resolution: Fixed Merged into master via 60bc87c0b83149c4f19d7e54af2d967087a277fb > Produce one intermediate dataset for multiple consumers consuming the same > data > --- > > Key: FLINK-28380 > URL: https://issues.apache.org/jira/browse/FLINK-28380 > Project: Flink > Issue Type: Sub-task > Components: Client / Job Submission, Runtime / Coordination, Runtime > / Network >Reporter: Yingjie Cao >Assignee: Yingjie Cao >Priority: Major > Labels: pull-request-available > Fix For: 1.16.0 > > > Currently, if one output of an upstream job vertex is consumed by multiple > downstream job vertices, the upstream vertex will produce multiple dataset. > For blocking shuffle, it means serialize and persist the same data multiple > times. This ticket aims to optimize this behavior and make the upstream job > vertex produce one dataset which will be read by multiple downstream vertex. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Closed] (FLINK-24614) Add array,map,row types support for parquet vectorized reader
[ https://issues.apache.org/jira/browse/FLINK-24614?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jark Wu closed FLINK-24614. --- Fix Version/s: 1.16.0 Resolution: Fixed Fixed in master: 5c731a37e1a8f71f9c9e813f6c741a1e203fa1a3 > Add array,map,row types support for parquet vectorized reader > - > > Key: FLINK-24614 > URL: https://issues.apache.org/jira/browse/FLINK-24614 > Project: Flink > Issue Type: Sub-task > Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile) >Affects Versions: 1.14.0 >Reporter: sujun >Assignee: dalongliu >Priority: Major > Labels: pull-request-available, stale-assigned > Fix For: 1.16.0 > > > Add array,map,row types support for parquet vectorized reader -- This message was sent by Atlassian Jira (v8.20.10#820010)
[GitHub] [flink] wuchong merged pull request #20442: [FLINK-24614][Connectors/PARQUET] Add Complex types support for parquet reader
wuchong merged PR #20442: URL: https://github.com/apache/flink/pull/20442 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-28878) PipelinedRegionSchedulingITCase.testRecoverFromPartitionException failed with AssertionError
[ https://issues.apache.org/jira/browse/FLINK-28878?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17577063#comment-17577063 ] Huang Xingbo commented on FLINK-28878: -- cc [~zhuzh] > PipelinedRegionSchedulingITCase.testRecoverFromPartitionException failed with > AssertionError > > > Key: FLINK-28878 > URL: https://issues.apache.org/jira/browse/FLINK-28878 > Project: Flink > Issue Type: Bug > Components: Runtime / Coordination >Affects Versions: 1.16.0 >Reporter: Huang Xingbo >Priority: Major > Labels: test-stability > > {code:java} > 2022-08-08T20:38:43.3934646Z Aug 08 20:38:43 [ERROR] > org.apache.flink.test.scheduling.PipelinedRegionSchedulingITCase.testRecoverFromPartitionException > Time elapsed: 20.288 s <<< FAILURE! > 2022-08-08T20:38:43.3935309Z Aug 08 20:38:43 java.lang.AssertionError: > 2022-08-08T20:38:43.3937070Z Aug 08 20:38:43 > 2022-08-08T20:38:43.3938015Z Aug 08 20:38:43 Expected: is > 2022-08-08T20:38:43.3940277Z Aug 08 20:38:43 but: was > 2022-08-08T20:38:43.3940927Z Aug 08 20:38:43 at > org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) > 2022-08-08T20:38:43.3941571Z Aug 08 20:38:43 at > org.junit.Assert.assertThat(Assert.java:964) > 2022-08-08T20:38:43.3942120Z Aug 08 20:38:43 at > org.junit.Assert.assertThat(Assert.java:930) > 2022-08-08T20:38:43.3943202Z Aug 08 20:38:43 at > org.apache.flink.test.scheduling.PipelinedRegionSchedulingITCase.testRecoverFromPartitionException(PipelinedRegionSchedulingITCase.java:98) > {code} > https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=39652=logs=a57e0635-3fad-5b08-57c7-a4142d7d6fa9=2ef0effc-1da1-50e5-c2bd-aab434b1c5b7=9994 -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Reopened] (FLINK-27916) HybridSourceReaderTest.testReader failed with AssertionError
[ https://issues.apache.org/jira/browse/FLINK-27916?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Huang Xingbo reopened FLINK-27916: -- > HybridSourceReaderTest.testReader failed with AssertionError > > > Key: FLINK-27916 > URL: https://issues.apache.org/jira/browse/FLINK-27916 > Project: Flink > Issue Type: Bug > Components: Connectors / Common >Affects Versions: 1.16.0 >Reporter: Huang Xingbo >Priority: Critical > Labels: test-stability > Attachments: Screen Shot 2022-07-21 at 5.51.40 PM.png > > > {code:java} > 2022-06-05T07:47:33.3332158Z Jun 05 07:47:33 [ERROR] Tests run: 3, Failures: > 1, Errors: 0, Skipped: 0, Time elapsed: 2.03 s <<< FAILURE! - in > org.apache.flink.connector.base.source.hybrid.HybridSourceReaderTest > 2022-06-05T07:47:33.3334366Z Jun 05 07:47:33 [ERROR] > org.apache.flink.connector.base.source.hybrid.HybridSourceReaderTest.testReader > Time elapsed: 0.108 s <<< FAILURE! > 2022-06-05T07:47:33.3335385Z Jun 05 07:47:33 java.lang.AssertionError: > 2022-06-05T07:47:33.3336049Z Jun 05 07:47:33 > 2022-06-05T07:47:33.3336682Z Jun 05 07:47:33 Expected size: 1 but was: 0 in: > 2022-06-05T07:47:33.3337316Z Jun 05 07:47:33 [] > 2022-06-05T07:47:33.3338437Z Jun 05 07:47:33 at > org.apache.flink.connector.base.source.hybrid.HybridSourceReaderTest.assertAndClearSourceReaderFinishedEvent(HybridSourceReaderTest.java:199) > 2022-06-05T07:47:33.3340082Z Jun 05 07:47:33 at > org.apache.flink.connector.base.source.hybrid.HybridSourceReaderTest.testReader(HybridSourceReaderTest.java:96) > 2022-06-05T07:47:33.3341373Z Jun 05 07:47:33 at > java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > 2022-06-05T07:47:33.3342540Z Jun 05 07:47:33 at > java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > 2022-06-05T07:47:33.3344124Z Jun 05 07:47:33 at > java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > 2022-06-05T07:47:33.3345283Z Jun 05 07:47:33 at > java.base/java.lang.reflect.Method.invoke(Method.java:566) > 2022-06-05T07:47:33.3346804Z Jun 05 07:47:33 at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) > 2022-06-05T07:47:33.3348218Z Jun 05 07:47:33 at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) > 2022-06-05T07:47:33.3349495Z Jun 05 07:47:33 at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) > 2022-06-05T07:47:33.3350779Z Jun 05 07:47:33 at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) > 2022-06-05T07:47:33.3351956Z Jun 05 07:47:33 at > org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) > 2022-06-05T07:47:33.3357032Z Jun 05 07:47:33 at > org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) > 2022-06-05T07:47:33.3358633Z Jun 05 07:47:33 at > org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) > 2022-06-05T07:47:33.3360003Z Jun 05 07:47:33 at > org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) > 2022-06-05T07:47:33.3361924Z Jun 05 07:47:33 at > org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) > 2022-06-05T07:47:33.3363427Z Jun 05 07:47:33 at > org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) > 2022-06-05T07:47:33.3364793Z Jun 05 07:47:33 at > org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) > 2022-06-05T07:47:33.3365619Z Jun 05 07:47:33 at > org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) > 2022-06-05T07:47:33.3366254Z Jun 05 07:47:33 at > org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) > 2022-06-05T07:47:33.3366939Z Jun 05 07:47:33 at > org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) > 2022-06-05T07:47:33.3367556Z Jun 05 07:47:33 at > org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) > 2022-06-05T07:47:33.3368268Z Jun 05 07:47:33 at > org.junit.runners.ParentRunner.run(ParentRunner.java:413) > 2022-06-05T07:47:33.3369166Z Jun 05 07:47:33 at > org.junit.runner.JUnitCore.run(JUnitCore.java:137) > 2022-06-05T07:47:33.3369993Z Jun 05 07:47:33 at > org.junit.runner.JUnitCore.run(JUnitCore.java:115) > 2022-06-05T07:47:33.3371021Z Jun 05 07:47:33 at > org.junit.vintage.engine.execution.RunnerExecutor.execute(RunnerExecutor.java:42) > 2022-06-05T07:47:33.3372128Z Jun 05 07:47:33 at > org.junit.vintage.engine.VintageTestEngine.executeAllChildren(VintageTestEngine.java:80) > 2022-06-05T07:47:33.3373622Z Jun 05 07:47:33 at > org.junit.vintage.engine.VintageTestEngine.execute(VintageTestEngine.java:72) > 2022-06-05T07:47:33.3374886Z Jun 05
[GitHub] [flink] wsry merged pull request #20351: [FLINK-28380][runtime] Produce one intermediate dataset for multiple consumer job vertices consuming the same data
wsry merged PR #20351: URL: https://github.com/apache/flink/pull/20351 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] wsry commented on pull request #20351: [FLINK-28380][runtime] Produce one intermediate dataset for multiple consumer job vertices consuming the same data
wsry commented on PR #20351: URL: https://github.com/apache/flink/pull/20351#issuecomment-1208823627 Tests passed: https://dev.azure.com/kevin-flink/flink/_build/results?buildId=624=results, merging. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Created] (FLINK-28878) PipelinedRegionSchedulingITCase.testRecoverFromPartitionException failed with AssertionError
Huang Xingbo created FLINK-28878: Summary: PipelinedRegionSchedulingITCase.testRecoverFromPartitionException failed with AssertionError Key: FLINK-28878 URL: https://issues.apache.org/jira/browse/FLINK-28878 Project: Flink Issue Type: Bug Components: Runtime / Coordination Affects Versions: 1.16.0 Reporter: Huang Xingbo {code:java} 2022-08-08T20:38:43.3934646Z Aug 08 20:38:43 [ERROR] org.apache.flink.test.scheduling.PipelinedRegionSchedulingITCase.testRecoverFromPartitionException Time elapsed: 20.288 s <<< FAILURE! 2022-08-08T20:38:43.3935309Z Aug 08 20:38:43 java.lang.AssertionError: 2022-08-08T20:38:43.3937070Z Aug 08 20:38:43 2022-08-08T20:38:43.3938015Z Aug 08 20:38:43 Expected: is 2022-08-08T20:38:43.3940277Z Aug 08 20:38:43 but: was 2022-08-08T20:38:43.3940927Z Aug 08 20:38:43at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20) 2022-08-08T20:38:43.3941571Z Aug 08 20:38:43at org.junit.Assert.assertThat(Assert.java:964) 2022-08-08T20:38:43.3942120Z Aug 08 20:38:43at org.junit.Assert.assertThat(Assert.java:930) 2022-08-08T20:38:43.3943202Z Aug 08 20:38:43at org.apache.flink.test.scheduling.PipelinedRegionSchedulingITCase.testRecoverFromPartitionException(PipelinedRegionSchedulingITCase.java:98) {code} https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=39652=logs=a57e0635-3fad-5b08-57c7-a4142d7d6fa9=2ef0effc-1da1-50e5-c2bd-aab434b1c5b7=9994 -- This message was sent by Atlassian Jira (v8.20.10#820010)
[GitHub] [flink] wsry commented on pull request #20333: [FLINK-28623][network] Optimize the use of off heap memory by blocking and hybrid shuffle reader
wsry commented on PR #20333: URL: https://github.com/apache/flink/pull/20333#issuecomment-1208822800 @reswqa Thanks for the change. I left a minor comment. Please rebase master and fix the conflicts. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] wsry commented on a diff in pull request #20333: [FLINK-28623][network] Optimize the use of off heap memory by blocking and hybrid shuffle reader
wsry commented on code in PR #20333: URL: https://github.com/apache/flink/pull/20333#discussion_r940808314 ## flink-runtime/src/main/java/org/apache/flink/runtime/io/network/partition/SortMergeResultPartitionReadScheduler.java: ## @@ -70,6 +71,17 @@ class SortMergeResultPartitionReadScheduler implements Runnable, BufferRecycler */ private static final Duration DEFAULT_BUFFER_REQUEST_TIMEOUT = Duration.ofMinutes(5); +/** Used to read buffers from file channel. */ Review Comment: buffers -> buffer headers -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] yunfengzhou-hub commented on a diff in pull request #20454: [FLINK-28639][Runtime/Checkpointing] Preserve consistency of events from subtask to OC
yunfengzhou-hub commented on code in PR #20454: URL: https://github.com/apache/flink/pull/20454#discussion_r940807420 ## flink-runtime/src/main/java/org/apache/flink/runtime/operators/coordination/OperatorCoordinatorHolder.java: ## @@ -204,18 +228,52 @@ public void close() throws Exception { public void handleEventFromOperator(int subtask, int attemptNumber, OperatorEvent event) throws Exception { mainThreadExecutor.assertRunningInMainThread(); + +if (event instanceof AcknowledgeCloseGatewayEvent) { +Preconditions.checkArgument( +subtask == ((AcknowledgeCloseGatewayEvent) event).getSubtaskIndex()); +completeAcknowledgeCloseGatewayFuture( +subtask, ((AcknowledgeCloseGatewayEvent) event).getCheckpointID()); +return; +} else if (event instanceof AcknowledgeCheckpointEvent) { +Preconditions.checkArgument( +subtask == ((AcknowledgeCheckpointEvent) event).getSubtaskIndex()); +subtaskGatewayMap +.get(subtask) +.openGatewayAndUnmarkCheckpoint( +((AcknowledgeCheckpointEvent) event).getCheckpointID()); +return; +} + coordinator.handleEventFromOperator(subtask, attemptNumber, event); } public void executionAttemptFailed(int subtask, int attemptNumber, @Nullable Throwable reason) { mainThreadExecutor.assertRunningInMainThread(); + +if (!context.isConcurrentExecutionAttemptsSupported() +&& acknowledgeCloseGatewayFutureMap.containsKey(subtask)) { +Exception exception = +new FlinkException(String.format("Subtask %d has failed.", subtask), reason); + acknowledgeCloseGatewayFutureMap.get(subtask).completeExceptionally(exception); Review Comment: There is no guarantee that a subtask would not fail or be reset during checkpoints. If this happens, the coordinator might not receive ACK events from that subtask and the checkpoint process might be infinitely blocked. Thus we need to update the map here to make sure that the checkpoint can be correctly aborted in the situation described above. Same for that in `subtaskReset`. According to our offline discussion, I'll add a method here to wrap this part of logic and reuse the method in `executionAttemptFailed()` and `subtaskReset()`. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-27916) HybridSourceReaderTest.testReader failed with AssertionError
[ https://issues.apache.org/jira/browse/FLINK-27916?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17577062#comment-17577062 ] Huang Xingbo commented on FLINK-27916: -- https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=39636=logs=aa18c3f6-13b8-5f58-86bb-c1cffb239496=502fb6c0-30a2-5e49-c5c2-a00fa3acb203 new instance > HybridSourceReaderTest.testReader failed with AssertionError > > > Key: FLINK-27916 > URL: https://issues.apache.org/jira/browse/FLINK-27916 > Project: Flink > Issue Type: Bug > Components: Connectors / Common >Affects Versions: 1.16.0 >Reporter: Huang Xingbo >Priority: Critical > Labels: test-stability > Attachments: Screen Shot 2022-07-21 at 5.51.40 PM.png > > > {code:java} > 2022-06-05T07:47:33.3332158Z Jun 05 07:47:33 [ERROR] Tests run: 3, Failures: > 1, Errors: 0, Skipped: 0, Time elapsed: 2.03 s <<< FAILURE! - in > org.apache.flink.connector.base.source.hybrid.HybridSourceReaderTest > 2022-06-05T07:47:33.3334366Z Jun 05 07:47:33 [ERROR] > org.apache.flink.connector.base.source.hybrid.HybridSourceReaderTest.testReader > Time elapsed: 0.108 s <<< FAILURE! > 2022-06-05T07:47:33.3335385Z Jun 05 07:47:33 java.lang.AssertionError: > 2022-06-05T07:47:33.3336049Z Jun 05 07:47:33 > 2022-06-05T07:47:33.3336682Z Jun 05 07:47:33 Expected size: 1 but was: 0 in: > 2022-06-05T07:47:33.3337316Z Jun 05 07:47:33 [] > 2022-06-05T07:47:33.3338437Z Jun 05 07:47:33 at > org.apache.flink.connector.base.source.hybrid.HybridSourceReaderTest.assertAndClearSourceReaderFinishedEvent(HybridSourceReaderTest.java:199) > 2022-06-05T07:47:33.3340082Z Jun 05 07:47:33 at > org.apache.flink.connector.base.source.hybrid.HybridSourceReaderTest.testReader(HybridSourceReaderTest.java:96) > 2022-06-05T07:47:33.3341373Z Jun 05 07:47:33 at > java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > 2022-06-05T07:47:33.3342540Z Jun 05 07:47:33 at > java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > 2022-06-05T07:47:33.3344124Z Jun 05 07:47:33 at > java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > 2022-06-05T07:47:33.3345283Z Jun 05 07:47:33 at > java.base/java.lang.reflect.Method.invoke(Method.java:566) > 2022-06-05T07:47:33.3346804Z Jun 05 07:47:33 at > org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) > 2022-06-05T07:47:33.3348218Z Jun 05 07:47:33 at > org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) > 2022-06-05T07:47:33.3349495Z Jun 05 07:47:33 at > org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) > 2022-06-05T07:47:33.3350779Z Jun 05 07:47:33 at > org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) > 2022-06-05T07:47:33.3351956Z Jun 05 07:47:33 at > org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) > 2022-06-05T07:47:33.3357032Z Jun 05 07:47:33 at > org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) > 2022-06-05T07:47:33.3358633Z Jun 05 07:47:33 at > org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) > 2022-06-05T07:47:33.3360003Z Jun 05 07:47:33 at > org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) > 2022-06-05T07:47:33.3361924Z Jun 05 07:47:33 at > org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) > 2022-06-05T07:47:33.3363427Z Jun 05 07:47:33 at > org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) > 2022-06-05T07:47:33.3364793Z Jun 05 07:47:33 at > org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) > 2022-06-05T07:47:33.3365619Z Jun 05 07:47:33 at > org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) > 2022-06-05T07:47:33.3366254Z Jun 05 07:47:33 at > org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) > 2022-06-05T07:47:33.3366939Z Jun 05 07:47:33 at > org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) > 2022-06-05T07:47:33.3367556Z Jun 05 07:47:33 at > org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) > 2022-06-05T07:47:33.3368268Z Jun 05 07:47:33 at > org.junit.runners.ParentRunner.run(ParentRunner.java:413) > 2022-06-05T07:47:33.3369166Z Jun 05 07:47:33 at > org.junit.runner.JUnitCore.run(JUnitCore.java:137) > 2022-06-05T07:47:33.3369993Z Jun 05 07:47:33 at > org.junit.runner.JUnitCore.run(JUnitCore.java:115) > 2022-06-05T07:47:33.3371021Z Jun 05 07:47:33 at > org.junit.vintage.engine.execution.RunnerExecutor.execute(RunnerExecutor.java:42) > 2022-06-05T07:47:33.3372128Z Jun 05 07:47:33 at >
[GitHub] [flink] pltbkd commented on pull request #20415: [FLINK-28711] Hive connector implements SupportsDynamicFiltering interface
pltbkd commented on PR #20415: URL: https://github.com/apache/flink/pull/20415#issuecomment-1208790289 @flinkbot run azure -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Created] (FLINK-28877) Elasticsearch7DynamicSinkITCase.testWritingDocumentsNoPrimaryKey case failed
lincoln lee created FLINK-28877: --- Summary: Elasticsearch7DynamicSinkITCase.testWritingDocumentsNoPrimaryKey case failed Key: FLINK-28877 URL: https://issues.apache.org/jira/browse/FLINK-28877 Project: Flink Issue Type: Bug Reporter: lincoln lee Fix For: 1.16.0 {code} Aug 08 16:00:39 Caused by: java.lang.RuntimeException: An error occurred in ElasticsearchSink. Aug 08 16:00:39 at org.apache.flink.streaming.connectors.elasticsearch.ElasticsearchSinkBase.checkErrorAndRethrow(ElasticsearchSinkBase.java:426) Aug 08 16:00:39 at org.apache.flink.streaming.connectors.elasticsearch.ElasticsearchSinkBase.close(ElasticsearchSinkBase.java:365) Aug 08 16:00:39 at org.apache.flink.api.common.functions.util.FunctionUtils.closeFunction(FunctionUtils.java:41) Aug 08 16:00:39 at org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator.close(AbstractUdfStreamOperator.java:114) Aug 08 16:00:39 at org.apache.flink.streaming.runtime.tasks.StreamOperatorWrapper.close(StreamOperatorWrapper.java:163) Aug 08 16:00:39 at org.apache.flink.streaming.runtime.tasks.RegularOperatorChain.closeAllOperators(RegularOperatorChain.java:125) Aug 08 16:00:39 at org.apache.flink.streaming.runtime.tasks.StreamTask.closeAllOperators(StreamTask.java:1022) Aug 08 16:00:39 at org.apache.flink.streaming.runtime.tasks.StreamTask.afterInvoke(StreamTask.java:900) Aug 08 16:00:39 at org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:783) Aug 08 16:00:39 at org.apache.flink.runtime.taskmanager.Task.runWithSystemExitMonitoring(Task.java:935) Aug 08 16:00:39 at org.apache.flink.runtime.taskmanager.Task.restoreAndInvoke(Task.java:914) Aug 08 16:00:39 at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:728) Aug 08 16:00:39 at org.apache.flink.runtime.taskmanager.Task.run(Task.java:550) Aug 08 16:00:39 at java.lang.Thread.run(Thread.java:748) Aug 08 16:00:39 Caused by: java.net.SocketTimeoutException: 30,000 milliseconds timeout on connection http-outgoing-0 [ACTIVE] Aug 08 16:00:39 at org.apache.http.nio.protocol.HttpAsyncRequestExecutor.timeout(HttpAsyncRequestExecutor.java:387) Aug 08 16:00:39 at org.apache.http.impl.nio.client.InternalIODispatch.onTimeout(InternalIODispatch.java:92) Aug 08 16:00:39 at org.apache.http.impl.nio.client.InternalIODispatch.onTimeout(InternalIODispatch.java:39) Aug 08 16:00:39 at org.apache.http.impl.nio.reactor.AbstractIODispatch.timeout(AbstractIODispatch.java:175) Aug 08 16:00:39 at org.apache.http.impl.nio.reactor.BaseIOReactor.sessionTimedOut(BaseIOReactor.java:261) Aug 08 16:00:39 at org.apache.http.impl.nio.reactor.AbstractIOReactor.timeoutCheck(AbstractIOReactor.java:502) Aug 08 16:00:39 at org.apache.http.impl.nio.reactor.BaseIOReactor.validate(BaseIOReactor.java:211) Aug 08 16:00:39 at org.apache.http.impl.nio.reactor.AbstractIOReactor.execute(AbstractIOReactor.java:280) Aug 08 16:00:39 at org.apache.http.impl.nio.reactor.BaseIOReactor.execute(BaseIOReactor.java:104) Aug 08 16:00:39 at org.apache.http.impl.nio.reactor.AbstractMultiworkerIOReactor$Worker.run(AbstractMultiworkerIOReactor.java:591) {code} https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=39617=logs=4eda0b4a-bd0d-521a-0916-8285b9be9bb5=2ff6d5fa-53a6-53ac-bff7-fa524ea361a9 -- This message was sent by Atlassian Jira (v8.20.10#820010)
[GitHub] [flink] lincoln-lil commented on pull request #20324: [FLINK-28568][table-runtime] Implements a new lookup join operator (sync mode only) with state to eliminate the non determinism
lincoln-lil commented on PR #20324: URL: https://github.com/apache/flink/pull/20324#issuecomment-1208787436 @lincoln-lil run azure -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] tisonkun commented on a diff in pull request #19972: [FLINK-27399][Connector/Pulsar] Change initial consuming position setting logic for better handle the checkpoint.
tisonkun commented on code in PR #19972: URL: https://github.com/apache/flink/pull/19972#discussion_r940792766 ## flink-connectors/flink-connector-pulsar/src/test/java/org/apache/flink/connector/pulsar/source/PulsarSourceITCase.java: ## @@ -33,8 +33,10 @@ import org.apache.flink.testutils.junit.FailsOnJava11; import org.junit.experimental.categories.Category; +import org.junit.jupiter.api.Disabled; /** Unite test class for {@link PulsarSource}. */ +@Disabled("Could not reproduce on local machine.") Review Comment: +1 This is the heaviest and most unstable test of Pulsar connector. If you simply disable it, you'd better do it in a separate patch with a reference. Otherwise, I'm unconfident to other changes in this patch. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] wsry commented on pull request #20457: [FLINK-28373][network] Read a full buffer of data per file IO read request for sort-shuffle
wsry commented on PR #20457: URL: https://github.com/apache/flink/pull/20457#issuecomment-1208780264 @flinkbot run azure -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[GitHub] [flink] luoyuxia commented on pull request #19556: [FLINK-26413][hive] Hive dialect supports "LOAD DATA INPATH"
luoyuxia commented on PR #19556: URL: https://github.com/apache/flink/pull/19556#issuecomment-1208778544 @flinkbot run azure -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-28765) Create a doc for protobuf format
[ https://issues.apache.org/jira/browse/FLINK-28765?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17577047#comment-17577047 ] Hui Yang commented on FLINK-28765: -- [~maosuhan] can you guide me how I can use the flink-protobuf now with the flink version 1.15? > Create a doc for protobuf format > > > Key: FLINK-28765 > URL: https://issues.apache.org/jira/browse/FLINK-28765 > Project: Flink > Issue Type: Improvement > Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile) >Reporter: Suhan Mao >Assignee: Suhan Mao >Priority: Major > Labels: pull-request-available > Fix For: 1.16.0 > > > After this feature has been done > https://issues.apache.org/jira/browse/FLINK-18202, we should write a doc to > introduce how to use the protobuf format in SQL. > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[GitHub] [flink] luoyuxia commented on pull request #19556: [FLINK-26413][hive] Hive dialect supports "LOAD DATA INPATH"
luoyuxia commented on PR #19556: URL: https://github.com/apache/flink/pull/19556#issuecomment-1208716644 @flinkbot run azure -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-28865) Add updated Print sink for new interfaces
[ https://issues.apache.org/jira/browse/FLINK-28865?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chesnay Schepler updated FLINK-28865: - Description: The built-in print sink still uses the old sink interfaces. Add a new implementation for the new sink interfaces. (was: The built-in print sink still uses the old sink interfaces. Add a new implementation for the new sink interfaces and deprecate the old sink.) > Add updated Print sink for new interfaces > - > > Key: FLINK-28865 > URL: https://issues.apache.org/jira/browse/FLINK-28865 > Project: Flink > Issue Type: Improvement > Components: API / DataStream >Reporter: Chesnay Schepler >Assignee: Chesnay Schepler >Priority: Major > Labels: pull-request-available > Fix For: 1.16.0 > > > The built-in print sink still uses the old sink interfaces. Add a new > implementation for the new sink interfaces. -- This message was sent by Atlassian Jira (v8.20.10#820010)