HunterHunter created FLINK-29032:
------------------------------------
Summary: Kafka Consume from timestamp catch exception : Caused by:
java.lang.IllegalArgumentException: Invalid negative offset
Key: FLINK-29032
URL: https://issues.apache.org/jira/browse/FLINK-29032
Project: Flink
Issue Type: Bug
Components: Connectors / Kafka
Affects Versions: 1.15.1
Reporter: HunterHunter
/*+ OPTIONS(
'scan.startup.mode' = 'timestamp',
'scan.startup.timestamp-millis'='1660809660000',
) */;
{code:java}
org.apache.flink.util.FlinkRuntimeException: Failed to initialize partition
splits due to
at
org.apache.flink.connector.kafka.source.enumerator.KafkaSourceEnumerator.handlePartitionSplitChanges(KafkaSourceEnumerator.java:299)
~[flink-sql-connector-kafka-1.15.1-vip.jar:1.15.1-vip]
at
org.apache.flink.runtime.source.coordinator.ExecutorNotifier.lambda$null$1(ExecutorNotifier.java:83)
~[flink-dist-1.15.1-vip.jar:1.15.1-vip]
at
org.apache.flink.util.ThrowableCatchingRunnable.run(ThrowableCatchingRunnable.java:40)
[flink-dist-1.15.1-vip.jar:1.15.1-vip]
at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
[?:1.8.0_201]
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
[?:1.8.0_201]
at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
[?:1.8.0_201]
at
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
[?:1.8.0_201]
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
[?:1.8.0_201]
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
[?:1.8.0_201]
at java.lang.Thread.run(Thread.java:748) [?:1.8.0_201]
Caused by: java.lang.IllegalArgumentException: Invalid negative offset
at
org.apache.flink.kafka.shaded.org.apache.kafka.clients.consumer.OffsetAndTimestamp.<init>(OffsetAndTimestamp.java:36)
~[flink-sql-connector-kafka-1.15.1-vip.jar:1.15.1-vip]
at
org.apache.flink.connector.kafka.source.enumerator.KafkaSourceEnumerator$PartitionOffsetsRetrieverImpl.lambda$offsetsForTimes$8(KafkaSourceEnumerator.java:622)
~[flink-sql-connector-kafka-1.15.1-vip.jar:1.15.1-vip]
at java.util.stream.Collectors.lambda$toMap$58(Collectors.java:1321)
~[?:1.8.0_201]
at java.util.stream.ReduceOps$3ReducingSink.accept(ReduceOps.java:169)
~[?:1.8.0_201]
at
java.util.HashMap$EntrySpliterator.forEachRemaining(HashMap.java:1699)
~[?:1.8.0_201]
at
java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
~[?:1.8.0_201]
at
java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
~[?:1.8.0_201]
at
java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
~[?:1.8.0_201]
at
java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
~[?:1.8.0_201]
at
java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
~[?:1.8.0_201]
at
org.apache.flink.connector.kafka.source.enumerator.KafkaSourceEnumerator$PartitionOffsetsRetrieverImpl.offsetsForTimes(KafkaSourceEnumerator.java:615)
~[flink-sql-connector-kafka-1.15.1-vip.jar:1.15.1-vip]
at
org.apache.flink.connector.kafka.source.enumerator.initializer.TimestampOffsetsInitializer.getPartitionOffsets(TimestampOffsetsInitializer.java:57)
~[flink-sql-connector-kafka-1.15.1-vip.jar:1.15.1-vip]
at
org.apache.flink.connector.kafka.source.enumerator.KafkaSourceEnumerator.initializePartitionSplits(KafkaSourceEnumerator.java:272)
~[flink-sql-connector-kafka-1.15.1-vip.jar:1.15.1-vip]
at
org.apache.flink.connector.kafka.source.enumerator.KafkaSourceEnumerator.lambda$checkPartitionChanges$0(KafkaSourceEnumerator.java:242)
~[flink-sql-connector-kafka-1.15.1-vip.jar:1.15.1-vip]
at
org.apache.flink.runtime.source.coordinator.ExecutorNotifier.lambda$notifyReadyAsync$2(ExecutorNotifier.java:80)
~[flink-dist-1.15.1-vip.jar:1.15.1-vip]
... 7 more {code}
--
This message was sent by Atlassian Jira
(v8.20.10#820010)