HunterHunter created FLINK-29032:
------------------------------------

             Summary: Kafka Consume from timestamp catch exception : Caused by: 
java.lang.IllegalArgumentException: Invalid negative offset
                 Key: FLINK-29032
                 URL: https://issues.apache.org/jira/browse/FLINK-29032
             Project: Flink
          Issue Type: Bug
          Components: Connectors / Kafka
    Affects Versions: 1.15.1
            Reporter: HunterHunter


/*+ OPTIONS(
'scan.startup.mode' = 'timestamp',
'scan.startup.timestamp-millis'='1660809660000',
) */;
{code:java}
org.apache.flink.util.FlinkRuntimeException: Failed to initialize partition 
splits due to 
        at 
org.apache.flink.connector.kafka.source.enumerator.KafkaSourceEnumerator.handlePartitionSplitChanges(KafkaSourceEnumerator.java:299)
 ~[flink-sql-connector-kafka-1.15.1-vip.jar:1.15.1-vip]
        at 
org.apache.flink.runtime.source.coordinator.ExecutorNotifier.lambda$null$1(ExecutorNotifier.java:83)
 ~[flink-dist-1.15.1-vip.jar:1.15.1-vip]
        at 
org.apache.flink.util.ThrowableCatchingRunnable.run(ThrowableCatchingRunnable.java:40)
 [flink-dist-1.15.1-vip.jar:1.15.1-vip]
        at 
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) 
[?:1.8.0_201]
        at java.util.concurrent.FutureTask.run(FutureTask.java:266) 
[?:1.8.0_201]
        at 
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
 [?:1.8.0_201]
        at 
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
 [?:1.8.0_201]
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) 
[?:1.8.0_201]
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
[?:1.8.0_201]
        at java.lang.Thread.run(Thread.java:748) [?:1.8.0_201]
Caused by: java.lang.IllegalArgumentException: Invalid negative offset
        at 
org.apache.flink.kafka.shaded.org.apache.kafka.clients.consumer.OffsetAndTimestamp.<init>(OffsetAndTimestamp.java:36)
 ~[flink-sql-connector-kafka-1.15.1-vip.jar:1.15.1-vip]
        at 
org.apache.flink.connector.kafka.source.enumerator.KafkaSourceEnumerator$PartitionOffsetsRetrieverImpl.lambda$offsetsForTimes$8(KafkaSourceEnumerator.java:622)
 ~[flink-sql-connector-kafka-1.15.1-vip.jar:1.15.1-vip]
        at java.util.stream.Collectors.lambda$toMap$58(Collectors.java:1321) 
~[?:1.8.0_201]
        at java.util.stream.ReduceOps$3ReducingSink.accept(ReduceOps.java:169) 
~[?:1.8.0_201]
        at 
java.util.HashMap$EntrySpliterator.forEachRemaining(HashMap.java:1699) 
~[?:1.8.0_201]
        at 
java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481) 
~[?:1.8.0_201]
        at 
java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471) 
~[?:1.8.0_201]
        at 
java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708) 
~[?:1.8.0_201]
        at 
java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234) 
~[?:1.8.0_201]
        at 
java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499) 
~[?:1.8.0_201]
        at 
org.apache.flink.connector.kafka.source.enumerator.KafkaSourceEnumerator$PartitionOffsetsRetrieverImpl.offsetsForTimes(KafkaSourceEnumerator.java:615)
 ~[flink-sql-connector-kafka-1.15.1-vip.jar:1.15.1-vip]
        at 
org.apache.flink.connector.kafka.source.enumerator.initializer.TimestampOffsetsInitializer.getPartitionOffsets(TimestampOffsetsInitializer.java:57)
 ~[flink-sql-connector-kafka-1.15.1-vip.jar:1.15.1-vip]
        at 
org.apache.flink.connector.kafka.source.enumerator.KafkaSourceEnumerator.initializePartitionSplits(KafkaSourceEnumerator.java:272)
 ~[flink-sql-connector-kafka-1.15.1-vip.jar:1.15.1-vip]
        at 
org.apache.flink.connector.kafka.source.enumerator.KafkaSourceEnumerator.lambda$checkPartitionChanges$0(KafkaSourceEnumerator.java:242)
 ~[flink-sql-connector-kafka-1.15.1-vip.jar:1.15.1-vip]
        at 
org.apache.flink.runtime.source.coordinator.ExecutorNotifier.lambda$notifyReadyAsync$2(ExecutorNotifier.java:80)
 ~[flink-dist-1.15.1-vip.jar:1.15.1-vip]
        ... 7 more {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

Reply via email to