baaaaast opened a new issue, #16587:
URL: https://github.com/apache/druid/issues/16587

   ### Affected Version
   29.0.1
   
   ### Description
   #### KIS configuration :
   - replica > 1
   - lateMessageRejectionPeriod of several days
   
   ```
   {
       "type": "kafka",
       "dataSchema":{
           "dataSource":"my-datasource",
           "parser":{
               "type":"avro_stream",
               "avroBytesDecoder" : {
                   "type" : "schema_registry",
                   "url" : "https://schema-registry.url";
               },
               "parseSpec": {
                   "format":"timeAndDims",
                   "timestampSpec": {
                       "column": "timestamp"
                   },
                   "dimensionsSpec":{
                       "dimensions":["my-dimension1","my-dimension2",...],
                       "spatialDimensions":[]
                   }
               }
           },
           "metricsSpec":[
             {
                 "type" : "longSum",
                 "name" : "my-name",
                 "fieldName": "my-name"
             },...
           ],
           "granularitySpec":{
               "type":"uniform",
               "segmentGranularity":"HOUR",
               "queryGranularity":"HOUR"
           }
       },
       "tuningConfig" : {
           "type" : "kafka",
           "maxRowsInMemory": 150000,
           "maxBytesInMemory" : 0,
           "maxRowsPerSegment": 5000000,
           "workerThreads": 5,
           "intermediatePersistPeriod" : "PT10M",
           "maxPendingPersists" : 0,
           "forceExtendableShardSpecs": true,
           "indexSpec" : {
               "bitmap" : {
                   "type" : "roaring"
               },
               "dimensionCompression" : "lz4",
               "metricCompression" : "lz4",
               "longEncoding" : "longs"
           },
           "resetOffsetAutomatically" : false
       },
       "ioConfig": {
           "type" : "kafka",
           "topic": "my-topic",
           "taskCount": 2,
           "replicas": 2,
           "taskDuration": "PT1H",
           "useEarliestOffset": false,
           "lateMessageRejectionPeriod": "P8D",
           "earlyMessageRejectionPeriod": "PT3H",
           "consumerProperties": {
               "bootstrap.servers": "broker-1:9092,broker-2:9092,broker-3:9092"
           }
       }
   }
   ```
   
   #### Particularity of the topic :
   - consumed topic can have several million messages per minute
   - messages with a timestamp to be dated several days in the past
   
   => KIS writes many segments in the past every hour.
   
   #### Problem
   After a while (after several intermediate persist periods), the overlord 
starts generating errors and returns a status 500 to the middlemanager. This 
causes the indexing task to fail.
   I can't reproduce the problem if the kafka topic has a low rate of messages 
produced.
   
   #### Overlord error
   
   ```
   2024-06-12T08:46:42,877 WARN [qtp1884568750-98] 
org.apache.druid.indexing.overlord.http.OverlordResource - Failed to perform 
task action
   java.lang.RuntimeException: java.util.concurrent.ExecutionException: 
org.skife.jdbi.v2.exceptions.CallbackFailedException: 
org.skife.jdbi.v2.exceptions.UnableToExecuteStatementException: 
java.sql.BatchUpdateException: Batch entry 1 INSERT INTO druid_pendingSegments 
(id, dataSource, created_date, start, "end", sequence_name, sequence_prev_id, 
sequence_name_prev_id_sha1, payload) VALUES 
(('my-datasource_2024-06-12T01:00:00.000Z_2024-06-12T02:00:00.000Z_2024-06-12T01:00:00.849Z_5'),
 ('my-datasource'), ('2024-06-12T08:46:42.875Z'), ('2024-06-12T01:00:00.000Z'), 
('2024-06-12T02:00:00.000Z'), ('index_kafka_my-datasource_0f3e4661a8cd3ee_0'), 
('my-datasource_2024-05-30T00:00:00.000Z_2024-05-31T00:00:00.000Z_2024-06-12T01:04:25.132Z_6'),
 ('A10883453F9AA74CDD231465CFF98E73DE3A2864'), ?) was aborted: ERROR: duplicate 
key value violates unique constraint 
"druid_pendingsegments_sequence_name_prev_id_sha1_key"
     Detail: Key 
(sequence_name_prev_id_sha1)=(A10883453F9AA74CDD231465CFF98E73DE3A2864) already 
exists.  Call getNextException to see other errors in the batch. 
[statement:"INSERT INTO druid_pendingSegments (id, dataSource, created_date, 
start, \"end\", sequence_name, sequence_prev_id, sequence_name_prev_id_sha1, 
payload) VALUES (:id, :dataSource, :created_date, :start, :end, :sequence_name, 
:sequence_prev_id, :sequence_name_prev_id_sha1, :payload)", located:"null", 
rewritten:"null", arguments:{ positional:{}, 
named:{sequence_prev_id:'my-datasource_2024-05-30T00:00:00.000Z_2024-05-31T00:00:00.000Z_2024-06-12T01:04:25.132Z_6',payload:[*,
 
*...],start:'2024-06-12T01:00:00.000Z',end:'2024-06-12T02:00:00.000Z',id:'my-datasource_2024-06-12T01:00:00.000Z_2024-06-12T02:00:00.000Z_2024-06-12T01:00:00.849Z_5',created_date:'2024-06-12T08:46:42.875Z',dataSource:'my-datasource',sequence_name:'index_kafka_my-datasource_0f3e4661a8cd3ee_0',sequence_name_prev_id_sha1:'A10883453F9AA74CDD231465CFF98E7
 3DE3A2864'}, finder:[]}]
        at 
org.apache.druid.indexing.common.actions.LocalTaskActionClient.performAction(LocalTaskActionClient.java:98)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.common.actions.LocalTaskActionClient.submit(LocalTaskActionClient.java:80)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.overlord.http.OverlordResource$3.apply(OverlordResource.java:627)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.overlord.http.OverlordResource$3.apply(OverlordResource.java:616)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.overlord.http.OverlordResource.asLeaderWith(OverlordResource.java:1108)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.overlord.http.OverlordResource.doAction(OverlordResource.java:613)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at jdk.internal.reflect.GeneratedMethodAccessor246.invoke(Unknown 
Source) ~[?:?]
        at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(Unknown 
Source) ~[?:?]
        at java.lang.reflect.Method.invoke(Unknown Source) ~[?:?]
        at 
com.sun.jersey.spi.container.JavaMethodInvokerFactory$1.invoke(JavaMethodInvokerFactory.java:60)
 ~[jersey-server-1.19.4.jar:1.19.4]
        at 
com.sun.jersey.server.impl.model.method.dispatch.AbstractResourceMethodDispatchProvider$ResponseOutInvoker._dispatch(AbstractResourceMethodDispatchProvider.java:205)
 ~[jersey-server-1.19.4.jar:1.19.4]
        at 
com.sun.jersey.server.impl.model.method.dispatch.ResourceJavaMethodDispatcher.dispatch(ResourceJavaMethodDispatcher.java:75)
 ~[jersey-server-1.19.4.jar:1.19.4]
        at 
com.sun.jersey.server.impl.uri.rules.HttpMethodRule.accept(HttpMethodRule.java:302)
 ~[jersey-server-1.19.4.jar:1.19.4]
        at 
com.sun.jersey.server.impl.uri.rules.RightHandPathRule.accept(RightHandPathRule.java:147)
 ~[jersey-server-1.19.4.jar:1.19.4]
        at 
com.sun.jersey.server.impl.uri.rules.ResourceClassRule.accept(ResourceClassRule.java:108)
 ~[jersey-server-1.19.4.jar:1.19.4]
        at 
com.sun.jersey.server.impl.uri.rules.RightHandPathRule.accept(RightHandPathRule.java:147)
 ~[jersey-server-1.19.4.jar:1.19.4]
        at 
com.sun.jersey.server.impl.uri.rules.RootResourceClassesRule.accept(RootResourceClassesRule.java:84)
 ~[jersey-server-1.19.4.jar:1.19.4]
        at 
com.sun.jersey.server.impl.application.WebApplicationImpl._handleRequest(WebApplicationImpl.java:1542)
 ~[jersey-server-1.19.4.jar:1.19.4]
        at 
com.sun.jersey.server.impl.application.WebApplicationImpl._handleRequest(WebApplicationImpl.java:1473)
 ~[jersey-server-1.19.4.jar:1.19.4]
        at 
com.sun.jersey.server.impl.application.WebApplicationImpl.handleRequest(WebApplicationImpl.java:1419)
 ~[jersey-server-1.19.4.jar:1.19.4]
        at 
com.sun.jersey.server.impl.application.WebApplicationImpl.handleRequest(WebApplicationImpl.java:1409)
 ~[jersey-server-1.19.4.jar:1.19.4]
        at 
com.sun.jersey.spi.container.servlet.WebComponent.service(WebComponent.java:409)
 ~[jersey-servlet-1.19.4.jar:1.19.4]
        at 
com.sun.jersey.spi.container.servlet.ServletContainer.service(ServletContainer.java:558)
 ~[jersey-servlet-1.19.4.jar:1.19.4]
        at 
com.sun.jersey.spi.container.servlet.ServletContainer.service(ServletContainer.java:733)
 ~[jersey-servlet-1.19.4.jar:1.19.4]
        at javax.servlet.http.HttpServlet.service(HttpServlet.java:790) 
~[javax.servlet-api-3.1.0.jar:3.1.0]
        at 
com.google.inject.servlet.ServletDefinition.doServiceImpl(ServletDefinition.java:286)
 ~[guice-servlet-4.1.0.jar:?]
        at 
com.google.inject.servlet.ServletDefinition.doService(ServletDefinition.java:276)
 ~[guice-servlet-4.1.0.jar:?]
        at 
com.google.inject.servlet.ServletDefinition.service(ServletDefinition.java:181) 
~[guice-servlet-4.1.0.jar:?]
        at 
com.google.inject.servlet.ManagedServletPipeline.service(ManagedServletPipeline.java:91)
 ~[guice-servlet-4.1.0.jar:?]
        at 
com.google.inject.servlet.FilterChainInvocation.doFilter(FilterChainInvocation.java:85)
 ~[guice-servlet-4.1.0.jar:?]
        at 
com.google.inject.servlet.ManagedFilterPipeline.dispatch(ManagedFilterPipeline.java:120)
 ~[guice-servlet-4.1.0.jar:?]
        at com.google.inject.servlet.GuiceFilter.doFilter(GuiceFilter.java:135) 
~[guice-servlet-4.1.0.jar:?]
        at 
org.eclipse.jetty.servlet.FilterHolder.doFilter(FilterHolder.java:193) 
~[jetty-servlet-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.servlet.ServletHandler$Chain.doFilter(ServletHandler.java:1626)
 ~[jetty-servlet-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.apache.druid.server.http.RedirectFilter.doFilter(RedirectFilter.java:73) 
~[druid-server-29.0.1.jar:29.0.1]
        at 
org.eclipse.jetty.servlet.FilterHolder.doFilter(FilterHolder.java:193) 
~[jetty-servlet-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.servlet.ServletHandler$Chain.doFilter(ServletHandler.java:1626)
 ~[jetty-servlet-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.apache.druid.server.security.PreResponseAuthorizationCheckFilter.doFilter(PreResponseAuthorizationCheckFilter.java:84)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.eclipse.jetty.servlet.FilterHolder.doFilter(FilterHolder.java:193) 
~[jetty-servlet-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.servlet.ServletHandler$Chain.doFilter(ServletHandler.java:1626)
 ~[jetty-servlet-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.apache.druid.server.initialization.jetty.StandardResponseHeaderFilterHolder$StandardResponseHeaderFilter.doFilter(StandardResponseHeaderFilterHolder.java:164)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.eclipse.jetty.servlet.FilterHolder.doFilter(FilterHolder.java:193) 
~[jetty-servlet-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.servlet.ServletHandler$Chain.doFilter(ServletHandler.java:1626)
 ~[jetty-servlet-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.apache.druid.server.security.AllowHttpMethodsResourceFilter.doFilter(AllowHttpMethodsResourceFilter.java:78)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.eclipse.jetty.servlet.FilterHolder.doFilter(FilterHolder.java:193) 
~[jetty-servlet-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.servlet.ServletHandler$Chain.doFilter(ServletHandler.java:1626)
 ~[jetty-servlet-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.apache.druid.server.security.AllowOptionsResourceFilter.doFilter(AllowOptionsResourceFilter.java:74)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.eclipse.jetty.servlet.FilterHolder.doFilter(FilterHolder.java:193) 
~[jetty-servlet-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.servlet.ServletHandler$Chain.doFilter(ServletHandler.java:1626)
 ~[jetty-servlet-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.apache.druid.server.security.AllowAllAuthenticator$1.doFilter(AllowAllAuthenticator.java:84)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.apache.druid.server.security.AuthenticationWrappingFilter.doFilter(AuthenticationWrappingFilter.java:59)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.eclipse.jetty.servlet.FilterHolder.doFilter(FilterHolder.java:193) 
~[jetty-servlet-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.servlet.ServletHandler$Chain.doFilter(ServletHandler.java:1626)
 ~[jetty-servlet-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.apache.druid.server.security.SecuritySanityCheckFilter.doFilter(SecuritySanityCheckFilter.java:77)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.eclipse.jetty.servlet.FilterHolder.doFilter(FilterHolder.java:193) 
~[jetty-servlet-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.servlet.ServletHandler$Chain.doFilter(ServletHandler.java:1626)
 ~[jetty-servlet-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:552) 
~[jetty-servlet-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.server.handler.ScopedHandler.nextHandle(ScopedHandler.java:233)
 ~[jetty-server-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:1624)
 ~[jetty-server-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.server.handler.ScopedHandler.nextHandle(ScopedHandler.java:233)
 ~[jetty-server-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1440)
 ~[jetty-server-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.server.handler.ScopedHandler.nextScope(ScopedHandler.java:188)
 ~[jetty-server-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:505) 
~[jetty-servlet-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:1594)
 ~[jetty-server-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.server.handler.ScopedHandler.nextScope(ScopedHandler.java:186)
 ~[jetty-server-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:1355)
 ~[jetty-server-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:141) 
~[jetty-server-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.server.handler.gzip.GzipHandler.handle(GzipHandler.java:772) 
~[jetty-server-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.server.handler.HandlerList.handle(HandlerList.java:59) 
~[jetty-server-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:127) 
~[jetty-server-9.4.53.v20231009.jar:9.4.53.v20231009]
        at org.eclipse.jetty.server.Server.handle(Server.java:516) 
~[jetty-server-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.server.HttpChannel.lambda$handle$1(HttpChannel.java:487) 
~[jetty-server-9.4.53.v20231009.jar:9.4.53.v20231009]
        at org.eclipse.jetty.server.HttpChannel.dispatch(HttpChannel.java:732) 
~[jetty-server-9.4.53.v20231009.jar:9.4.53.v20231009]
        at org.eclipse.jetty.server.HttpChannel.handle(HttpChannel.java:479) 
~[jetty-server-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.server.HttpConnection.onFillable(HttpConnection.java:277) 
~[jetty-server-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.io.AbstractConnection$ReadCallback.succeeded(AbstractConnection.java:311)
 ~[jetty-io-9.4.53.v20231009.jar:9.4.53.v20231009]
        at org.eclipse.jetty.io.FillInterest.fillable(FillInterest.java:105) 
~[jetty-io-9.4.53.v20231009.jar:9.4.53.v20231009]
        at org.eclipse.jetty.io.ChannelEndPoint$1.run(ChannelEndPoint.java:104) 
~[jetty-io-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.runTask(EatWhatYouKill.java:338)
 ~[jetty-util-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.doProduce(EatWhatYouKill.java:315)
 ~[jetty-util-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.tryProduce(EatWhatYouKill.java:173)
 ~[jetty-util-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.util.thread.strategy.EatWhatYouKill.run(EatWhatYouKill.java:131)
 ~[jetty-util-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.util.thread.ReservedThreadExecutor$ReservedThread.run(ReservedThreadExecutor.java:409)
 ~[jetty-util-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.util.thread.QueuedThreadPool.runJob(QueuedThreadPool.java:883)
 ~[jetty-util-9.4.53.v20231009.jar:9.4.53.v20231009]
        at 
org.eclipse.jetty.util.thread.QueuedThreadPool$Runner.run(QueuedThreadPool.java:1034)
 ~[jetty-util-9.4.53.v20231009.jar:9.4.53.v20231009]
        at java.lang.Thread.run(Unknown Source) ~[?:?]
   Caused by: java.util.concurrent.ExecutionException: 
org.skife.jdbi.v2.exceptions.CallbackFailedException: 
org.skife.jdbi.v2.exceptions.UnableToExecuteStatementException: 
java.sql.BatchUpdateException: Batch entry 1 INSERT INTO druid_pendingSegments 
(id, dataSource, created_date, start, "end", sequence_name, sequence_prev_id, 
sequence_name_prev_id_sha1, payload) VALUES 
(('my-datasource_2024-06-12T01:00:00.000Z_2024-06-12T02:00:00.000Z_2024-06-12T01:00:00.849Z_5'),
 ('my-datasource'), ('2024-06-12T08:46:42.875Z'), ('2024-06-12T01:00:00.000Z'), 
('2024-06-12T02:00:00.000Z'), ('index_kafka_my-datasource_0f3e4661a8cd3ee_0'), 
('my-datasource_2024-05-30T00:00:00.000Z_2024-05-31T00:00:00.000Z_2024-06-12T01:04:25.132Z_6'),
 ('A10883453F9AA74CDD231465CFF98E73DE3A2864'), ?) was aborted: ERROR: duplicate 
key value violates unique constraint 
"druid_pendingsegments_sequence_name_prev_id_sha1_key"
     Detail: Key 
(sequence_name_prev_id_sha1)=(A10883453F9AA74CDD231465CFF98E73DE3A2864) already 
exists.  Call getNextException to see other errors in the batch. 
[statement:"INSERT INTO druid_pendingSegments (id, dataSource, created_date, 
start, \"end\", sequence_name, sequence_prev_id, sequence_name_prev_id_sha1, 
payload) VALUES (:id, :dataSource, :created_date, :start, :end, :sequence_name, 
:sequence_prev_id, :sequence_name_prev_id_sha1, :payload)", located:"null", 
rewritten:"null", arguments:{ positional:{}, 
named:{sequence_prev_id:'my-datasource_2024-05-30T00:00:00.000Z_2024-05-31T00:00:00.000Z_2024-06-12T01:04:25.132Z_6',payload:[*,
 
*...],start:'2024-06-12T01:00:00.000Z',end:'2024-06-12T02:00:00.000Z',id:'my-datasource_2024-06-12T01:00:00.000Z_2024-06-12T02:00:00.000Z_2024-06-12T01:00:00.849Z_5',created_date:'2024-06-12T08:46:42.875Z',dataSource:'my-datasource',sequence_name:'index_kafka_my-datasource_0f3e4661a8cd3ee_0',sequence_name_prev_id_sha1:'A10883453F9AA74CDD231465CFF98E7
 3DE3A2864'}, finder:[]}]
        at java.util.concurrent.CompletableFuture.reportGet(Unknown Source) 
~[?:?]
        at java.util.concurrent.CompletableFuture.get(Unknown Source) ~[?:?]
        at 
org.apache.druid.indexing.common.actions.LocalTaskActionClient.performAction(LocalTaskActionClient.java:90)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        ... 85 more
   Caused by: org.skife.jdbi.v2.exceptions.CallbackFailedException: 
org.skife.jdbi.v2.exceptions.UnableToExecuteStatementException: 
java.sql.BatchUpdateException: Batch entry 1 INSERT INTO druid_pendingSegments 
(id, dataSource, created_date, start, "end", sequence_name, sequence_prev_id, 
sequence_name_prev_id_sha1, payload) VALUES 
(('my-datasource_2024-06-12T01:00:00.000Z_2024-06-12T02:00:00.000Z_2024-06-12T01:00:00.849Z_5'),
 ('my-datasource'), ('2024-06-12T08:46:42.875Z'), ('2024-06-12T01:00:00.000Z'), 
('2024-06-12T02:00:00.000Z'), ('index_kafka_my-datasource_0f3e4661a8cd3ee_0'), 
('my-datasource_2024-05-30T00:00:00.000Z_2024-05-31T00:00:00.000Z_2024-06-12T01:04:25.132Z_6'),
 ('A10883453F9AA74CDD231465CFF98E73DE3A2864'), ?) was aborted: ERROR: duplicate 
key value violates unique constraint 
"druid_pendingsegments_sequence_name_prev_id_sha1_key"
     Detail: Key 
(sequence_name_prev_id_sha1)=(A10883453F9AA74CDD231465CFF98E73DE3A2864) already 
exists.  Call getNextException to see other errors in the batch. 
[statement:"INSERT INTO druid_pendingSegments (id, dataSource, created_date, 
start, \"end\", sequence_name, sequence_prev_id, sequence_name_prev_id_sha1, 
payload) VALUES (:id, :dataSource, :created_date, :start, :end, :sequence_name, 
:sequence_prev_id, :sequence_name_prev_id_sha1, :payload)", located:"null", 
rewritten:"null", arguments:{ positional:{}, 
named:{sequence_prev_id:'my-datasource_2024-05-30T00:00:00.000Z_2024-05-31T00:00:00.000Z_2024-06-12T01:04:25.132Z_6',payload:[*,
 
*...],start:'2024-06-12T01:00:00.000Z',end:'2024-06-12T02:00:00.000Z',id:'my-datasource_2024-06-12T01:00:00.000Z_2024-06-12T02:00:00.000Z_2024-06-12T01:00:00.849Z_5',created_date:'2024-06-12T08:46:42.875Z',dataSource:'my-datasource',sequence_name:'index_kafka_my-datasource_0f3e4661a8cd3ee_0',sequence_name_prev_id_sha1:'A10883453F9AA74CDD231465CFF98E7
 3DE3A2864'}, finder:[]}]
        at org.skife.jdbi.v2.DBI.withHandle(DBI.java:284) 
~[jdbi-2.63.1.jar:2.63.1]
        at 
org.apache.druid.metadata.SQLMetadataConnector.lambda$retryWithHandle$0(SQLMetadataConnector.java:151)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.apache.druid.java.util.common.RetryUtils.retry(RetryUtils.java:129) 
~[druid-processing-29.0.1.jar:29.0.1]
        at 
org.apache.druid.java.util.common.RetryUtils.retry(RetryUtils.java:81) 
~[druid-processing-29.0.1.jar:29.0.1]
        at 
org.apache.druid.java.util.common.RetryUtils.retry(RetryUtils.java:163) 
~[druid-processing-29.0.1.jar:29.0.1]
        at 
org.apache.druid.java.util.common.RetryUtils.retry(RetryUtils.java:153) 
~[druid-processing-29.0.1.jar:29.0.1]
        at 
org.apache.druid.metadata.SQLMetadataConnector.retryWithHandle(SQLMetadataConnector.java:151)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.apache.druid.metadata.SQLMetadataConnector.retryWithHandle(SQLMetadataConnector.java:161)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.apache.druid.metadata.IndexerSQLMetadataStorageCoordinator.allocatePendingSegments(IndexerSQLMetadataStorageCoordinator.java:645)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.overlord.TaskLockbox.allocateSegmentIds(TaskLockbox.java:696)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.overlord.TaskLockbox.allocateSegments(TaskLockbox.java:466)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.common.actions.SegmentAllocationQueue.allocateSegmentsForInterval(SegmentAllocationQueue.java:491)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.common.actions.SegmentAllocationQueue.allocateSegmentsForBatch(SegmentAllocationQueue.java:433)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.common.actions.SegmentAllocationQueue.processBatch(SegmentAllocationQueue.java:351)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.common.actions.SegmentAllocationQueue.processBatchesDue(SegmentAllocationQueue.java:266)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) 
~[?:?]
        at java.util.concurrent.FutureTask.run(Unknown Source) ~[?:?]
        at 
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(Unknown
 Source) ~[?:?]
        at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) 
~[?:?]
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) 
~[?:?]
        ... 1 more
   Caused by: org.skife.jdbi.v2.exceptions.UnableToExecuteStatementException: 
java.sql.BatchUpdateException: Batch entry 1 INSERT INTO druid_pendingSegments 
(id, dataSource, created_date, start, "end", sequence_name, sequence_prev_id, 
sequence_name_prev_id_sha1, payload) VALUES 
(('my-datasource_2024-06-12T01:00:00.000Z_2024-06-12T02:00:00.000Z_2024-06-12T01:00:00.849Z_5'),
 ('my-datasource'), ('2024-06-12T08:46:42.875Z'), ('2024-06-12T01:00:00.000Z'), 
('2024-06-12T02:00:00.000Z'), ('index_kafka_my-datasource_0f3e4661a8cd3ee_0'), 
('my-datasource_2024-05-30T00:00:00.000Z_2024-05-31T00:00:00.000Z_2024-06-12T01:04:25.132Z_6'),
 ('A10883453F9AA74CDD231465CFF98E73DE3A2864'), ?) was aborted: ERROR: duplicate 
key value violates unique constraint 
"druid_pendingsegments_sequence_name_prev_id_sha1_key"
     Detail: Key 
(sequence_name_prev_id_sha1)=(A10883453F9AA74CDD231465CFF98E73DE3A2864) already 
exists.  Call getNextException to see other errors in the batch. 
[statement:"INSERT INTO druid_pendingSegments (id, dataSource, created_date, 
start, \"end\", sequence_name, sequence_prev_id, sequence_name_prev_id_sha1, 
payload) VALUES (:id, :dataSource, :created_date, :start, :end, :sequence_name, 
:sequence_prev_id, :sequence_name_prev_id_sha1, :payload)", located:"null", 
rewritten:"null", arguments:{ positional:{}, 
named:{sequence_prev_id:'my-datasource_2024-05-30T00:00:00.000Z_2024-05-31T00:00:00.000Z_2024-06-12T01:04:25.132Z_6',payload:[*,
 
*...],start:'2024-06-12T01:00:00.000Z',end:'2024-06-12T02:00:00.000Z',id:'my-datasource_2024-06-12T01:00:00.000Z_2024-06-12T02:00:00.000Z_2024-06-12T01:00:00.849Z_5',created_date:'2024-06-12T08:46:42.875Z',dataSource:'my-datasource',sequence_name:'index_kafka_my-datasource_0f3e4661a8cd3ee_0',sequence_name_prev_id_sha1:'A10883453F9AA74CDD231465CFF98E7
 3DE3A2864'}, finder:[]}]
        at org.skife.jdbi.v2.PreparedBatch.execute(PreparedBatch.java:148) 
~[jdbi-2.63.1.jar:2.63.1]
        at 
org.apache.druid.metadata.IndexerSQLMetadataStorageCoordinator.insertPendingSegmentsIntoMetastore(IndexerSQLMetadataStorageCoordinator.java:1394)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.apache.druid.metadata.IndexerSQLMetadataStorageCoordinator.allocatePendingSegments(IndexerSQLMetadataStorageCoordinator.java:976)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.apache.druid.metadata.IndexerSQLMetadataStorageCoordinator.lambda$allocatePendingSegments$9(IndexerSQLMetadataStorageCoordinator.java:646)
 ~[druid-server-29.0.1.jar:29.0.1]
        at org.skife.jdbi.v2.DBI.withHandle(DBI.java:281) 
~[jdbi-2.63.1.jar:2.63.1]
        at 
org.apache.druid.metadata.SQLMetadataConnector.lambda$retryWithHandle$0(SQLMetadataConnector.java:151)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.apache.druid.java.util.common.RetryUtils.retry(RetryUtils.java:129) 
~[druid-processing-29.0.1.jar:29.0.1]
        at 
org.apache.druid.java.util.common.RetryUtils.retry(RetryUtils.java:81) 
~[druid-processing-29.0.1.jar:29.0.1]
        at 
org.apache.druid.java.util.common.RetryUtils.retry(RetryUtils.java:163) 
~[druid-processing-29.0.1.jar:29.0.1]
        at 
org.apache.druid.java.util.common.RetryUtils.retry(RetryUtils.java:153) 
~[druid-processing-29.0.1.jar:29.0.1]
        at 
org.apache.druid.metadata.SQLMetadataConnector.retryWithHandle(SQLMetadataConnector.java:151)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.apache.druid.metadata.SQLMetadataConnector.retryWithHandle(SQLMetadataConnector.java:161)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.apache.druid.metadata.IndexerSQLMetadataStorageCoordinator.allocatePendingSegments(IndexerSQLMetadataStorageCoordinator.java:645)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.overlord.TaskLockbox.allocateSegmentIds(TaskLockbox.java:696)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.overlord.TaskLockbox.allocateSegments(TaskLockbox.java:466)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.common.actions.SegmentAllocationQueue.allocateSegmentsForInterval(SegmentAllocationQueue.java:491)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.common.actions.SegmentAllocationQueue.allocateSegmentsForBatch(SegmentAllocationQueue.java:433)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.common.actions.SegmentAllocationQueue.processBatch(SegmentAllocationQueue.java:351)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.common.actions.SegmentAllocationQueue.processBatchesDue(SegmentAllocationQueue.java:266)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) 
~[?:?]
        at java.util.concurrent.FutureTask.run(Unknown Source) ~[?:?]
        at 
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(Unknown
 Source) ~[?:?]
        at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) 
~[?:?]
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) 
~[?:?]
        ... 1 more
   Caused by: java.sql.BatchUpdateException: Batch entry 1 INSERT INTO 
druid_pendingSegments (id, dataSource, created_date, start, "end", 
sequence_name, sequence_prev_id, sequence_name_prev_id_sha1, payload) VALUES 
(('my-datasource_2024-06-12T01:00:00.000Z_2024-06-12T02:00:00.000Z_2024-06-12T01:00:00.849Z_5'),
 ('my-datasource'), ('2024-06-12T08:46:42.875Z'), ('2024-06-12T01:00:00.000Z'), 
('2024-06-12T02:00:00.000Z'), ('index_kafka_my-datasource_0f3e4661a8cd3ee_0'), 
('my-datasource_2024-05-30T00:00:00.000Z_2024-05-31T00:00:00.000Z_2024-06-12T01:04:25.132Z_6'),
 ('A10883453F9AA74CDD231465CFF98E73DE3A2864'), ?) was aborted: ERROR: duplicate 
key value violates unique constraint 
"druid_pendingsegments_sequence_name_prev_id_sha1_key"
     Detail: Key 
(sequence_name_prev_id_sha1)=(A10883453F9AA74CDD231465CFF98E73DE3A2864) already 
exists.  Call getNextException to see other errors in the batch.
        at 
org.postgresql.jdbc.BatchResultHandler.handleCompletion(BatchResultHandler.java:186)
 ~[postgresql-42.7.2.jar:42.7.2]
        at 
org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:590) 
~[postgresql-42.7.2.jar:42.7.2]
        at 
org.postgresql.jdbc.PgStatement.internalExecuteBatch(PgStatement.java:912) 
~[postgresql-42.7.2.jar:42.7.2]
        at org.postgresql.jdbc.PgStatement.executeBatch(PgStatement.java:936) 
~[postgresql-42.7.2.jar:42.7.2]
        at 
org.postgresql.jdbc.PgPreparedStatement.executeBatch(PgPreparedStatement.java:1733)
 ~[postgresql-42.7.2.jar:42.7.2]
        at 
org.apache.commons.dbcp2.DelegatingStatement.executeBatch(DelegatingStatement.java:345)
 ~[commons-dbcp2-2.0.1.jar:2.0.1]
        at 
org.apache.commons.dbcp2.DelegatingStatement.executeBatch(DelegatingStatement.java:345)
 ~[commons-dbcp2-2.0.1.jar:2.0.1]
        at org.skife.jdbi.v2.PreparedBatch.execute(PreparedBatch.java:138) 
~[jdbi-2.63.1.jar:2.63.1]
        at 
org.apache.druid.metadata.IndexerSQLMetadataStorageCoordinator.insertPendingSegmentsIntoMetastore(IndexerSQLMetadataStorageCoordinator.java:1394)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.apache.druid.metadata.IndexerSQLMetadataStorageCoordinator.allocatePendingSegments(IndexerSQLMetadataStorageCoordinator.java:976)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.apache.druid.metadata.IndexerSQLMetadataStorageCoordinator.lambda$allocatePendingSegments$9(IndexerSQLMetadataStorageCoordinator.java:646)
 ~[druid-server-29.0.1.jar:29.0.1]
        at org.skife.jdbi.v2.DBI.withHandle(DBI.java:281) 
~[jdbi-2.63.1.jar:2.63.1]
        at 
org.apache.druid.metadata.SQLMetadataConnector.lambda$retryWithHandle$0(SQLMetadataConnector.java:151)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.apache.druid.java.util.common.RetryUtils.retry(RetryUtils.java:129) 
~[druid-processing-29.0.1.jar:29.0.1]
        at 
org.apache.druid.java.util.common.RetryUtils.retry(RetryUtils.java:81) 
~[druid-processing-29.0.1.jar:29.0.1]
        at 
org.apache.druid.java.util.common.RetryUtils.retry(RetryUtils.java:163) 
~[druid-processing-29.0.1.jar:29.0.1]
        at 
org.apache.druid.java.util.common.RetryUtils.retry(RetryUtils.java:153) 
~[druid-processing-29.0.1.jar:29.0.1]
        at 
org.apache.druid.metadata.SQLMetadataConnector.retryWithHandle(SQLMetadataConnector.java:151)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.apache.druid.metadata.SQLMetadataConnector.retryWithHandle(SQLMetadataConnector.java:161)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.apache.druid.metadata.IndexerSQLMetadataStorageCoordinator.allocatePendingSegments(IndexerSQLMetadataStorageCoordinator.java:645)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.overlord.TaskLockbox.allocateSegmentIds(TaskLockbox.java:696)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.overlord.TaskLockbox.allocateSegments(TaskLockbox.java:466)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.common.actions.SegmentAllocationQueue.allocateSegmentsForInterval(SegmentAllocationQueue.java:491)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.common.actions.SegmentAllocationQueue.allocateSegmentsForBatch(SegmentAllocationQueue.java:433)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.common.actions.SegmentAllocationQueue.processBatch(SegmentAllocationQueue.java:351)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.common.actions.SegmentAllocationQueue.processBatchesDue(SegmentAllocationQueue.java:266)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) 
~[?:?]
        at java.util.concurrent.FutureTask.run(Unknown Source) ~[?:?]
        at 
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(Unknown
 Source) ~[?:?]
        at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) 
~[?:?]
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) 
~[?:?]
        ... 1 more
   Caused by: org.postgresql.util.PSQLException: ERROR: duplicate key value 
violates unique constraint 
"druid_pendingsegments_sequence_name_prev_id_sha1_key"
     Detail: Key 
(sequence_name_prev_id_sha1)=(A10883453F9AA74CDD231465CFF98E73DE3A2864) already 
exists.
        at 
org.postgresql.core.v3.QueryExecutorImpl.receiveErrorResponse(QueryExecutorImpl.java:2725)
 ~[postgresql-42.7.2.jar:42.7.2]
        at 
org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:2412)
 ~[postgresql-42.7.2.jar:42.7.2]
        at 
org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:579) 
~[postgresql-42.7.2.jar:42.7.2]
        at 
org.postgresql.jdbc.PgStatement.internalExecuteBatch(PgStatement.java:912) 
~[postgresql-42.7.2.jar:42.7.2]
        at org.postgresql.jdbc.PgStatement.executeBatch(PgStatement.java:936) 
~[postgresql-42.7.2.jar:42.7.2]
        at 
org.postgresql.jdbc.PgPreparedStatement.executeBatch(PgPreparedStatement.java:1733)
 ~[postgresql-42.7.2.jar:42.7.2]
        at 
org.apache.commons.dbcp2.DelegatingStatement.executeBatch(DelegatingStatement.java:345)
 ~[commons-dbcp2-2.0.1.jar:2.0.1]
        at 
org.apache.commons.dbcp2.DelegatingStatement.executeBatch(DelegatingStatement.java:345)
 ~[commons-dbcp2-2.0.1.jar:2.0.1]
        at org.skife.jdbi.v2.PreparedBatch.execute(PreparedBatch.java:138) 
~[jdbi-2.63.1.jar:2.63.1]
        at 
org.apache.druid.metadata.IndexerSQLMetadataStorageCoordinator.insertPendingSegmentsIntoMetastore(IndexerSQLMetadataStorageCoordinator.java:1394)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.apache.druid.metadata.IndexerSQLMetadataStorageCoordinator.allocatePendingSegments(IndexerSQLMetadataStorageCoordinator.java:976)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.apache.druid.metadata.IndexerSQLMetadataStorageCoordinator.lambda$allocatePendingSegments$9(IndexerSQLMetadataStorageCoordinator.java:646)
 ~[druid-server-29.0.1.jar:29.0.1]
        at org.skife.jdbi.v2.DBI.withHandle(DBI.java:281) 
~[jdbi-2.63.1.jar:2.63.1]
        at 
org.apache.druid.metadata.SQLMetadataConnector.lambda$retryWithHandle$0(SQLMetadataConnector.java:151)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.apache.druid.java.util.common.RetryUtils.retry(RetryUtils.java:129) 
~[druid-processing-29.0.1.jar:29.0.1]
        at 
org.apache.druid.java.util.common.RetryUtils.retry(RetryUtils.java:81) 
~[druid-processing-29.0.1.jar:29.0.1]
        at 
org.apache.druid.java.util.common.RetryUtils.retry(RetryUtils.java:163) 
~[druid-processing-29.0.1.jar:29.0.1]
        at 
org.apache.druid.java.util.common.RetryUtils.retry(RetryUtils.java:153) 
~[druid-processing-29.0.1.jar:29.0.1]
        at 
org.apache.druid.metadata.SQLMetadataConnector.retryWithHandle(SQLMetadataConnector.java:151)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.apache.druid.metadata.SQLMetadataConnector.retryWithHandle(SQLMetadataConnector.java:161)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.apache.druid.metadata.IndexerSQLMetadataStorageCoordinator.allocatePendingSegments(IndexerSQLMetadataStorageCoordinator.java:645)
 ~[druid-server-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.overlord.TaskLockbox.allocateSegmentIds(TaskLockbox.java:696)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.overlord.TaskLockbox.allocateSegments(TaskLockbox.java:466)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.common.actions.SegmentAllocationQueue.allocateSegmentsForInterval(SegmentAllocationQueue.java:491)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.common.actions.SegmentAllocationQueue.allocateSegmentsForBatch(SegmentAllocationQueue.java:433)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.common.actions.SegmentAllocationQueue.processBatch(SegmentAllocationQueue.java:351)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at 
org.apache.druid.indexing.common.actions.SegmentAllocationQueue.processBatchesDue(SegmentAllocationQueue.java:266)
 ~[druid-indexing-service-29.0.1.jar:29.0.1]
        at java.util.concurrent.Executors$RunnableAdapter.call(Unknown Source) 
~[?:?]
        at java.util.concurrent.FutureTask.run(Unknown Source) ~[?:?]
        at 
java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(Unknown
 Source) ~[?:?]
        at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) 
~[?:?]
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) 
~[?:?]
        ... 1 more
   ```
   
   
   #### Postgresql error
   
   ```
   2024-06-12 08:02:13.731 UTC [424455] druid_u@druid_db LOG:  execute S_7: 
INSERT INTO druid_pendingSegments (id, dataSource, created_date, start, "end", 
sequence_name, sequence_prev_id, sequence_name_prev_id_sha1, payload) VALUES 
($1, $2, $3, $4, $5, $6, $7, $8, $9)
   2024-06-12 08:02:13.731 UTC [424455] druid_u@druid_db DETAIL:  parameters: 
$1 = 
'my-datasource_2024-06-12T07:00:00.000Z_2024-06-12T08:00:00.000Z_2024-06-12T07:03:58.908Z_4',
 $2 = 'my-datasource', $3 = '2024-06-12T08:02:13.731Z', $4 = 
'2024-06-12T07:00:00.000Z', $5 = '2024-06-12T08:00:00.000Z', $6 = 
'index_kafka_my-datasource_658fb00f8020ee2_0', $7 = 
'my-datasource_2024-06-09T03:00:00.000Z_2024-06-09T04:00:00.000Z_2024-06-09T03:00:00.440Z_35',
 $8 = '53051B2F1036FED82C2E66195320F7A1EC8609C1', $9 = '\x***'
   2024-06-12 08:02:13.731 UTC [424455] druid_u@druid_db LOG:  execute S_7: 
INSERT INTO druid_pendingSegments (id, dataSource, created_date, start, "end", 
sequence_name, sequence_prev_id, sequence_name_prev_id_sha1, payload) VALUES 
($1, $2, $3, $4, $5, $6, $7, $8, $9)
   2024-06-12 08:02:13.731 UTC [424455] druid_u@druid_db DETAIL:  parameters: 
$1 = 
'my-datasource_2024-06-12T07:00:00.000Z_2024-06-12T08:00:00.000Z_2024-06-12T07:03:58.908Z_3',
 $2 = 'my-datasource', $3 = '2024-06-12T08:02:13.731Z', $4 = 
'2024-06-12T07:00:00.000Z', $5 = '2024-06-12T08:00:00.000Z', $6 = 
'index_kafka_my-datasource_658fb00f8020ee2_0', $7 = '', $8 = 
'53051B2F1036FED82C2E66195320F7A1EC8609C1', $9 = '\x***'
   2024-06-12 08:02:13.731 UTC [424455] druid_u@druid_db ERROR:  duplicate key 
value violates unique constraint 
"druid_pendingsegments_sequence_name_prev_id_sha1_key"
   ```
   
   => We can see that the overlord tries to perform 2 consecutive INSERTs with 
the same sequence_name_prev_id_sha1. 
   
   
   ### Temporary solution
   If I disable the batching segments allocation option by setting the 
`druid.indexer.tasklock.batchSegmentAllocation` property to `false`, the 
problem disappears.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]


Reply via email to