thomaskwscott commented on a change in pull request #10760: URL: https://github.com/apache/kafka/pull/10760#discussion_r657007003
########## File path: clients/src/main/java/org/apache/kafka/clients/admin/KafkaAdminClient.java ########## @@ -4298,6 +4296,39 @@ void handleFailure(Throwable throwable) { } } } + + @Override + boolean handleUnsupportedVersionException(UnsupportedVersionException exception) { + if (supportsMaxTimestamp) { + supportsMaxTimestamp = false; + + // fail any unsupported futures and remove partitions from the downgraded retry + List<ListOffsetsTopic> topicsToRemove = new ArrayList<>(); + partitionsToQuery.stream().forEach( + t -> { + List<ListOffsetsPartition> partitionsToRemove = new ArrayList<>(); + t.partitions().stream() + .filter(p -> p.timestamp() == ListOffsetsRequest.MAX_TIMESTAMP) + .forEach( + p -> { + futures.get(new TopicPartition(t.name(), p.partitionIndex())) + .completeExceptionally( + new UnsupportedVersionException( + "Broker " + brokerId + + " does not support MAX_TIMESTAMP offset spec")); + partitionsToRemove.add(p); + + }); + t.partitions().removeAll(partitionsToRemove); + if (t.partitions().isEmpty()) topicsToRemove.add(t); + } + ); + partitionsToQuery.removeAll(topicsToRemove); + + return !partitionsToQuery.isEmpty(); Review comment: good point, I've added a check for this. ########## File path: clients/src/test/java/org/apache/kafka/clients/admin/KafkaAdminClientTest.java ########## @@ -4226,6 +4235,124 @@ public void testListOffsetsNonRetriableErrors() throws Exception { } } + @Test + public void testListOffsetsMaxTimestampUnsupportedSingleOffsetSpec() { + + Node node = new Node(0, "localhost", 8120); + List<Node> nodes = Collections.singletonList(node); + final Cluster cluster = new Cluster( + "mockClusterId", + nodes, + Collections.singleton(new PartitionInfo("foo", 0, node, new Node[]{node}, new Node[]{node})), + Collections.emptySet(), + Collections.emptySet(), + node); + final TopicPartition tp0 = new TopicPartition("foo", 0); + + try (AdminClientUnitTestEnv env = new AdminClientUnitTestEnv(cluster, AdminClientConfig.RETRIES_CONFIG, "2")) { + env.kafkaClient().setNodeApiVersions(NodeApiVersions.create()); + env.kafkaClient().prepareResponse(prepareMetadataResponse(cluster, Errors.NONE)); + + // listoffsets response from broker 0 + env.kafkaClient().prepareUnsupportedVersionResponse( + request -> request instanceof ListOffsetsRequest); + + ListOffsetsResult result = env.adminClient().listOffsets(Collections.singletonMap(tp0, OffsetSpec.maxTimestamp())); + + TestUtils.assertFutureThrows(result.all(), UnsupportedVersionException.class); + } + } + + @Test + public void testListOffsetsMaxTimestampUnsupportedMultipleOffsetSpec() throws Exception { + + Node node = new Node(0, "localhost", 8120); + List<Node> nodes = Collections.singletonList(node); + List<PartitionInfo> pInfos = new ArrayList<>(); + pInfos.add(new PartitionInfo("foo", 0, node, new Node[]{node}, new Node[]{node})); + pInfos.add(new PartitionInfo("foo", 1, node, new Node[]{node}, new Node[]{node})); + final Cluster cluster = new Cluster( + "mockClusterId", + nodes, + pInfos, + Collections.emptySet(), + Collections.emptySet(), + node); + final TopicPartition tp0 = new TopicPartition("foo", 0); + final TopicPartition tp1 = new TopicPartition("foo", 1); + + try (AdminClientUnitTestEnv env = new AdminClientUnitTestEnv(cluster, + AdminClientConfig.RETRIES_CONFIG, "2")) { + + env.kafkaClient().setNodeApiVersions(NodeApiVersions.create()); + env.kafkaClient().prepareResponse(prepareMetadataResponse(cluster, Errors.NONE)); + + // listoffsets response from broker 0 + env.kafkaClient().prepareUnsupportedVersionResponse( + request -> request instanceof ListOffsetsRequest); Review comment: sure thing, updated. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org