This is an automated email from the ASF dual-hosted git repository.

hepin pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/pekko-connectors-kafka.git

commit 5d1fab24b9c83c67c211164c3125aa8c6f469a10
Author: Scala Steward <[email protected]>
AuthorDate: Sun Feb 1 04:31:55 2026 +0000

    Update sbt-java-formatter to 0.10.0
---
 .sbtopts                                           | 10 ++++
 .../src/test/java/docs/javadsl/AssignmentTest.java | 28 +++++-----
 .../java/docs/javadsl/ClusterShardingExample.java  |  1 +
 .../java/docs/javadsl/ConsumerExampleTest.java     | 63 +++++++++++-----------
 .../java/docs/javadsl/ConsumerSettingsTest.java    |  6 +--
 .../test/java/docs/javadsl/MetadataClientTest.java | 35 ++++++------
 .../java/docs/javadsl/ProducerSettingsTest.java    |  4 +-
 .../src/test/java/docs/javadsl/ProducerTest.java   | 31 +++++------
 .../javadsl/SchemaRegistrySerializationTest.java   | 47 ++++++++--------
 .../test/java/docs/javadsl/SerializationTest.java  | 45 ++++++++--------
 .../java/docs/javadsl/proto/OrderMessages.java     | 17 ++++--
 project/plugins.sbt                                |  2 +-
 12 files changed, 155 insertions(+), 134 deletions(-)

diff --git a/.sbtopts b/.sbtopts
new file mode 100644
index 00000000..1ea8ae46
--- /dev/null
+++ b/.sbtopts
@@ -0,0 +1,10 @@
+-J--add-exports=jdk.compiler/com.sun.tools.javac.api=ALL-UNNAMED
+-J--add-exports=jdk.compiler/com.sun.tools.javac.file=ALL-UNNAMED
+-J--add-exports=jdk.compiler/com.sun.tools.javac.main=ALL-UNNAMED
+-J--add-exports=jdk.compiler/com.sun.tools.javac.model=ALL-UNNAMED
+-J--add-exports=jdk.compiler/com.sun.tools.javac.parser=ALL-UNNAMED
+-J--add-exports=jdk.compiler/com.sun.tools.javac.processing=ALL-UNNAMED
+-J--add-exports=jdk.compiler/com.sun.tools.javac.tree=ALL-UNNAMED
+-J--add-exports=jdk.compiler/com.sun.tools.javac.util=ALL-UNNAMED
+-J--add-opens=jdk.compiler/com.sun.tools.javac.code=ALL-UNNAMED
+-J--add-opens=jdk.compiler/com.sun.tools.javac.comp=ALL-UNNAMED
diff --git a/java-tests/src/test/java/docs/javadsl/AssignmentTest.java 
b/java-tests/src/test/java/docs/javadsl/AssignmentTest.java
index 6f4f2c70..00154cf7 100644
--- a/java-tests/src/test/java/docs/javadsl/AssignmentTest.java
+++ b/java-tests/src/test/java/docs/javadsl/AssignmentTest.java
@@ -14,6 +14,17 @@
 
 package docs.javadsl;
 
+import static org.junit.Assert.assertEquals;
+
+import java.util.Arrays;
+import java.util.List;
+import java.util.concurrent.CompletionStage;
+import java.util.stream.Collectors;
+import java.util.stream.IntStream;
+import org.apache.kafka.clients.consumer.ConsumerRecord;
+import org.apache.kafka.clients.producer.ProducerRecord;
+import org.apache.kafka.common.TopicPartition;
+// #testkit
 import org.apache.pekko.Done;
 import org.apache.pekko.actor.ActorSystem;
 import org.apache.pekko.kafka.AutoSubscription;
@@ -22,31 +33,20 @@ import org.apache.pekko.kafka.ProducerMessage;
 import org.apache.pekko.kafka.Subscriptions;
 import org.apache.pekko.kafka.javadsl.Consumer;
 // #testkit
+import org.apache.pekko.kafka.javadsl.Producer;
 import org.apache.pekko.kafka.testkit.TestcontainersKafkaJunit4Test;
 // #testkit
-import org.apache.pekko.kafka.javadsl.Producer;
 import org.apache.pekko.kafka.tests.javadsl.LogCapturingJunit4;
 import org.apache.pekko.stream.javadsl.Sink;
 import org.apache.pekko.stream.javadsl.Source;
 // #testkit
 import org.apache.pekko.testkit.javadsl.TestKit;
 // #testkit
-import org.apache.kafka.clients.consumer.ConsumerRecord;
-import org.apache.kafka.clients.producer.ProducerRecord;
-import org.apache.kafka.common.TopicPartition;
-// #testkit
 import org.junit.AfterClass;
-import org.junit.Test;
-// #testkit
 import org.junit.Rule;
+import org.junit.Test;
 
-import java.util.Arrays;
-import java.util.List;
-import java.util.concurrent.CompletionStage;
-import java.util.stream.Collectors;
-import java.util.stream.IntStream;
-
-import static org.junit.Assert.assertEquals;
+// #testkit
 
 // #testkit
 
diff --git a/java-tests/src/test/java/docs/javadsl/ClusterShardingExample.java 
b/java-tests/src/test/java/docs/javadsl/ClusterShardingExample.java
index 7b4e28c2..e79330f0 100644
--- a/java-tests/src/test/java/docs/javadsl/ClusterShardingExample.java
+++ b/java-tests/src/test/java/docs/javadsl/ClusterShardingExample.java
@@ -49,6 +49,7 @@ public class ClusterShardingExample {
       this.mame = mame;
     }
   }
+
   // #user-entity
 
   public static Behavior<User> userBehaviour() {
diff --git a/java-tests/src/test/java/docs/javadsl/ConsumerExampleTest.java 
b/java-tests/src/test/java/docs/javadsl/ConsumerExampleTest.java
index a5405703..0067070e 100644
--- a/java-tests/src/test/java/docs/javadsl/ConsumerExampleTest.java
+++ b/java-tests/src/test/java/docs/javadsl/ConsumerExampleTest.java
@@ -14,6 +14,30 @@
 
 package docs.javadsl;
 
+import static org.hamcrest.CoreMatchers.*;
+import static org.hamcrest.MatcherAssert.*;
+import static org.junit.Assert.assertEquals;
+
+import com.typesafe.config.Config;
+import java.time.Duration;
+import java.util.List;
+import java.util.Map;
+import java.util.Set;
+import java.util.concurrent.CompletableFuture;
+import java.util.concurrent.CompletionStage;
+import java.util.concurrent.Executor;
+import java.util.concurrent.Executors;
+import java.util.concurrent.atomic.AtomicLong;
+import java.util.concurrent.atomic.AtomicReference;
+import java.util.function.Function;
+import org.apache.kafka.clients.consumer.ConsumerConfig;
+import org.apache.kafka.clients.consumer.ConsumerRecord;
+import org.apache.kafka.clients.producer.ProducerRecord;
+import org.apache.kafka.common.Metric;
+import org.apache.kafka.common.MetricName;
+import org.apache.kafka.common.TopicPartition;
+import org.apache.kafka.common.serialization.ByteArrayDeserializer;
+import org.apache.kafka.common.serialization.StringDeserializer;
 import org.apache.pekko.Done;
 import org.apache.pekko.NotUsed;
 import org.apache.pekko.actor.AbstractLoggingActor;
@@ -22,54 +46,29 @@ import org.apache.pekko.actor.ActorSystem;
 import org.apache.pekko.actor.Props;
 import org.apache.pekko.actor.typed.Behavior;
 import org.apache.pekko.actor.typed.javadsl.ActorContext;
+import org.apache.pekko.actor.typed.javadsl.Adapter;
+// #consumerActorTyped
+// #withTypedRebalanceListenerActor
 import org.apache.pekko.actor.typed.javadsl.Behaviors;
 // #withTypedRebalanceListenerActor
 // #consumerActorTyped
 // adds support for actors to a classic actor system and context
-import org.apache.pekko.actor.typed.javadsl.Adapter;
-// #consumerActorTyped
-// #withTypedRebalanceListenerActor
 import org.apache.pekko.japi.Pair;
 import org.apache.pekko.kafka.*;
 import org.apache.pekko.kafka.javadsl.Committer;
 import org.apache.pekko.kafka.javadsl.Consumer;
-import org.apache.pekko.kafka.javadsl.Producer;
 import org.apache.pekko.kafka.javadsl.PartitionAssignmentHandler;
+import org.apache.pekko.kafka.javadsl.Producer;
 import org.apache.pekko.kafka.testkit.TestcontainersKafkaTest;
 import org.apache.pekko.kafka.tests.javadsl.LogCapturingExtension;
 import org.apache.pekko.stream.RestartSettings;
 import org.apache.pekko.stream.javadsl.*;
 import org.apache.pekko.testkit.javadsl.TestKit;
-import com.typesafe.config.Config;
-import org.apache.kafka.clients.consumer.ConsumerConfig;
-import org.apache.kafka.clients.consumer.ConsumerRecord;
-import org.apache.kafka.clients.producer.ProducerRecord;
-import org.apache.kafka.common.Metric;
-import org.apache.kafka.common.MetricName;
-import org.apache.kafka.common.TopicPartition;
-import org.apache.kafka.common.serialization.ByteArrayDeserializer;
-import org.apache.kafka.common.serialization.StringDeserializer;
 import org.junit.jupiter.api.AfterAll;
 import org.junit.jupiter.api.Test;
 import org.junit.jupiter.api.TestInstance;
 import org.junit.jupiter.api.extension.ExtendWith;
 
-import java.time.Duration;
-import java.util.List;
-import java.util.Map;
-import java.util.Set;
-import java.util.concurrent.CompletableFuture;
-import java.util.concurrent.CompletionStage;
-import java.util.concurrent.Executor;
-import java.util.concurrent.Executors;
-import java.util.concurrent.atomic.AtomicLong;
-import java.util.concurrent.atomic.AtomicReference;
-import java.util.function.Function;
-
-import static org.hamcrest.CoreMatchers.*;
-import static org.hamcrest.MatcherAssert.*;
-import static org.junit.Assert.assertEquals;
-
 @TestInstance(TestInstance.Lifecycle.PER_CLASS)
 @ExtendWith(LogCapturingExtension.class)
 class ConsumerExampleTest extends TestcontainersKafkaTest {
@@ -108,6 +107,7 @@ class ConsumerExampleTest extends TestcontainersKafkaTest {
       consumerSettings
           .withProperty(ConsumerConfig.ENABLE_AUTO_COMMIT_CONFIG, "true")
           .withProperty(ConsumerConfig.AUTO_COMMIT_INTERVAL_MS_CONFIG, "5000");
+
   // #settings-autocommit
 
   @Test
@@ -165,6 +165,7 @@ class ConsumerExampleTest extends TestcontainersKafkaTest {
 
     // #plainSource
   }
+
   // #plainSource
 
   @Test
@@ -650,7 +651,9 @@ class ConsumerExampleTest extends TestcontainersKafkaTest {
     // #consumerMetrics
     sleepMillis(
         100,
-        "to let the control establish itself (fails with 
`java.lang.IllegalStateException: not yet initialized: only setHandler is 
allowed in GraphStageLogic constructor)` otherwise");
+        "to let the control establish itself (fails with 
`java.lang.IllegalStateException: not yet"
+            + " initialized: only setHandler is allowed in GraphStageLogic 
constructor)`"
+            + " otherwise");
     // #consumerMetrics
     CompletionStage<Map<MetricName, Metric>> metrics = control.getMetrics();
     metrics.thenAccept(map -> System.out.println("Metrics: " + map));
diff --git a/java-tests/src/test/java/docs/javadsl/ConsumerSettingsTest.java 
b/java-tests/src/test/java/docs/javadsl/ConsumerSettingsTest.java
index 4781dc4f..92d36bcb 100644
--- a/java-tests/src/test/java/docs/javadsl/ConsumerSettingsTest.java
+++ b/java-tests/src/test/java/docs/javadsl/ConsumerSettingsTest.java
@@ -14,6 +14,9 @@
 
 package docs.javadsl;
 
+import com.typesafe.config.Config;
+import com.typesafe.config.ConfigFactory;
+import org.apache.kafka.common.serialization.StringDeserializer;
 import org.apache.pekko.actor.ActorSystem;
 import org.apache.pekko.kafka.ConsumerSettings;
 import org.apache.pekko.kafka.ConsumerSettingsSpec$;
@@ -21,9 +24,6 @@ import org.apache.pekko.kafka.ConsumerSettingsSpec$;
 import org.apache.pekko.kafka.javadsl.DiscoverySupport;
 // #discovery-settings
 import org.apache.pekko.testkit.javadsl.TestKit;
-import com.typesafe.config.Config;
-import com.typesafe.config.ConfigFactory;
-import org.apache.kafka.common.serialization.StringDeserializer;
 import org.junit.jupiter.api.Test;
 
 public class ConsumerSettingsTest {
diff --git a/java-tests/src/test/java/docs/javadsl/MetadataClientTest.java 
b/java-tests/src/test/java/docs/javadsl/MetadataClientTest.java
index c65106dc..8d6acd18 100644
--- a/java-tests/src/test/java/docs/javadsl/MetadataClientTest.java
+++ b/java-tests/src/test/java/docs/javadsl/MetadataClientTest.java
@@ -15,37 +15,36 @@
 package docs.javadsl;
 
 // #metadataClient
+import static java.util.stream.Collectors.toSet;
+import static org.hamcrest.MatcherAssert.assertThat;
+import static org.hamcrest.Matchers.containsInAnyOrder;
+import static org.hamcrest.core.Is.is;
+
+import java.util.Collections;
+import java.util.List;
+import java.util.Map;
+import java.util.Set;
+import java.util.concurrent.CompletionException;
+import java.util.concurrent.CompletionStage;
+import java.util.concurrent.Executor;
+import java.util.concurrent.Executors;
+import java.util.concurrent.TimeUnit;
+import org.apache.kafka.common.PartitionInfo;
+import org.apache.kafka.common.TopicPartition;
 import org.apache.pekko.actor.ActorSystem;
 import org.apache.pekko.kafka.ConsumerSettings;
 import org.apache.pekko.kafka.javadsl.MetadataClient;
 import org.apache.pekko.kafka.testkit.TestcontainersKafkaJunit4Test;
+import org.apache.pekko.kafka.tests.javadsl.LogCapturingJunit4;
 import org.apache.pekko.testkit.javadsl.TestKit;
 import org.apache.pekko.util.Timeout;
 // #metadataClient
-import org.apache.pekko.kafka.tests.javadsl.LogCapturingJunit4;
-import org.apache.kafka.common.PartitionInfo;
-import org.apache.kafka.common.TopicPartition;
 import org.hamcrest.core.IsInstanceOf;
 import org.junit.AfterClass;
 import org.junit.Rule;
 import org.junit.Test;
 import org.junit.rules.ExpectedException;
 
-import java.util.Collections;
-import java.util.List;
-import java.util.Map;
-import java.util.Set;
-import java.util.concurrent.CompletionException;
-import java.util.concurrent.CompletionStage;
-import java.util.concurrent.Executor;
-import java.util.concurrent.Executors;
-import java.util.concurrent.TimeUnit;
-
-import static java.util.stream.Collectors.toSet;
-import static org.hamcrest.MatcherAssert.assertThat;
-import static org.hamcrest.Matchers.containsInAnyOrder;
-import static org.hamcrest.core.Is.is;
-
 public class MetadataClientTest extends TestcontainersKafkaJunit4Test {
 
   @Rule public final LogCapturingJunit4 logCapturing = new 
LogCapturingJunit4();
diff --git a/java-tests/src/test/java/docs/javadsl/ProducerSettingsTest.java 
b/java-tests/src/test/java/docs/javadsl/ProducerSettingsTest.java
index 38dc76e6..de19ca6a 100644
--- a/java-tests/src/test/java/docs/javadsl/ProducerSettingsTest.java
+++ b/java-tests/src/test/java/docs/javadsl/ProducerSettingsTest.java
@@ -14,6 +14,8 @@
 
 package docs.javadsl;
 
+import com.typesafe.config.Config;
+import com.typesafe.config.ConfigFactory;
 import org.apache.kafka.common.serialization.StringSerializer;
 import org.apache.pekko.actor.ActorSystem;
 import org.apache.pekko.kafka.ProducerSettings;
@@ -22,8 +24,6 @@ import org.apache.pekko.kafka.ProducerSettingsSpec$;
 import org.apache.pekko.kafka.javadsl.DiscoverySupport;
 // #discovery-settings
 import org.apache.pekko.testkit.javadsl.TestKit;
-import com.typesafe.config.Config;
-import com.typesafe.config.ConfigFactory;
 import org.junit.jupiter.api.Test;
 
 public class ProducerSettingsTest {
diff --git a/java-tests/src/test/java/docs/javadsl/ProducerTest.java 
b/java-tests/src/test/java/docs/javadsl/ProducerTest.java
index 930f7b54..012d0021 100644
--- a/java-tests/src/test/java/docs/javadsl/ProducerTest.java
+++ b/java-tests/src/test/java/docs/javadsl/ProducerTest.java
@@ -14,6 +14,19 @@
 
 package docs.javadsl;
 
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertFalse;
+
+import com.typesafe.config.Config;
+import java.util.Arrays;
+import java.util.List;
+import java.util.Map;
+import java.util.concurrent.*;
+import org.apache.kafka.clients.consumer.ConsumerRecord;
+import org.apache.kafka.clients.producer.ProducerRecord;
+import org.apache.kafka.clients.producer.RecordMetadata;
+import org.apache.kafka.common.serialization.StringSerializer;
+// #testkit
 import org.apache.pekko.Done;
 import org.apache.pekko.actor.ActorSystem;
 import org.apache.pekko.kafka.*;
@@ -28,24 +41,11 @@ import org.apache.pekko.stream.javadsl.Source;
 // #testkit
 import org.apache.pekko.testkit.javadsl.TestKit;
 // #testkit
-import com.typesafe.config.Config;
-import org.apache.kafka.clients.consumer.ConsumerRecord;
-import org.apache.kafka.clients.producer.ProducerRecord;
-import org.apache.kafka.clients.producer.RecordMetadata;
-import org.apache.kafka.common.serialization.StringSerializer;
-// #testkit
 import org.junit.jupiter.api.*;
 import org.junit.jupiter.api.TestInstance.Lifecycle;
 import org.junit.jupiter.api.extension.ExtendWith;
-// #testkit
-
-import java.util.Arrays;
-import java.util.List;
-import java.util.Map;
-import java.util.concurrent.*;
 
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertFalse;
+// #testkit
 
 // #testkit
 
@@ -209,7 +209,8 @@ class ProducerTest extends TestcontainersKafkaTest {
                         + res.offset()
                         + ": "
                         + record.value();
-                  } else if (result instanceof 
ProducerMessage.MultiResult<String, String, Integer> res) {
+                  } else if (result
+                      instanceof ProducerMessage.MultiResult<String, String, 
Integer> res) {
                     return res.getParts().stream()
                         .map(
                             part -> {
diff --git 
a/java-tests/src/test/java/docs/javadsl/SchemaRegistrySerializationTest.java 
b/java-tests/src/test/java/docs/javadsl/SchemaRegistrySerializationTest.java
index 31e84508..5111a7da 100644
--- a/java-tests/src/test/java/docs/javadsl/SchemaRegistrySerializationTest.java
+++ b/java-tests/src/test/java/docs/javadsl/SchemaRegistrySerializationTest.java
@@ -14,24 +14,23 @@
 
 package docs.javadsl;
 
-import org.apache.pekko.Done;
-import org.apache.pekko.actor.ActorSystem;
-import org.apache.pekko.kafka.ConsumerSettings;
-import org.apache.pekko.kafka.ProducerSettings;
-import org.apache.pekko.kafka.Subscriptions;
-import org.apache.pekko.kafka.javadsl.Consumer;
-import org.apache.pekko.kafka.javadsl.Producer;
-import org.apache.pekko.kafka.testkit.KafkaTestkitTestcontainersSettings;
-import org.apache.pekko.kafka.testkit.TestcontainersKafkaJunit4Test;
-import org.apache.pekko.stream.javadsl.Sink;
-import org.apache.pekko.stream.javadsl.Source;
-import org.apache.pekko.testkit.javadsl.TestKit;
+import static org.hamcrest.CoreMatchers.is;
+import static org.hamcrest.MatcherAssert.assertThat;
+
 import docs.scaladsl.SampleAvroClass;
 // #imports
 import io.confluent.kafka.serializers.AbstractKafkaSchemaSerDeConfig;
 import io.confluent.kafka.serializers.KafkaAvroDeserializer;
 import io.confluent.kafka.serializers.KafkaAvroSerializer;
 // #imports
+import java.util.Arrays;
+import java.util.HashMap;
+import java.util.List;
+import java.util.Map;
+import java.util.concurrent.CompletionStage;
+import java.util.concurrent.Executor;
+import java.util.concurrent.Executors;
+import java.util.concurrent.TimeUnit;
 import org.apache.kafka.clients.consumer.ConsumerConfig;
 import org.apache.kafka.clients.consumer.ConsumerRecord;
 import org.apache.kafka.clients.producer.ProducerRecord;
@@ -41,21 +40,21 @@ import org.apache.kafka.common.serialization.Serializer;
 import org.apache.kafka.common.serialization.StringDeserializer;
 import org.apache.kafka.common.serialization.StringSerializer;
 // #imports
+import org.apache.pekko.Done;
+import org.apache.pekko.actor.ActorSystem;
+import org.apache.pekko.kafka.ConsumerSettings;
+import org.apache.pekko.kafka.ProducerSettings;
+import org.apache.pekko.kafka.Subscriptions;
+import org.apache.pekko.kafka.javadsl.Consumer;
+import org.apache.pekko.kafka.javadsl.Producer;
+import org.apache.pekko.kafka.testkit.KafkaTestkitTestcontainersSettings;
+import org.apache.pekko.kafka.testkit.TestcontainersKafkaJunit4Test;
+import org.apache.pekko.stream.javadsl.Sink;
+import org.apache.pekko.stream.javadsl.Source;
+import org.apache.pekko.testkit.javadsl.TestKit;
 import org.junit.AfterClass;
 import org.junit.Test;
 
-import java.util.Arrays;
-import java.util.HashMap;
-import java.util.List;
-import java.util.Map;
-import java.util.concurrent.CompletionStage;
-import java.util.concurrent.Executor;
-import java.util.concurrent.Executors;
-import java.util.concurrent.TimeUnit;
-
-import static org.hamcrest.CoreMatchers.is;
-import static org.hamcrest.MatcherAssert.assertThat;
-
 // #schema-registry-settings
 public class SchemaRegistrySerializationTest extends 
TestcontainersKafkaJunit4Test {
 
diff --git a/java-tests/src/test/java/docs/javadsl/SerializationTest.java 
b/java-tests/src/test/java/docs/javadsl/SerializationTest.java
index f1392dce..21983f2f 100644
--- a/java-tests/src/test/java/docs/javadsl/SerializationTest.java
+++ b/java-tests/src/test/java/docs/javadsl/SerializationTest.java
@@ -14,6 +14,28 @@
 
 package docs.javadsl;
 
+import static org.hamcrest.CoreMatchers.is;
+import static org.hamcrest.MatcherAssert.assertThat;
+
+import com.fasterxml.jackson.core.JsonParseException;
+// #jackson-imports
+// #protobuf-imports
+// the Protobuf generated class
+import com.fasterxml.jackson.databind.ObjectMapper;
+import com.fasterxml.jackson.databind.ObjectReader;
+import com.fasterxml.jackson.databind.ObjectWriter;
+import docs.javadsl.proto.OrderMessages;
+import java.nio.charset.StandardCharsets;
+import java.util.Arrays;
+import java.util.List;
+import java.util.concurrent.*;
+import org.apache.kafka.clients.consumer.ConsumerRecord;
+import org.apache.kafka.clients.producer.ProducerRecord;
+import org.apache.kafka.common.serialization.ByteArrayDeserializer;
+import org.apache.kafka.common.serialization.ByteArraySerializer;
+// #protobuf-imports
+import org.apache.kafka.common.serialization.StringDeserializer;
+import org.apache.kafka.common.serialization.StringSerializer;
 import org.apache.pekko.Done;
 import org.apache.pekko.actor.ActorSystem;
 import org.apache.pekko.kafka.ConsumerSettings;
@@ -28,34 +50,11 @@ import org.apache.pekko.stream.javadsl.Sink;
 import org.apache.pekko.stream.javadsl.Source;
 import org.apache.pekko.testkit.javadsl.TestKit;
 // #jackson-imports
-import com.fasterxml.jackson.databind.ObjectMapper;
-import com.fasterxml.jackson.databind.ObjectReader;
-import com.fasterxml.jackson.databind.ObjectWriter;
-import com.fasterxml.jackson.core.JsonParseException;
-// #jackson-imports
-// #protobuf-imports
-// the Protobuf generated class
-import docs.javadsl.proto.OrderMessages;
-import org.apache.kafka.common.serialization.ByteArrayDeserializer;
-import org.apache.kafka.common.serialization.ByteArraySerializer;
-// #protobuf-imports
-import org.apache.kafka.clients.consumer.ConsumerRecord;
-import org.apache.kafka.clients.producer.ProducerRecord;
-import org.apache.kafka.common.serialization.StringDeserializer;
-import org.apache.kafka.common.serialization.StringSerializer;
 import org.junit.jupiter.api.AfterAll;
 import org.junit.jupiter.api.Test;
 import org.junit.jupiter.api.TestInstance;
 import org.junit.jupiter.api.extension.ExtendWith;
 
-import java.nio.charset.StandardCharsets;
-import java.util.Arrays;
-import java.util.List;
-import java.util.concurrent.*;
-
-import static org.hamcrest.CoreMatchers.is;
-import static org.hamcrest.MatcherAssert.assertThat;
-
 @TestInstance(TestInstance.Lifecycle.PER_CLASS)
 @ExtendWith(LogCapturingExtension.class)
 public class SerializationTest extends TestcontainersKafkaTest {
diff --git a/java-tests/src/test/java/docs/javadsl/proto/OrderMessages.java 
b/java-tests/src/test/java/docs/javadsl/proto/OrderMessages.java
index 39e5c520..9d46a6bd 100644
--- a/java-tests/src/test/java/docs/javadsl/proto/OrderMessages.java
+++ b/java-tests/src/test/java/docs/javadsl/proto/OrderMessages.java
@@ -39,6 +39,7 @@ public final class OrderMessages {
      * @return The id.
      */
     java.lang.String getId();
+
     /**
      * <code>string id = 1;</code>
      *
@@ -46,12 +47,14 @@ public final class OrderMessages {
      */
     com.google.protobuf.ByteString getIdBytes();
   }
+
   /** Protobuf type {@code Order} */
   public static final class Order extends 
com.google.protobuf.GeneratedMessageV3
       implements
       // @@protoc_insertion_point(message_implements:Order)
       OrderOrBuilder {
     private static final long serialVersionUID = 0L;
+
     // Use Order.newBuilder() to construct.
     private Order(com.google.protobuf.GeneratedMessageV3.Builder<?> builder) {
       super(builder);
@@ -84,6 +87,7 @@ public final class OrderMessages {
 
     @SuppressWarnings("serial")
     private volatile java.lang.Object id_ = "";
+
     /**
      * <code>string id = 1;</code>
      *
@@ -101,6 +105,7 @@ public final class OrderMessages {
         return s;
       }
     }
+
     /**
      * <code>string id = 1;</code>
      *
@@ -110,8 +115,7 @@ public final class OrderMessages {
     public com.google.protobuf.ByteString getIdBytes() {
       java.lang.Object ref = id_;
       if (ref instanceof java.lang.String s) {
-        com.google.protobuf.ByteString b =
-            com.google.protobuf.ByteString.copyFromUtf8(s);
+        com.google.protobuf.ByteString b = 
com.google.protobuf.ByteString.copyFromUtf8(s);
         id_ = b;
         return b;
       } else {
@@ -277,6 +281,7 @@ public final class OrderMessages {
       Builder builder = new Builder(parent);
       return builder;
     }
+
     /** Protobuf type {@code Order} */
     public static final class Builder
         extends com.google.protobuf.GeneratedMessageV3.Builder<Builder>
@@ -452,6 +457,7 @@ public final class OrderMessages {
       private int bitField0_;
 
       private java.lang.Object id_ = "";
+
       /**
        * <code>string id = 1;</code>
        *
@@ -468,6 +474,7 @@ public final class OrderMessages {
           return s;
         }
       }
+
       /**
        * <code>string id = 1;</code>
        *
@@ -476,14 +483,14 @@ public final class OrderMessages {
       public com.google.protobuf.ByteString getIdBytes() {
         java.lang.Object ref = id_;
         if (ref instanceof String s) {
-          com.google.protobuf.ByteString b =
-              com.google.protobuf.ByteString.copyFromUtf8(s);
+          com.google.protobuf.ByteString b = 
com.google.protobuf.ByteString.copyFromUtf8(s);
           id_ = b;
           return b;
         } else {
           return (com.google.protobuf.ByteString) ref;
         }
       }
+
       /**
        * <code>string id = 1;</code>
        *
@@ -499,6 +506,7 @@ public final class OrderMessages {
         onChanged();
         return this;
       }
+
       /**
        * <code>string id = 1;</code>
        *
@@ -510,6 +518,7 @@ public final class OrderMessages {
         onChanged();
         return this;
       }
+
       /**
        * <code>string id = 1;</code>
        *
diff --git a/project/plugins.sbt b/project/plugins.sbt
index e681aac6..a066b90e 100644
--- a/project/plugins.sbt
+++ b/project/plugins.sbt
@@ -17,7 +17,7 @@ addSbtPlugin("com.github.sbt" % "sbt-license-report" % 
"1.9.0")
 addSbtPlugin("com.github.sbt" % "sbt-header" % "5.11.0")
 addSbtPlugin("org.scalameta" % "sbt-scalafmt" % "2.5.6")
 addSbtPlugin("com.typesafe" % "sbt-mima-plugin" % "1.1.4")
-addSbtPlugin("com.github.sbt" % "sbt-java-formatter" % "0.9.0")
+addSbtPlugin("com.github.sbt" % "sbt-java-formatter" % "0.10.0")
 // docs
 addSbtPlugin("com.github.sbt" % "sbt-unidoc" % "0.6.1")
 // Java 11 module names are not added 
https://github.com/ThoughtWorksInc/sbt-api-mappings/issues/58


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to