This is an automated email from the ASF dual-hosted git repository.
pabloem pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/beam.git
The following commit(s) were added to refs/heads/master by this push:
new 09bbb48 [BEAM-11936] Fix errorprone UnusedVariable in IO (#16036)
09bbb48 is described below
commit 09bbb48187301f18bec6d9110741c69b955e2b5a
Author: Benjamin Gonzalez <[email protected]>
AuthorDate: Fri Dec 3 14:14:32 2021 -0600
[BEAM-11936] Fix errorprone UnusedVariable in IO (#16036)
* [BEAM-11936] Fix errorprone UnusedVariable in io
* [BEAM-11936] Change todo issue ref, and spotlessApply
* [BEAM-11936] Remove unused code
* [BEAM-11936] Remove suppressUnusedVariable flag
* [BEAM-11936] Remove unsued code after merge
---
sdks/java/io/amazon-web-services/build.gradle | 1 -
sdks/java/io/amazon-web-services2/build.gradle | 1 -
.../apache/beam/sdk/io/aws2/kinesis/KinesisIO.java | 6 -----
.../beam/sdk/io/aws2/s3/S3FileSystemTest.java | 2 +-
sdks/java/io/amqp/build.gradle | 2 +-
.../java/org/apache/beam/sdk/io/amqp/AmqpIO.java | 8 ++----
.../beam/sdk/io/amqp/AmqpMessageCoderTest.java | 2 +-
sdks/java/io/azure/build.gradle | 1 -
.../beam/sdk/io/azure/options/AzureModule.java | 6 -----
sdks/java/io/bigquery-io-perf-tests/build.gradle | 2 +-
.../beam/sdk/bigqueryioperftests/BigQueryIOIT.java | 4 ---
sdks/java/io/cassandra/build.gradle | 1 -
.../apache/beam/sdk/io/cassandra/CassandraIO.java | 1 -
.../beam/sdk/io/cassandra/CassandraIOTest.java | 1 -
sdks/java/io/clickhouse/build.gradle | 1 -
sdks/java/io/common/build.gradle | 2 +-
sdks/java/io/contextualtextio/build.gradle | 1 -
.../contextualtextio/ContextualTextIOSource.java | 4 ---
sdks/java/io/debezium/build.gradle | 1 -
.../io/debezium/expansion-service/build.gradle | 1 -
.../io/debezium/DebeziumTransformRegistrar.java | 3 ---
.../apache/beam/io/debezium/SourceRecordJson.java | 4 +--
.../apache/beam/io/debezium/DebeziumIOTest.java | 4 +--
.../io/debezium/KafkaSourceConsumerFnTest.java | 23 ++++++++---------
.../elasticsearch-tests-5/build.gradle | 1 -
.../elasticsearch-tests-6/build.gradle | 1 -
.../elasticsearch-tests-7/build.gradle | 1 -
.../elasticsearch-tests-common/build.gradle | 1 -
sdks/java/io/elasticsearch/build.gradle | 2 +-
sdks/java/io/expansion-service/build.gradle | 1 -
sdks/java/io/file-based-io-tests/build.gradle | 2 +-
.../java/org/apache/beam/sdk/io/avro/AvroIOIT.java | 4 ---
.../apache/beam/sdk/io/parquet/ParquetIOIT.java | 4 ---
.../java/org/apache/beam/sdk/io/text/TextIOIT.java | 7 -----
.../apache/beam/sdk/io/tfrecord/TFRecordIOIT.java | 4 ---
.../java/org/apache/beam/sdk/io/xml/XmlIOIT.java | 4 ---
sdks/java/io/google-cloud-platform/build.gradle | 1 -
.../expansion-service/build.gradle | 1 -
.../sdk/io/gcp/bigquery/BigQueryQuerySource.java | 4 ---
.../sdk/io/gcp/bigquery/BigQueryServicesImpl.java | 3 ---
.../gcp/bigquery/BigQueryStorageArrowReader.java | 5 ----
.../sdk/io/gcp/bigquery/BigQueryTableSource.java | 3 ---
.../beam/sdk/io/gcp/bigquery/CreateTables.java | 6 -----
.../beam/sdk/io/gcp/bigquery/StorageApiLoads.java | 3 ---
.../bigquery/StorageApiWritesShardedRecords.java | 5 +++-
.../beam/sdk/io/gcp/bigtable/BigtableIO.java | 1 -
.../apache/beam/sdk/io/gcp/healthcare/FhirIO.java | 1 -
.../io/gcp/healthcare/HttpHealthcareApiClient.java | 2 +-
.../sdk/io/gcp/pubsub/PubsubUnboundedSink.java | 4 ---
.../beam/sdk/io/gcp/pubsub/TestPubsubSignal.java | 9 ++-----
.../sdk/io/gcp/spanner/MutationKeyEncoder.java | 4 +--
.../sdk/io/gcp/testing/FakeDatasetService.java | 1 -
.../io/gcp/bigquery/BigQueryIOStorageReadTest.java | 3 ---
.../io/gcp/bigquery/BigQueryNestedRecordsIT.java | 2 --
.../io/gcp/bigquery/BigQueryServicesImplTest.java | 4 ---
.../sdk/io/gcp/bigquery/BigQueryToTableIT.java | 5 ++--
.../beam/sdk/io/gcp/bigtable/BigtableIOTest.java | 2 --
.../beam/sdk/io/gcp/datastore/DatastoreV1Test.java | 3 ---
.../gcp/firestore/BaseFirestoreV1WriteFnTest.java | 5 +---
.../beam/sdk/io/gcp/healthcare/FhirIOLROIT.java | 4 +--
.../beam/sdk/io/gcp/healthcare/FhirIOTestUtil.java | 3 ---
.../beam/sdk/io/gcp/pubsub/PubsubReadIT.java | 3 ---
.../sdk/io/gcp/spanner/ReadSpannerSchemaTest.java | 2 --
.../beam/sdk/io/gcp/spanner/SpannerIOReadTest.java | 30 ++++++++++------------
.../sdk/io/gcp/spanner/SpannerIOWriteTest.java | 1 -
sdks/java/io/hadoop-common/build.gradle | 2 +-
sdks/java/io/hadoop-file-system/build.gradle | 1 -
sdks/java/io/hadoop-format/build.gradle | 1 -
.../sdk/io/hadoop/format/HadoopFormatIOIT.java | 4 ---
.../io/hadoop/format/HadoopFormatIOReadTest.java | 3 ---
sdks/java/io/hbase/build.gradle | 2 +-
.../java/org/apache/beam/sdk/io/hbase/HBaseIO.java | 2 --
sdks/java/io/hcatalog/build.gradle | 2 +-
sdks/java/io/influxdb/build.gradle | 3 +--
.../apache/beam/sdk/io/influxdb/InfluxDbIO.java | 3 ---
sdks/java/io/jdbc/build.gradle | 1 -
.../java/org/apache/beam/sdk/io/jdbc/JdbcIOIT.java | 4 ---
.../org/apache/beam/sdk/io/jdbc/JdbcIOTest.java | 26 ++++++++-----------
sdks/java/io/jms/build.gradle | 1 -
sdks/java/io/kafka/build.gradle | 1 -
.../beam/sdk/io/kafka/KafkaExactlyOnceSink.java | 4 ++-
.../sdk/io/kafka/WatchKafkaTopicPartitionDoFn.java | 5 +++-
.../beam/sdk/io/kafka/KafkaIOExternalTest.java | 3 +--
sdks/java/io/kinesis/build.gradle | 2 +-
.../java/io/kinesis/expansion-service/build.gradle | 1 -
.../beam/sdk/io/kinesis/KinesisIOReadTest.java | 2 --
.../beam/sdk/io/kinesis/serde/AwsModuleTest.java | 2 --
sdks/java/io/kudu/build.gradle | 1 -
.../java/org/apache/beam/sdk/io/kudu/KuduIO.java | 3 ---
.../java/org/apache/beam/sdk/io/kudu/KuduIOIT.java | 3 ---
.../org/apache/beam/sdk/io/kudu/KuduIOTest.java | 1 -
sdks/java/io/mongodb/build.gradle | 2 +-
.../org/apache/beam/sdk/io/mongodb/MongoDbIO.java | 1 -
.../apache/beam/sdk/io/mongodb/MongoDBIOIT.java | 4 ---
sdks/java/io/mqtt/build.gradle | 2 +-
sdks/java/io/parquet/build.gradle | 1 -
.../apache/beam/sdk/io/parquet/ParquetIOTest.java | 1 -
sdks/java/io/rabbitmq/build.gradle | 2 +-
.../beam/sdk/io/rabbitmq/RabbitMqIOTest.java | 3 ---
sdks/java/io/redis/build.gradle | 2 +-
sdks/java/io/snowflake/build.gradle | 1 -
.../io/snowflake/expansion-service/build.gradle | 1 -
.../apache/beam/sdk/io/snowflake/SnowflakeIO.java | 3 +--
.../services/SnowflakeBatchServiceImpl.java | 3 ---
.../services/SnowflakeStreamingServiceImpl.java | 4 +--
.../test/unit/DataSourceConfigurationTest.java | 4 ---
.../test/unit/write/SchemaDispositionTest.java | 1 -
.../test/unit/write/StreamingWriteTest.java | 4 +--
sdks/java/io/solr/build.gradle | 1 -
sdks/java/io/splunk/build.gradle | 2 +-
.../beam/sdk/io/splunk/SplunkEventWriter.java | 3 ++-
sdks/java/io/synthetic/build.gradle | 1 -
sdks/java/io/thrift/build.gradle | 1 -
.../apache/beam/sdk/io/thrift/ThriftIOTest.java | 6 ++---
sdks/java/io/tika/build.gradle | 2 +-
sdks/java/io/xml/build.gradle | 2 +-
116 files changed, 87 insertions(+), 287 deletions(-)
diff --git a/sdks/java/io/amazon-web-services/build.gradle
b/sdks/java/io/amazon-web-services/build.gradle
index 28ce7443..442b2ff 100644
--- a/sdks/java/io/amazon-web-services/build.gradle
+++ b/sdks/java/io/amazon-web-services/build.gradle
@@ -23,7 +23,6 @@ plugins {
}
applyJavaNature(
- suppressUnusedVariable: true,
automaticModuleName: 'org.apache.beam.sdk.io.aws',
)
diff --git a/sdks/java/io/amazon-web-services2/build.gradle
b/sdks/java/io/amazon-web-services2/build.gradle
index 248a49d..fa0b97c 100644
--- a/sdks/java/io/amazon-web-services2/build.gradle
+++ b/sdks/java/io/amazon-web-services2/build.gradle
@@ -20,7 +20,6 @@ import groovy.json.JsonOutput
plugins { id 'org.apache.beam.module' }
applyJavaNature(
- suppressUnusedVariable: true,
automaticModuleName: 'org.apache.beam.sdk.io.aws2',
)
provideIntegrationTestingDependencies()
diff --git
a/sdks/java/io/amazon-web-services2/src/main/java/org/apache/beam/sdk/io/aws2/kinesis/KinesisIO.java
b/sdks/java/io/amazon-web-services2/src/main/java/org/apache/beam/sdk/io/aws2/kinesis/KinesisIO.java
index b636544..6c80310 100644
---
a/sdks/java/io/amazon-web-services2/src/main/java/org/apache/beam/sdk/io/aws2/kinesis/KinesisIO.java
+++
b/sdks/java/io/amazon-web-services2/src/main/java/org/apache/beam/sdk/io/aws2/kinesis/KinesisIO.java
@@ -30,8 +30,6 @@ import org.apache.beam.sdk.values.PCollection;
import org.checkerframework.checker.nullness.qual.Nullable;
import org.joda.time.Duration;
import org.joda.time.Instant;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
import software.amazon.awssdk.auth.credentials.AwsBasicCredentials;
import software.amazon.awssdk.auth.credentials.AwsCredentialsProvider;
import software.amazon.awssdk.auth.credentials.StaticCredentialsProvider;
@@ -237,10 +235,6 @@ import
software.amazon.kinesis.common.InitialPositionInStream;
})
public final class KinesisIO {
- private static final Logger LOG = LoggerFactory.getLogger(KinesisIO.class);
-
- private static final int DEFAULT_NUM_RETRIES = 6;
-
/** Returns a new {@link Read} transform for reading from Kinesis. */
public static Read read() {
return new AutoValue_KinesisIO_Read.Builder()
diff --git
a/sdks/java/io/amazon-web-services2/src/test/java/org/apache/beam/sdk/io/aws2/s3/S3FileSystemTest.java
b/sdks/java/io/amazon-web-services2/src/test/java/org/apache/beam/sdk/io/aws2/s3/S3FileSystemTest.java
index 4149e37..2a7814f 100644
---
a/sdks/java/io/amazon-web-services2/src/test/java/org/apache/beam/sdk/io/aws2/s3/S3FileSystemTest.java
+++
b/sdks/java/io/amazon-web-services2/src/test/java/org/apache/beam/sdk/io/aws2/s3/S3FileSystemTest.java
@@ -181,7 +181,7 @@ public class S3FileSystemTest {
HeadObjectResponse bigHeadObjectResponse =
headObjectResponse.toBuilder().contentLength(5_368_709_120L).build();
assertGetObjectHead(
- s3FileSystem, createObjectHeadRequest(sourcePath, options), options,
headObjectResponse);
+ s3FileSystem, createObjectHeadRequest(sourcePath, options), options,
bigHeadObjectResponse);
try {
s3FileSystem.copy(sourcePath, destinationPath);
diff --git a/sdks/java/io/amqp/build.gradle b/sdks/java/io/amqp/build.gradle
index 0b6ca45..29dd7f6 100644
--- a/sdks/java/io/amqp/build.gradle
+++ b/sdks/java/io/amqp/build.gradle
@@ -17,7 +17,7 @@
*/
plugins { id 'org.apache.beam.module' }
-applyJavaNature(suppressUnusedVariable: true, automaticModuleName:
'org.apache.beam.sdk.io.amqp')
+applyJavaNature( automaticModuleName: 'org.apache.beam.sdk.io.amqp')
description = "Apache Beam :: SDKs :: Java :: IO :: AMQP"
ext.summary = "IO to read and write using AMQP 1.0 protocol
(http://www.amqp.org)."
diff --git
a/sdks/java/io/amqp/src/main/java/org/apache/beam/sdk/io/amqp/AmqpIO.java
b/sdks/java/io/amqp/src/main/java/org/apache/beam/sdk/io/amqp/AmqpIO.java
index c4c5460..6f6effd 100644
--- a/sdks/java/io/amqp/src/main/java/org/apache/beam/sdk/io/amqp/AmqpIO.java
+++ b/sdks/java/io/amqp/src/main/java/org/apache/beam/sdk/io/amqp/AmqpIO.java
@@ -327,19 +327,15 @@ public class AmqpIO {
@Override
public PDone expand(PCollection<Message> input) {
- input.apply(ParDo.of(new WriteFn(this)));
+ input.apply(ParDo.of(new WriteFn()));
return PDone.in(input.getPipeline());
}
private static class WriteFn extends DoFn<Message, Void> {
- private final Write spec;
-
private transient Messenger messenger;
- public WriteFn(Write spec) {
- this.spec = spec;
- }
+ public WriteFn() {}
@Setup
public void setup() throws Exception {
diff --git
a/sdks/java/io/amqp/src/test/java/org/apache/beam/sdk/io/amqp/AmqpMessageCoderTest.java
b/sdks/java/io/amqp/src/test/java/org/apache/beam/sdk/io/amqp/AmqpMessageCoderTest.java
index 6c5c039..548ccce 100644
---
a/sdks/java/io/amqp/src/test/java/org/apache/beam/sdk/io/amqp/AmqpMessageCoderTest.java
+++
b/sdks/java/io/amqp/src/test/java/org/apache/beam/sdk/io/amqp/AmqpMessageCoderTest.java
@@ -63,7 +63,7 @@ public class AmqpMessageCoderTest {
AmqpMessageCoder coder = AmqpMessageCoder.of();
- byte[] encoded = CoderUtils.encodeToByteArray(coder, message);
+ CoderUtils.encodeToByteArray(coder, message);
}
@Test
diff --git a/sdks/java/io/azure/build.gradle b/sdks/java/io/azure/build.gradle
index a2158e7..8f30a48 100644
--- a/sdks/java/io/azure/build.gradle
+++ b/sdks/java/io/azure/build.gradle
@@ -21,7 +21,6 @@ plugins {
}
applyJavaNature(
- suppressUnusedVariable: true,
automaticModuleName: 'org.apache.beam.sdk.io.azure')
description = "Apache Beam :: SDKs :: Java :: IO :: Azure"
diff --git
a/sdks/java/io/azure/src/main/java/org/apache/beam/sdk/io/azure/options/AzureModule.java
b/sdks/java/io/azure/src/main/java/org/apache/beam/sdk/io/azure/options/AzureModule.java
index ea538cf..d199b6a 100644
---
a/sdks/java/io/azure/src/main/java/org/apache/beam/sdk/io/azure/options/AzureModule.java
+++
b/sdks/java/io/azure/src/main/java/org/apache/beam/sdk/io/azure/options/AzureModule.java
@@ -51,7 +51,6 @@ import com.google.auto.service.AutoService;
import java.io.IOException;
import java.lang.reflect.Field;
import java.util.Map;
-import
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.collect.ImmutableSet;
/**
* A Jackson {@link Module} that registers a {@link JsonSerializer} and {@link
JsonDeserializer} for
@@ -60,7 +59,6 @@ import
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.collect.Immutabl
@AutoService(Module.class)
public class AzureModule extends SimpleModule {
- private static final String TYPE_PROPERTY = "@type";
private static final String AZURE_CLIENT_ID = "azureClientId";
private static final String AZURE_TENANT_ID = "azureTenantId";
private static final String AZURE_CLIENT_SECRET = "azureClientSecret";
@@ -154,10 +152,6 @@ public class AzureModule extends SimpleModule {
}
private static class TokenCredentialSerializer extends
JsonSerializer<TokenCredential> {
- // These providers are singletons, so don't require any serialization,
other than type.
- // add any singleton credentials...
- private static final ImmutableSet<Object> SINGLETON_CREDENTIAL_PROVIDERS =
ImmutableSet.of();
-
@Override
public void serialize(
TokenCredential tokenCredential,
diff --git a/sdks/java/io/bigquery-io-perf-tests/build.gradle
b/sdks/java/io/bigquery-io-perf-tests/build.gradle
index 1c3909d..16df776 100644
--- a/sdks/java/io/bigquery-io-perf-tests/build.gradle
+++ b/sdks/java/io/bigquery-io-perf-tests/build.gradle
@@ -17,7 +17,7 @@
*/
plugins { id 'org.apache.beam.module' }
-applyJavaNature(suppressUnusedVariable: true, exportJavadoc: false, publish:
false)
+applyJavaNature( exportJavadoc: false, publish: false)
provideIntegrationTestingDependencies()
enableJavaPerformanceTesting()
diff --git
a/sdks/java/io/bigquery-io-perf-tests/src/test/java/org/apache/beam/sdk/bigqueryioperftests/BigQueryIOIT.java
b/sdks/java/io/bigquery-io-perf-tests/src/test/java/org/apache/beam/sdk/bigqueryioperftests/BigQueryIOIT.java
index ce0ec67..cc5dffe 100644
---
a/sdks/java/io/bigquery-io-perf-tests/src/test/java/org/apache/beam/sdk/bigqueryioperftests/BigQueryIOIT.java
+++
b/sdks/java/io/bigquery-io-perf-tests/src/test/java/org/apache/beam/sdk/bigqueryioperftests/BigQueryIOIT.java
@@ -85,8 +85,6 @@ public class BigQueryIOIT {
private static final String READ_TIME_METRIC_NAME = "read_time";
private static final String WRITE_TIME_METRIC_NAME = "write_time";
private static final String AVRO_WRITE_TIME_METRIC_NAME = "avro_write_time";
- private static String metricsBigQueryTable;
- private static String metricsBigQueryDataset;
private static String testBigQueryDataset;
private static String testBigQueryTable;
private static SyntheticSourceOptions sourceOptions;
@@ -102,8 +100,6 @@ public class BigQueryIOIT {
tempRoot = options.getTempRoot();
sourceOptions =
SyntheticOptions.fromJsonString(options.getSourceOptions(),
SyntheticSourceOptions.class);
- metricsBigQueryDataset = options.getMetricsBigQueryDataset();
- metricsBigQueryTable = options.getMetricsBigQueryTable();
testBigQueryDataset = options.getTestBigQueryDataset();
testBigQueryTable = options.getTestBigQueryTable();
writeFormat = WriteFormat.valueOf(options.getWriteFormat());
diff --git a/sdks/java/io/cassandra/build.gradle
b/sdks/java/io/cassandra/build.gradle
index 0fbd97d..94b584b 100644
--- a/sdks/java/io/cassandra/build.gradle
+++ b/sdks/java/io/cassandra/build.gradle
@@ -20,7 +20,6 @@ plugins { id 'org.apache.beam.module' }
// Do not relocate guava to avoid issues with Cassandra's version.
applyJavaNature(
- suppressUnusedVariable: true,
automaticModuleName: 'org.apache.beam.sdk.io.cassandra',
)
provideIntegrationTestingDependencies()
diff --git
a/sdks/java/io/cassandra/src/main/java/org/apache/beam/sdk/io/cassandra/CassandraIO.java
b/sdks/java/io/cassandra/src/main/java/org/apache/beam/sdk/io/cassandra/CassandraIO.java
index bd919dd..c337834 100644
---
a/sdks/java/io/cassandra/src/main/java/org/apache/beam/sdk/io/cassandra/CassandraIO.java
+++
b/sdks/java/io/cassandra/src/main/java/org/apache/beam/sdk/io/cassandra/CassandraIO.java
@@ -411,7 +411,6 @@ public class CassandraIO {
@Element CassandraIO.Read<T> read, OutputReceiver<Read<T>>
outputReceiver) {
Set<RingRange> ringRanges = getRingRanges(read);
for (RingRange rr : ringRanges) {
- Set<RingRange> subset = ImmutableSet.<RingRange>of(rr);
outputReceiver.output(read.withRingRanges(ImmutableSet.of(rr)));
}
}
diff --git
a/sdks/java/io/cassandra/src/test/java/org/apache/beam/sdk/io/cassandra/CassandraIOTest.java
b/sdks/java/io/cassandra/src/test/java/org/apache/beam/sdk/io/cassandra/CassandraIOTest.java
index eb61c3a..83f28ed 100644
---
a/sdks/java/io/cassandra/src/test/java/org/apache/beam/sdk/io/cassandra/CassandraIOTest.java
+++
b/sdks/java/io/cassandra/src/test/java/org/apache/beam/sdk/io/cassandra/CassandraIOTest.java
@@ -98,7 +98,6 @@ public class CassandraIOTest implements Serializable {
private static final String CASSANDRA_TABLE = "scientist";
private static final Logger LOG =
LoggerFactory.getLogger(CassandraIOTest.class);
private static final String STORAGE_SERVICE_MBEAN =
"org.apache.cassandra.db:type=StorageService";
- private static final float ACCEPTABLE_EMPTY_SPLITS_PERCENTAGE = 0.5f;
private static final int FLUSH_TIMEOUT = 30000;
private static final int JMX_CONF_TIMEOUT = 1000;
private static int jmxPort;
diff --git a/sdks/java/io/clickhouse/build.gradle
b/sdks/java/io/clickhouse/build.gradle
index 1544714..3fd2abf 100644
--- a/sdks/java/io/clickhouse/build.gradle
+++ b/sdks/java/io/clickhouse/build.gradle
@@ -21,7 +21,6 @@ plugins {
id 'ca.coglinc.javacc'
}
applyJavaNature(
- suppressUnusedVariable: true,
automaticModuleName: 'org.apache.beam.sdk.io.clickhouse',
generatedClassPatterns: [
/^org\.apache\.beam\.sdk\.io\.clickhouse\.impl\.parser\..*/,
diff --git a/sdks/java/io/common/build.gradle b/sdks/java/io/common/build.gradle
index 96bfadc..a9d611b 100644
--- a/sdks/java/io/common/build.gradle
+++ b/sdks/java/io/common/build.gradle
@@ -17,7 +17,7 @@
*/
plugins { id 'org.apache.beam.module' }
-applyJavaNature(suppressUnusedVariable: true, exportJavadoc: false,
automaticModuleName: 'org.apache.beam.sdk.io.common')
+applyJavaNature( exportJavadoc: false, automaticModuleName:
'org.apache.beam.sdk.io.common')
description = "Apache Beam :: SDKs :: Java :: IO :: Common"
ext.summary = "Code used by all Beam IOs"
diff --git a/sdks/java/io/contextualtextio/build.gradle
b/sdks/java/io/contextualtextio/build.gradle
index e507522..340e994 100644
--- a/sdks/java/io/contextualtextio/build.gradle
+++ b/sdks/java/io/contextualtextio/build.gradle
@@ -18,7 +18,6 @@
plugins { id 'org.apache.beam.module' }
applyJavaNature(
- suppressUnusedVariable: true,
automaticModuleName: 'org.apache.beam.sdk.io.contextualtextio',
)
diff --git
a/sdks/java/io/contextualtextio/src/main/java/org/apache/beam/sdk/io/contextualtextio/ContextualTextIOSource.java
b/sdks/java/io/contextualtextio/src/main/java/org/apache/beam/sdk/io/contextualtextio/ContextualTextIOSource.java
index 2a95120..82ce3ac 100644
---
a/sdks/java/io/contextualtextio/src/main/java/org/apache/beam/sdk/io/contextualtextio/ContextualTextIOSource.java
+++
b/sdks/java/io/contextualtextio/src/main/java/org/apache/beam/sdk/io/contextualtextio/ContextualTextIOSource.java
@@ -34,8 +34,6 @@ import
org.apache.beam.vendor.grpc.v1p36p0.com.google.protobuf.ByteString;
import
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.annotations.VisibleForTesting;
import
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.base.Preconditions;
import org.checkerframework.checker.nullness.qual.Nullable;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
/**
* Implementation detail of {@link ContextualTextIO.Read}.
@@ -57,8 +55,6 @@ import org.slf4j.LoggerFactory;
class ContextualTextIOSource extends FileBasedSource<Row> {
byte[] delimiter;
- private static final Logger LOG =
LoggerFactory.getLogger(ContextualTextIOSource.class);
-
// Used to Override isSplittable
private boolean hasMultilineCSVRecords;
diff --git a/sdks/java/io/debezium/build.gradle
b/sdks/java/io/debezium/build.gradle
index 9c7fc2b..29964fc 100644
--- a/sdks/java/io/debezium/build.gradle
+++ b/sdks/java/io/debezium/build.gradle
@@ -19,7 +19,6 @@ import groovy.json.JsonOutput
plugins { id 'org.apache.beam.module' }
applyJavaNature(
- suppressUnusedVariable: true,
automaticModuleName: 'org.apache.beam.sdk.io.debezium',
mavenRepositories: [
[id: 'io.confluent', url:
'https://packages.confluent.io/maven/']
diff --git a/sdks/java/io/debezium/expansion-service/build.gradle
b/sdks/java/io/debezium/expansion-service/build.gradle
index 85b8c1f..a183c91 100644
--- a/sdks/java/io/debezium/expansion-service/build.gradle
+++ b/sdks/java/io/debezium/expansion-service/build.gradle
@@ -21,7 +21,6 @@ apply plugin: 'application'
mainClassName = "org.apache.beam.sdk.expansion.service.ExpansionService"
applyJavaNature(
- suppressUnusedVariable: true,
automaticModuleName:
'org.apache.beam.sdk.io.debezium.expansion.service',
exportJavadoc: false,
validateShadowJar: false,
diff --git
a/sdks/java/io/debezium/src/main/java/org/apache/beam/io/debezium/DebeziumTransformRegistrar.java
b/sdks/java/io/debezium/src/main/java/org/apache/beam/io/debezium/DebeziumTransformRegistrar.java
index 0eae9c9..13a099a 100644
---
a/sdks/java/io/debezium/src/main/java/org/apache/beam/io/debezium/DebeziumTransformRegistrar.java
+++
b/sdks/java/io/debezium/src/main/java/org/apache/beam/io/debezium/DebeziumTransformRegistrar.java
@@ -28,8 +28,6 @@ import org.apache.beam.sdk.values.PBegin;
import org.apache.beam.sdk.values.PCollection;
import
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.collect.ImmutableMap;
import org.checkerframework.checker.nullness.qual.Nullable;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
/** Exposes {@link DebeziumIO.Read} as an external transform for
cross-language usage. */
@Experimental(Experimental.Kind.PORTABILITY)
@@ -38,7 +36,6 @@ import org.slf4j.LoggerFactory;
"nullness" // TODO(https://issues.apache.org/jira/browse/BEAM-10402)
})
public class DebeziumTransformRegistrar implements ExternalTransformRegistrar {
- private static final Logger LOG =
LoggerFactory.getLogger(DebeziumTransformRegistrar.class);
public static final String READ_JSON_URN =
"beam:transform:org.apache.beam:debezium_read:v1";
@Override
diff --git
a/sdks/java/io/debezium/src/main/java/org/apache/beam/io/debezium/SourceRecordJson.java
b/sdks/java/io/debezium/src/main/java/org/apache/beam/io/debezium/SourceRecordJson.java
index e61884f..44460a5 100644
---
a/sdks/java/io/debezium/src/main/java/org/apache/beam/io/debezium/SourceRecordJson.java
+++
b/sdks/java/io/debezium/src/main/java/org/apache/beam/io/debezium/SourceRecordJson.java
@@ -61,7 +61,6 @@ import org.checkerframework.checker.nullness.qual.Nullable;
*/
@SuppressWarnings({"nullness"})
public class SourceRecordJson {
- private final @Nullable SourceRecord sourceRecord;
private final @Nullable Struct value;
private final @Nullable Event event;
@@ -75,7 +74,6 @@ public class SourceRecordJson {
throw new IllegalArgumentException();
}
- this.sourceRecord = sourceRecord;
this.value = (Struct) sourceRecord.value();
if (this.value == null) {
@@ -190,6 +188,7 @@ public class SourceRecordJson {
}
/** Depicts a SourceRecord as an Event in order for it to be mapped as JSON.
*/
+ @SuppressWarnings("unused") // Unused params are validated as JSON
static class Event implements Serializable {
private final SourceRecordJson.Event.Metadata metadata;
private final SourceRecordJson.Event.Before before;
@@ -222,6 +221,7 @@ public class SourceRecordJson {
}
/** Depicts the metadata within a SourceRecord. It has valuable fields. */
+ @SuppressWarnings("unused") // Unused params are validated as JSON
static class Metadata implements Serializable {
private final @Nullable String connector;
private final @Nullable String version;
diff --git
a/sdks/java/io/debezium/src/test/java/org/apache/beam/io/debezium/DebeziumIOTest.java
b/sdks/java/io/debezium/src/test/java/org/apache/beam/io/debezium/DebeziumIOTest.java
index ccf57b6..12d1d61 100644
---
a/sdks/java/io/debezium/src/test/java/org/apache/beam/io/debezium/DebeziumIOTest.java
+++
b/sdks/java/io/debezium/src/test/java/org/apache/beam/io/debezium/DebeziumIOTest.java
@@ -30,13 +30,11 @@ import org.apache.kafka.common.config.ConfigValue;
import org.junit.Test;
import org.junit.runner.RunWith;
import org.junit.runners.JUnit4;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
/** Test on the DebeziumIO. */
@RunWith(JUnit4.class)
public class DebeziumIOTest implements Serializable {
- private static final Logger LOG =
LoggerFactory.getLogger(DebeziumIOTest.class);
+
private static final ConnectorConfiguration MYSQL_CONNECTOR_CONFIGURATION =
ConnectorConfiguration.create()
.withUsername("debezium")
diff --git
a/sdks/java/io/debezium/src/test/java/org/apache/beam/io/debezium/KafkaSourceConsumerFnTest.java
b/sdks/java/io/debezium/src/test/java/org/apache/beam/io/debezium/KafkaSourceConsumerFnTest.java
index c22f8a3..85ca71c 100644
---
a/sdks/java/io/debezium/src/test/java/org/apache/beam/io/debezium/KafkaSourceConsumerFnTest.java
+++
b/sdks/java/io/debezium/src/test/java/org/apache/beam/io/debezium/KafkaSourceConsumerFnTest.java
@@ -87,18 +87,17 @@ public class KafkaSourceConsumerFnTest implements
Serializable {
Pipeline pipeline = Pipeline.create();
- PCollection<Integer> counts =
- pipeline
- .apply(
- Create.of(Lists.newArrayList(config))
- .withCoder(MapCoder.of(StringUtf8Coder.of(),
StringUtf8Coder.of())))
- .apply(
- ParDo.of(
- new KafkaSourceConsumerFn<>(
- CounterSourceConnector.class,
- sourceRecord -> (Integer) sourceRecord.value(),
- 1)))
- .setCoder(VarIntCoder.of());
+ pipeline
+ .apply(
+ Create.of(Lists.newArrayList(config))
+ .withCoder(MapCoder.of(StringUtf8Coder.of(),
StringUtf8Coder.of())))
+ .apply(
+ ParDo.of(
+ new KafkaSourceConsumerFn<>(
+ CounterSourceConnector.class,
+ sourceRecord -> (Integer) sourceRecord.value(),
+ 1)))
+ .setCoder(VarIntCoder.of());
pipeline.run().waitUntilFinish();
Assert.assertEquals(3, CounterTask.getCountTasks());
diff --git
a/sdks/java/io/elasticsearch-tests/elasticsearch-tests-5/build.gradle
b/sdks/java/io/elasticsearch-tests/elasticsearch-tests-5/build.gradle
index 01e1750..6bccb6c 100644
--- a/sdks/java/io/elasticsearch-tests/elasticsearch-tests-5/build.gradle
+++ b/sdks/java/io/elasticsearch-tests/elasticsearch-tests-5/build.gradle
@@ -18,7 +18,6 @@
plugins { id 'org.apache.beam.module' }
applyJavaNature(
- suppressUnusedVariable: true,
publish: false,
archivesBaseName: 'beam-sdks-java-io-elasticsearch-tests-5'
)
diff --git
a/sdks/java/io/elasticsearch-tests/elasticsearch-tests-6/build.gradle
b/sdks/java/io/elasticsearch-tests/elasticsearch-tests-6/build.gradle
index 0f7115f..c4a97c4 100644
--- a/sdks/java/io/elasticsearch-tests/elasticsearch-tests-6/build.gradle
+++ b/sdks/java/io/elasticsearch-tests/elasticsearch-tests-6/build.gradle
@@ -18,7 +18,6 @@
plugins { id 'org.apache.beam.module' }
applyJavaNature(
- suppressUnusedVariable: true,
publish: false,
archivesBaseName: 'beam-sdks-java-io-elasticsearch-tests-6'
)
diff --git
a/sdks/java/io/elasticsearch-tests/elasticsearch-tests-7/build.gradle
b/sdks/java/io/elasticsearch-tests/elasticsearch-tests-7/build.gradle
index 513adfc..4d40504 100644
--- a/sdks/java/io/elasticsearch-tests/elasticsearch-tests-7/build.gradle
+++ b/sdks/java/io/elasticsearch-tests/elasticsearch-tests-7/build.gradle
@@ -18,7 +18,6 @@
plugins { id 'org.apache.beam.module' }
applyJavaNature(
- suppressUnusedVariable: true,
publish: false,
archivesBaseName: 'beam-sdks-java-io-elasticsearch-tests-7'
)
diff --git
a/sdks/java/io/elasticsearch-tests/elasticsearch-tests-common/build.gradle
b/sdks/java/io/elasticsearch-tests/elasticsearch-tests-common/build.gradle
index 6718d6c..6f470f2 100644
--- a/sdks/java/io/elasticsearch-tests/elasticsearch-tests-common/build.gradle
+++ b/sdks/java/io/elasticsearch-tests/elasticsearch-tests-common/build.gradle
@@ -18,7 +18,6 @@
plugins { id 'org.apache.beam.module' }
applyJavaNature(
- suppressUnusedVariable: true,
publish: false,
archivesBaseName: 'beam-sdks-java-io-elasticsearch-tests-common'
)
diff --git a/sdks/java/io/elasticsearch/build.gradle
b/sdks/java/io/elasticsearch/build.gradle
index b7cf19f..c3bacdb 100644
--- a/sdks/java/io/elasticsearch/build.gradle
+++ b/sdks/java/io/elasticsearch/build.gradle
@@ -17,7 +17,7 @@
*/
plugins { id 'org.apache.beam.module' }
-applyJavaNature(suppressUnusedVariable: true, automaticModuleName:
'org.apache.beam.sdk.io.elasticsearch')
+applyJavaNature( automaticModuleName: 'org.apache.beam.sdk.io.elasticsearch')
description = "Apache Beam :: SDKs :: Java :: IO :: Elasticsearch"
ext.summary = "IO to read and write on Elasticsearch"
diff --git a/sdks/java/io/expansion-service/build.gradle
b/sdks/java/io/expansion-service/build.gradle
index 529ccef..c9aab1c 100644
--- a/sdks/java/io/expansion-service/build.gradle
+++ b/sdks/java/io/expansion-service/build.gradle
@@ -21,7 +21,6 @@ apply plugin: 'application'
mainClassName = "org.apache.beam.sdk.expansion.service.ExpansionService"
applyJavaNature(
- suppressUnusedVariable: true,
automaticModuleName: 'org.apache.beam.sdk.io.expansion.service',
exportJavadoc: false,
validateShadowJar: false,
diff --git a/sdks/java/io/file-based-io-tests/build.gradle
b/sdks/java/io/file-based-io-tests/build.gradle
index 6b5924f..0c46d80 100644
--- a/sdks/java/io/file-based-io-tests/build.gradle
+++ b/sdks/java/io/file-based-io-tests/build.gradle
@@ -17,7 +17,7 @@
*/
plugins { id 'org.apache.beam.module' }
-applyJavaNature(suppressUnusedVariable: true, exportJavadoc: false, publish:
false)
+applyJavaNature( exportJavadoc: false, publish: false)
provideIntegrationTestingDependencies()
enableJavaPerformanceTesting()
diff --git
a/sdks/java/io/file-based-io-tests/src/test/java/org/apache/beam/sdk/io/avro/AvroIOIT.java
b/sdks/java/io/file-based-io-tests/src/test/java/org/apache/beam/sdk/io/avro/AvroIOIT.java
index 8789e35..f9f6cb8 100644
---
a/sdks/java/io/file-based-io-tests/src/test/java/org/apache/beam/sdk/io/avro/AvroIOIT.java
+++
b/sdks/java/io/file-based-io-tests/src/test/java/org/apache/beam/sdk/io/avro/AvroIOIT.java
@@ -93,8 +93,6 @@ public class AvroIOIT {
+ "}");
private static String filenamePrefix;
- private static String bigQueryDataset;
- private static String bigQueryTable;
private static final String AVRO_NAMESPACE = AvroIOIT.class.getName();
private static Integer numberOfTextLines;
private static Integer datasetSize;
@@ -108,8 +106,6 @@ public class AvroIOIT {
FileBasedIOTestPipelineOptions options =
readFileBasedIOITPipelineOptions();
filenamePrefix = appendTimestampSuffix(options.getFilenamePrefix());
- bigQueryDataset = options.getBigQueryDataset();
- bigQueryTable = options.getBigQueryTable();
datasetSize = options.getDatasetSize();
expectedHash = options.getExpectedHash();
numberOfTextLines = options.getNumberOfRecords();
diff --git
a/sdks/java/io/file-based-io-tests/src/test/java/org/apache/beam/sdk/io/parquet/ParquetIOIT.java
b/sdks/java/io/file-based-io-tests/src/test/java/org/apache/beam/sdk/io/parquet/ParquetIOIT.java
index 600652c..8fb39b4 100644
---
a/sdks/java/io/file-based-io-tests/src/test/java/org/apache/beam/sdk/io/parquet/ParquetIOIT.java
+++
b/sdks/java/io/file-based-io-tests/src/test/java/org/apache/beam/sdk/io/parquet/ParquetIOIT.java
@@ -93,8 +93,6 @@ public class ParquetIOIT {
+ "}");
private static String filenamePrefix;
- private static String bigQueryDataset;
- private static String bigQueryTable;
private static Integer numberOfTextLines;
private static Integer datasetSize;
private static String expectedHash;
@@ -110,8 +108,6 @@ public class ParquetIOIT {
datasetSize = options.getDatasetSize();
expectedHash = options.getExpectedHash();
filenamePrefix = appendTimestampSuffix(options.getFilenamePrefix());
- bigQueryDataset = options.getBigQueryDataset();
- bigQueryTable = options.getBigQueryTable();
settings =
InfluxDBSettings.builder()
.withHost(options.getInfluxHost())
diff --git
a/sdks/java/io/file-based-io-tests/src/test/java/org/apache/beam/sdk/io/text/TextIOIT.java
b/sdks/java/io/file-based-io-tests/src/test/java/org/apache/beam/sdk/io/text/TextIOIT.java
index 8fe65dd..5b3a4f5 100644
---
a/sdks/java/io/file-based-io-tests/src/test/java/org/apache/beam/sdk/io/text/TextIOIT.java
+++
b/sdks/java/io/file-based-io-tests/src/test/java/org/apache/beam/sdk/io/text/TextIOIT.java
@@ -52,8 +52,6 @@ import org.junit.Rule;
import org.junit.Test;
import org.junit.runner.RunWith;
import org.junit.runners.JUnit4;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
/**
* Integration tests for {@link org.apache.beam.sdk.io.TextIO}.
@@ -78,7 +76,6 @@ import org.slf4j.LoggerFactory;
*/
@RunWith(JUnit4.class)
public class TextIOIT {
- private static final Logger LOG = LoggerFactory.getLogger(TextIOIT.class);
private static String filenamePrefix;
private static Integer numberOfTextLines;
@@ -86,8 +83,6 @@ public class TextIOIT {
private static String expectedHash;
private static Compression compressionType;
private static Integer numShards;
- private static String bigQueryDataset;
- private static String bigQueryTable;
private static boolean gatherGcsPerformanceMetrics;
private static InfluxDBSettings settings;
private static final String FILEIOIT_NAMESPACE = TextIOIT.class.getName();
@@ -103,8 +98,6 @@ public class TextIOIT {
compressionType = Compression.valueOf(options.getCompressionType());
filenamePrefix = appendTimestampSuffix(options.getFilenamePrefix());
numShards = options.getNumberOfShards();
- bigQueryDataset = options.getBigQueryDataset();
- bigQueryTable = options.getBigQueryTable();
gatherGcsPerformanceMetrics = options.getReportGcsPerformanceMetrics();
settings =
InfluxDBSettings.builder()
diff --git
a/sdks/java/io/file-based-io-tests/src/test/java/org/apache/beam/sdk/io/tfrecord/TFRecordIOIT.java
b/sdks/java/io/file-based-io-tests/src/test/java/org/apache/beam/sdk/io/tfrecord/TFRecordIOIT.java
index 09cb467..38e4c8c 100644
---
a/sdks/java/io/file-based-io-tests/src/test/java/org/apache/beam/sdk/io/tfrecord/TFRecordIOIT.java
+++
b/sdks/java/io/file-based-io-tests/src/test/java/org/apache/beam/sdk/io/tfrecord/TFRecordIOIT.java
@@ -88,8 +88,6 @@ public class TFRecordIOIT {
private static final String RUN_TIME = "run_time";
private static String filenamePrefix;
- private static String bigQueryDataset;
- private static String bigQueryTable;
private static Integer numberOfTextLines;
private static Integer datasetSize;
private static String expectedHash;
@@ -108,8 +106,6 @@ public class TFRecordIOIT {
numberOfTextLines = options.getNumberOfRecords();
compressionType = Compression.valueOf(options.getCompressionType());
filenamePrefix = appendTimestampSuffix(options.getFilenamePrefix());
- bigQueryDataset = options.getBigQueryDataset();
- bigQueryTable = options.getBigQueryTable();
settings =
InfluxDBSettings.builder()
.withHost(options.getInfluxHost())
diff --git
a/sdks/java/io/file-based-io-tests/src/test/java/org/apache/beam/sdk/io/xml/XmlIOIT.java
b/sdks/java/io/file-based-io-tests/src/test/java/org/apache/beam/sdk/io/xml/XmlIOIT.java
index 7b5f401..bd4d254 100644
---
a/sdks/java/io/file-based-io-tests/src/test/java/org/apache/beam/sdk/io/xml/XmlIOIT.java
+++
b/sdks/java/io/file-based-io-tests/src/test/java/org/apache/beam/sdk/io/xml/XmlIOIT.java
@@ -93,8 +93,6 @@ public class XmlIOIT {
}
private static String filenamePrefix;
- private static String bigQueryDataset;
- private static String bigQueryTable;
private static Integer numberOfTextLines;
private static Integer datasetSize;
private static String expectedHash;
@@ -111,8 +109,6 @@ public class XmlIOIT {
XmlIOITPipelineOptions options =
readIOTestPipelineOptions(XmlIOITPipelineOptions.class);
filenamePrefix = appendTimestampSuffix(options.getFilenamePrefix());
charset = Charset.forName(options.getCharset());
- bigQueryDataset = options.getBigQueryDataset();
- bigQueryTable = options.getBigQueryTable();
datasetSize = options.getDatasetSize();
expectedHash = options.getExpectedHash();
numberOfTextLines = options.getNumberOfRecords();
diff --git a/sdks/java/io/google-cloud-platform/build.gradle
b/sdks/java/io/google-cloud-platform/build.gradle
index 51f5f7f..d16fb16 100644
--- a/sdks/java/io/google-cloud-platform/build.gradle
+++ b/sdks/java/io/google-cloud-platform/build.gradle
@@ -20,7 +20,6 @@ import groovy.json.JsonOutput
plugins { id 'org.apache.beam.module' }
applyJavaNature(
- suppressUnusedVariable: true,
automaticModuleName: 'org.apache.beam.sdk.io.gcp',
enableSpotbugs: false,
classesTriggerCheckerBugs: [
diff --git a/sdks/java/io/google-cloud-platform/expansion-service/build.gradle
b/sdks/java/io/google-cloud-platform/expansion-service/build.gradle
index 0052d42..dea464b 100644
--- a/sdks/java/io/google-cloud-platform/expansion-service/build.gradle
+++ b/sdks/java/io/google-cloud-platform/expansion-service/build.gradle
@@ -21,7 +21,6 @@ apply plugin: 'application'
mainClassName = "org.apache.beam.sdk.expansion.service.ExpansionService"
applyJavaNature(
- suppressUnusedVariable: true,
automaticModuleName: 'org.apache.beam.sdk.io.gcp.expansion.service',
exportJavadoc: false,
validateShadowJar: false,
diff --git
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryQuerySource.java
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryQuerySource.java
index 40559f6..4ca2c64 100644
---
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryQuerySource.java
+++
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryQuerySource.java
@@ -24,15 +24,11 @@ import org.apache.beam.sdk.options.PipelineOptions;
import org.apache.beam.sdk.transforms.SerializableFunction;
import org.apache.beam.sdk.transforms.display.DisplayData;
import
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.annotations.VisibleForTesting;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
/** A {@link BigQuerySourceBase} for querying BigQuery tables. */
@VisibleForTesting
class BigQueryQuerySource<T> extends BigQuerySourceBase<T> {
- private static final Logger LOG =
LoggerFactory.getLogger(BigQueryQuerySource.class);
-
static <T> BigQueryQuerySource<T> create(
String stepUuid,
BigQueryQuerySourceDef queryDef,
diff --git
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryServicesImpl.java
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryServicesImpl.java
index a62b6fc..5a270d4 100644
---
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryServicesImpl.java
+++
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryServicesImpl.java
@@ -145,9 +145,6 @@ import org.slf4j.LoggerFactory;
class BigQueryServicesImpl implements BigQueryServices {
private static final Logger LOG =
LoggerFactory.getLogger(BigQueryServicesImpl.class);
- // How frequently to log while polling.
- private static final Duration POLLING_LOG_GAP = Duration.standardMinutes(10);
-
// The maximum number of retries to execute a BigQuery RPC.
private static final int MAX_RPC_RETRIES = 9;
diff --git
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryStorageArrowReader.java
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryStorageArrowReader.java
index 323b69c..92d23bf 100644
---
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryStorageArrowReader.java
+++
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryStorageArrowReader.java
@@ -34,7 +34,6 @@ import org.apache.beam.sdk.values.Row;
@SuppressWarnings("nullness")
class BigQueryStorageArrowReader implements BigQueryStorageReader {
- private org.apache.beam.sdk.schemas.Schema arrowBeamSchema;
private @Nullable RecordBatchRowIterator recordBatchIterator;
private long rowCount;
private ArrowSchema protoSchema;
@@ -42,10 +41,6 @@ class BigQueryStorageArrowReader implements
BigQueryStorageReader {
BigQueryStorageArrowReader(ReadSession readSession) throws IOException {
protoSchema = readSession.getArrowSchema();
- InputStream input = protoSchema.getSerializedSchema().newInput();
- this.arrowBeamSchema =
- ArrowConversion.ArrowSchemaTranslator.toBeamSchema(
- ArrowConversion.arrowSchemaFromInput(input));
this.rowCount = 0;
this.alloc = null;
}
diff --git
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryTableSource.java
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryTableSource.java
index e2e2bb4..86aa87e 100644
---
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryTableSource.java
+++
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryTableSource.java
@@ -26,8 +26,6 @@ import org.apache.beam.sdk.options.PipelineOptions;
import org.apache.beam.sdk.transforms.SerializableFunction;
import org.apache.beam.sdk.transforms.display.DisplayData;
import
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.annotations.VisibleForTesting;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
/** A {@link BigQuerySourceBase} for reading BigQuery tables. */
@VisibleForTesting
@@ -35,7 +33,6 @@ import org.slf4j.LoggerFactory;
"nullness" // TODO(https://issues.apache.org/jira/browse/BEAM-10402)
})
class BigQueryTableSource<T> extends BigQuerySourceBase<T> {
- private static final Logger LOG =
LoggerFactory.getLogger(BigQueryTableSource.class);
static <T> BigQueryTableSource<T> create(
String stepUuid,
diff --git
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/CreateTables.java
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/CreateTables.java
index d9bb6dd..8eef067 100644
---
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/CreateTables.java
+++
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/CreateTables.java
@@ -20,11 +20,8 @@ package org.apache.beam.sdk.io.gcp.bigquery;
import static
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.base.Preconditions.checkArgument;
import com.google.api.services.bigquery.model.TableSchema;
-import java.util.Collections;
import java.util.List;
import java.util.Map;
-import java.util.Set;
-import java.util.concurrent.ConcurrentHashMap;
import org.apache.beam.sdk.io.gcp.bigquery.BigQueryIO.Write.CreateDisposition;
import org.apache.beam.sdk.transforms.DoFn;
import org.apache.beam.sdk.transforms.PTransform;
@@ -57,9 +54,6 @@ public class CreateTables<DestinationT, ElementT>
*
* <p>TODO: We should put a bound on memory usage of this. Use guava cache
instead.
*/
- private static Set<String> createdTables =
- Collections.newSetFromMap(new ConcurrentHashMap<String, Boolean>());
-
public CreateTables(
CreateDisposition createDisposition,
DynamicDestinations<?, DestinationT> dynamicDestinations) {
diff --git
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/StorageApiLoads.java
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/StorageApiLoads.java
index 2d079e3..8d12556 100644
---
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/StorageApiLoads.java
+++
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/StorageApiLoads.java
@@ -38,13 +38,10 @@ import org.apache.beam.sdk.values.PCollection;
import org.apache.beam.sdk.values.TupleTag;
import org.apache.beam.sdk.values.TypeDescriptor;
import org.joda.time.Duration;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
/** This {@link PTransform} manages loads into BigQuery using the Storage API.
*/
public class StorageApiLoads<DestinationT, ElementT>
extends PTransform<PCollection<KV<DestinationT, ElementT>>, WriteResult> {
- private static final Logger LOG =
LoggerFactory.getLogger(StorageApiLoads.class);
static final int MAX_BATCH_SIZE_BYTES = 2 * 1024 * 1024;
private final Coder<DestinationT> destinationCoder;
diff --git
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/StorageApiWritesShardedRecords.java
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/StorageApiWritesShardedRecords.java
index aacab97..fde0085 100644
---
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/StorageApiWritesShardedRecords.java
+++
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigquery/StorageApiWritesShardedRecords.java
@@ -84,7 +84,10 @@ import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
/** A transform to write sharded records to BigQuery using the Storage API. */
-@SuppressWarnings("FutureReturnValueIgnored")
+@SuppressWarnings({
+ "FutureReturnValueIgnored",
+ "unused" // TODO(BEAM-13271): Remove when new version of errorprone is
released (2.11.0)
+})
public class StorageApiWritesShardedRecords<DestinationT, ElementT>
extends PTransform<
PCollection<KV<ShardedKey<DestinationT>, Iterable<byte[]>>>,
PCollection<Void>> {
diff --git
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigtable/BigtableIO.java
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigtable/BigtableIO.java
index f9df078..8fbadf2 100644
---
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigtable/BigtableIO.java
+++
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/bigtable/BigtableIO.java
@@ -912,7 +912,6 @@ public class BigtableIO {
private final BigtableConfig config;
private final BigtableReadOptions readOptions;
private @Nullable Long estimatedSizeBytes;
- private transient @Nullable List<SampleRowKeysResponse> sampleRowKeys;
/** Creates a new {@link BigtableSource} with just one {@link
ByteKeyRange}. */
protected BigtableSource withSingleRange(ByteKeyRange range) {
diff --git
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/healthcare/FhirIO.java
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/healthcare/FhirIO.java
index 7e7d351..f36d943 100644
---
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/healthcare/FhirIO.java
+++
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/healthcare/FhirIO.java
@@ -924,7 +924,6 @@ public class FhirIO {
@Override
public Result expand(PCollection<String> input) {
- PCollectionTuple bundles;
switch (this.getWriteMethod()) {
case IMPORT:
LOG.warn(
diff --git
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/healthcare/HttpHealthcareApiClient.java
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/healthcare/HttpHealthcareApiClient.java
index 13dfe28..1b09ff4 100644
---
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/healthcare/HttpHealthcareApiClient.java
+++
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/healthcare/HttpHealthcareApiClient.java
@@ -250,7 +250,7 @@ public class HttpHealthcareApiClient implements
HealthcareApiClient, Serializabl
.setEntity(requestEntity)
.addHeader("Content-Type", "application/dicom")
.build();
- HttpResponse response = httpClient.execute(request);
+ httpClient.execute(request);
return new Empty();
}
diff --git
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/pubsub/PubsubUnboundedSink.java
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/pubsub/PubsubUnboundedSink.java
index 038dc6f..fa182d2 100644
---
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/pubsub/PubsubUnboundedSink.java
+++
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/pubsub/PubsubUnboundedSink.java
@@ -24,7 +24,6 @@ import java.io.InputStream;
import java.io.OutputStream;
import java.util.ArrayList;
import java.util.List;
-import java.util.Map;
import java.util.UUID;
import java.util.concurrent.ThreadLocalRandom;
import org.apache.beam.sdk.coders.AtomicCoder;
@@ -33,7 +32,6 @@ import org.apache.beam.sdk.coders.ByteArrayCoder;
import org.apache.beam.sdk.coders.Coder;
import org.apache.beam.sdk.coders.CoderException;
import org.apache.beam.sdk.coders.KvCoder;
-import org.apache.beam.sdk.coders.MapCoder;
import org.apache.beam.sdk.coders.NullableCoder;
import org.apache.beam.sdk.coders.StringUtf8Coder;
import org.apache.beam.sdk.coders.VarIntCoder;
@@ -100,8 +98,6 @@ public class PubsubUnboundedSink extends
PTransform<PCollection<PubsubMessage>,
private static class OutgoingMessageCoder extends
AtomicCoder<OutgoingMessage> {
private static final NullableCoder<String> RECORD_ID_CODER =
NullableCoder.of(StringUtf8Coder.of());
- private static final NullableCoder<Map<String, String>> ATTRIBUTES_CODER =
- NullableCoder.of(MapCoder.of(StringUtf8Coder.of(),
StringUtf8Coder.of()));
@Override
public void encode(OutgoingMessage value, OutputStream outStream)
diff --git
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/pubsub/TestPubsubSignal.java
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/pubsub/TestPubsubSignal.java
index 12c879b..27567a4 100644
---
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/pubsub/TestPubsubSignal.java
+++
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/pubsub/TestPubsubSignal.java
@@ -74,7 +74,8 @@ import org.slf4j.LoggerFactory;
* <p>Uses a random temporary Pubsub topic for synchronization.
*/
@SuppressWarnings({
- "nullness" // TODO(https://issues.apache.org/jira/browse/BEAM-10402)
+ "nullness", // TODO(https://issues.apache.org/jira/browse/BEAM-10402)
+ "unused" // TODO(BEAM-13271): Remove when new version of errorprone is
released (2.11.0)
})
public class TestPubsubSignal implements TestRule {
private static final Logger LOG =
LoggerFactory.getLogger(TestPubsubSignal.class);
@@ -82,10 +83,6 @@ public class TestPubsubSignal implements TestRule {
private static final String RESULT_SUCCESS_MESSAGE = "SUCCESS";
private static final String START_TOPIC_NAME = "start";
private static final String START_SIGNAL_MESSAGE = "START SIGNAL";
- private static final Integer DEFAULT_ACK_DEADLINE_SECONDS = 60;
-
- private static final String NO_ID_ATTRIBUTE = null;
- private static final String NO_TIMESTAMP_ATTRIBUTE = null;
private final TestPubsubOptions pipelineOptions;
private final String pubsubEndpoint;
@@ -404,7 +401,6 @@ public class TestPubsubSignal implements TestRule {
* "FAILURE".
*/
static class StatefulPredicateCheck<T> extends DoFn<KV<String, ? extends T>,
String> {
- private final SerializableFunction<T, String> formatter;
private SerializableFunction<Set<T>, Boolean> successPredicate;
// keep all events seen so far in the state cell
@@ -418,7 +414,6 @@ public class TestPubsubSignal implements TestRule {
SerializableFunction<T, String> formatter,
SerializableFunction<Set<T>, Boolean> successPredicate) {
this.seenEvents = StateSpecs.bag(coder);
- this.formatter = formatter;
this.successPredicate = successPredicate;
}
diff --git
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/spanner/MutationKeyEncoder.java
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/spanner/MutationKeyEncoder.java
index 77d84e9..5e68d55 100644
---
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/spanner/MutationKeyEncoder.java
+++
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/spanner/MutationKeyEncoder.java
@@ -39,15 +39,13 @@ import
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.annotations.Visi
import org.joda.time.DateTime;
import org.joda.time.Days;
import org.joda.time.MutableDateTime;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
/**
* Given the Schema, Encodes the table name and Key into a lexicographically
sortable {@code
* byte[]}.
*/
class MutationKeyEncoder {
- private static final Logger LOG =
LoggerFactory.getLogger(MutationKeyEncoder.class);
+
private static final int ROWS_PER_UNKNOWN_TABLE_LOG_MESSAGE = 10000;
private static final DateTime MIN_DATE = new DateTime(1, 1, 1, 0, 0);
private final SpannerSchema schema;
diff --git
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/testing/FakeDatasetService.java
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/testing/FakeDatasetService.java
index bb3ef9f..239c2ad 100644
---
a/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/testing/FakeDatasetService.java
+++
b/sdks/java/io/google-cloud-platform/src/main/java/org/apache/beam/sdk/io/gcp/testing/FakeDatasetService.java
@@ -188,7 +188,6 @@ public class FakeDatasetService implements DatasetService,
Serializable {
public List<TableRow> getAllRows(String projectId, String datasetId, String
tableId)
throws InterruptedException, IOException {
synchronized (tables) {
- TableContainer tableContainer = getTableContainer(projectId, datasetId,
tableId);
return getTableContainer(projectId, datasetId, tableId).getRows();
}
}
diff --git
a/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryIOStorageReadTest.java
b/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryIOStorageReadTest.java
index 135a536..fcf335c 100644
---
a/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryIOStorageReadTest.java
+++
b/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryIOStorageReadTest.java
@@ -842,7 +842,6 @@ public class BigQueryIOStorageReadTest {
TableRowJsonCoder.of(),
new FakeBigQueryServices().withStorageClient(fakeStorageClient));
- List<TableRow> rows = new ArrayList<>();
BoundedReader<TableRow> reader = streamSource.createReader(options);
// Before call to BoundedReader#start, fraction consumed must be zero.
@@ -928,7 +927,6 @@ public class BigQueryIOStorageReadTest {
TableRowJsonCoder.of(),
new FakeBigQueryServices().withStorageClient(fakeStorageClient));
- List<TableRow> rows = new ArrayList<>();
BoundedReader<TableRow> reader = streamSource.createReader(options);
// Before call to BoundedReader#start, fraction consumed must be zero.
@@ -1618,7 +1616,6 @@ public class BigQueryIOStorageReadTest {
TableRowJsonCoder.of(),
new FakeBigQueryServices().withStorageClient(fakeStorageClient));
- List<TableRow> rows = new ArrayList<>();
BoundedReader<TableRow> reader = streamSource.createReader(options);
// Before call to BoundedReader#start, fraction consumed must be zero.
diff --git
a/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryNestedRecordsIT.java
b/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryNestedRecordsIT.java
index 872faa2..698ef66 100644
---
a/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryNestedRecordsIT.java
+++
b/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryNestedRecordsIT.java
@@ -41,8 +41,6 @@ public class BigQueryNestedRecordsIT {
private static final String UNFLATTENABLE_QUERY =
"SELECT * FROM [apache-beam-testing:big_query_nested_test.genomics_2]";
- private static Integer stringifyCount = 0;
-
/** Options supported by this class. */
public interface Options extends PipelineOptions {
diff --git
a/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryServicesImplTest.java
b/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryServicesImplTest.java
index 16c40cf..4ea3d97 100644
---
a/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryServicesImplTest.java
+++
b/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryServicesImplTest.java
@@ -1043,8 +1043,6 @@ public class BigQueryServicesImplTest {
List<FailsafeValueInSingleWindow<TableRow, TableRow>> rows = new
ArrayList<>();
rows.add(wrapValue(new TableRow()));
- final TableDataInsertAllResponse allRowsSucceeded =
- new TableDataInsertAllResponse().setInsertErrors(ImmutableList.of());
// First response is 403 non-{rate-limited, quota-exceeded}, second
response has valid payload
// but should not be invoked.
setupMockResponses(
@@ -1318,8 +1316,6 @@ public class BigQueryServicesImplTest {
@Test
public void testGetErrorInfo() throws IOException {
- DatasetServiceImpl dataService =
- new DatasetServiceImpl(bigquery, null,
PipelineOptionsFactory.create());
ErrorInfo info = new ErrorInfo();
List<ErrorInfo> infoList = new ArrayList<>();
infoList.add(info);
diff --git
a/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryToTableIT.java
b/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryToTableIT.java
index 34bf8a0..4f4354f 100644
---
a/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryToTableIT.java
+++
b/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/bigquery/BigQueryToTableIT.java
@@ -185,9 +185,8 @@ public class BigQueryToTableIT {
"abc=,2000-01-01,00:00:00",
"dec=,3000-12-31,23:59:59.990000",
"xyw=,2011-01-01,23:59:59.999999");
- QueryResponse response =
- BQ_CLIENT.queryWithRetries(
- String.format("SELECT bytes, date, time FROM [%s];", outputTable),
project);
+ BQ_CLIENT.queryWithRetries(
+ String.format("SELECT bytes, date, time FROM [%s];", outputTable),
project);
List<TableRow> tableRows =
getTableRowsFromQuery(
String.format("SELECT bytes, date, time FROM [%s];", outputTable),
MAX_RETRY);
diff --git
a/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/bigtable/BigtableIOTest.java
b/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/bigtable/BigtableIOTest.java
index abe2749..eb7429f 100644
---
a/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/bigtable/BigtableIOTest.java
+++
b/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/bigtable/BigtableIOTest.java
@@ -1284,8 +1284,6 @@ public class BigtableIOTest {
@Test
public void testWritingEmitsResultsWhenDoneInFixedWindow() throws Exception {
final String table = "table";
- final String key = "key";
- final String value = "value";
service.createTable(table);
diff --git
a/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/datastore/DatastoreV1Test.java
b/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/datastore/DatastoreV1Test.java
index 9c42a45..8371582 100644
---
a/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/datastore/DatastoreV1Test.java
+++
b/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/datastore/DatastoreV1Test.java
@@ -138,9 +138,6 @@ public class DatastoreV1Test {
public void setUp() {
MockitoAnnotations.initMocks(this);
- DatastoreV1.Read initialRead =
-
DatastoreIO.v1().read().withProjectId(PROJECT_ID).withQuery(QUERY).withNamespace(NAMESPACE);
-
when(mockDatastoreFactory.getDatastore(any(PipelineOptions.class),
any(String.class), any()))
.thenReturn(mockDatastore);
when(mockDatastoreFactory.getQuerySplitter()).thenReturn(mockQuerySplitter);
diff --git
a/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/firestore/BaseFirestoreV1WriteFnTest.java
b/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/firestore/BaseFirestoreV1WriteFnTest.java
index 727fdc0..16c90af 100644
---
a/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/firestore/BaseFirestoreV1WriteFnTest.java
+++
b/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/firestore/BaseFirestoreV1WriteFnTest.java
@@ -410,7 +410,7 @@ public abstract class BaseFirestoreV1WriteFnTest<
LOG.debug("request = {}", request);
long requestDurationMs = 0;
BatchWriteResponse.Builder builder =
BatchWriteResponse.newBuilder();
- for (Write w : request.getWritesList()) {
+ for (Write ignored : request.getWritesList()) {
builder.addWriteResults(WriteResult.newBuilder().build());
if (clock.prev.isBefore(threshold)) {
requestDurationMs += defaultDocumentWriteLatency;
@@ -853,19 +853,16 @@ public abstract class BaseFirestoreV1WriteFnTest<
private static class MyCounter implements Counter {
private final MetricName named;
- private long incInvocationCount;
private final List<Long> incInvocations;
public MyCounter(String namespace, String name) {
named = MetricName.named(namespace, name);
- incInvocationCount = 0;
incInvocations = new ArrayList<>();
}
@Override
public void inc() {
LOG.trace("{} {}:inc()", METRIC_MARKER, named);
- incInvocationCount++;
}
@Override
diff --git
a/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/healthcare/FhirIOLROIT.java
b/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/healthcare/FhirIOLROIT.java
index 5ba4f7c..ddadf57 100644
---
a/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/healthcare/FhirIOLROIT.java
+++
b/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/healthcare/FhirIOLROIT.java
@@ -24,7 +24,6 @@ import
com.google.api.services.healthcare.v1.model.DeidentifyConfig;
import java.io.IOException;
import java.security.SecureRandom;
import org.apache.beam.sdk.testing.TestPipeline;
-import org.apache.beam.sdk.values.PCollection;
import org.junit.After;
import org.junit.AfterClass;
import org.junit.Before;
@@ -92,8 +91,7 @@ public class FhirIOLROIT {
String fhirStoreName = healthcareDataset + "/fhirStores/" + fhirStoreId;
String exportGcsUriPrefix =
"gs://" + DEFAULT_TEMP_BUCKET + "/export/" + new
SecureRandom().nextInt(32);
- PCollection<String> resources =
- pipeline.apply(FhirIO.exportResourcesToGcs(fhirStoreName,
exportGcsUriPrefix));
+ pipeline.apply(FhirIO.exportResourcesToGcs(fhirStoreName,
exportGcsUriPrefix));
pipeline.run();
}
diff --git
a/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/healthcare/FhirIOTestUtil.java
b/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/healthcare/FhirIOTestUtil.java
index 6f2a6ce..c5c2c9e 100644
---
a/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/healthcare/FhirIOTestUtil.java
+++
b/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/healthcare/FhirIOTestUtil.java
@@ -17,7 +17,6 @@
*/
package org.apache.beam.sdk.io.gcp.healthcare;
-import com.fasterxml.jackson.databind.ObjectMapper;
import com.google.api.client.http.HttpHeaders;
import com.google.api.client.http.HttpRequestInitializer;
import com.google.api.client.http.javanet.NetHttpTransport;
@@ -46,11 +45,9 @@ import java.util.stream.Stream;
import
org.apache.beam.sdk.io.gcp.healthcare.HttpHealthcareApiClient.HealthcareHttpException;
class FhirIOTestUtil {
- private static final ObjectMapper mapper = new ObjectMapper();
public static final String DEFAULT_TEMP_BUCKET =
"temp-storage-for-healthcare-io-tests";
private static Stream<String> readPrettyBundles(String version) {
- ClassLoader classLoader = FhirIOTestUtil.class.getClassLoader();
Path resourceDir = Paths.get("build", "resources", "test", version);
String absolutePath = resourceDir.toFile().getAbsolutePath();
File dir = new File(absolutePath);
diff --git
a/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/pubsub/PubsubReadIT.java
b/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/pubsub/PubsubReadIT.java
index 66b4582..7b370eb 100644
---
a/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/pubsub/PubsubReadIT.java
+++
b/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/pubsub/PubsubReadIT.java
@@ -30,13 +30,10 @@ import org.junit.Rule;
import org.junit.Test;
import org.junit.runner.RunWith;
import org.junit.runners.JUnit4;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
/** Integration test for PubsubIO. */
@RunWith(JUnit4.class)
public class PubsubReadIT {
- private static final Logger LOG =
LoggerFactory.getLogger(PubsubReadIT.class);
@Rule public transient TestPubsubSignal signal = TestPubsubSignal.create();
@Rule public transient TestPipeline pipeline = TestPipeline.create();
diff --git
a/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/spanner/ReadSpannerSchemaTest.java
b/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/spanner/ReadSpannerSchemaTest.java
index 306039a..33888cd 100644
---
a/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/spanner/ReadSpannerSchemaTest.java
+++
b/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/spanner/ReadSpannerSchemaTest.java
@@ -47,7 +47,6 @@ public class ReadSpannerSchemaTest {
@Rule public final transient ExpectedException thrown =
ExpectedException.none();
private FakeServiceFactory serviceFactory;
- private ReadOnlyTransaction mockTx;
private static Struct columnMetadata(String tableName, String columnName,
String type) {
return Struct.newBuilder()
@@ -122,7 +121,6 @@ public class ReadSpannerSchemaTest {
@SuppressWarnings("unchecked")
public void setUp() throws Exception {
serviceFactory = new FakeServiceFactory();
- mockTx = mock(ReadOnlyTransaction.class);
}
@Test
diff --git
a/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/spanner/SpannerIOReadTest.java
b/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/spanner/SpannerIOReadTest.java
index 3c886d1..8e1c833 100644
---
a/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/spanner/SpannerIOReadTest.java
+++
b/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/spanner/SpannerIOReadTest.java
@@ -288,14 +288,13 @@ public class SpannerIOReadTest implements Serializable {
SpannerConfig spannerConfig = getSpannerConfig();
- PCollection<Struct> one =
- pipeline.apply(
- "read q",
- SpannerIO.read()
- .withSpannerConfig(spannerConfig)
- .withQuery("SELECT * FROM users")
- .withQueryName("queryName")
- .withTimestampBound(timestampBound));
+ pipeline.apply(
+ "read q",
+ SpannerIO.read()
+ .withSpannerConfig(spannerConfig)
+ .withQuery("SELECT * FROM users")
+ .withQueryName("queryName")
+ .withTimestampBound(timestampBound));
FakeBatchTransactionId id = new FakeBatchTransactionId("runQueryTest");
when(mockBatchTx.getBatchTransactionId()).thenReturn(id);
@@ -336,14 +335,13 @@ public class SpannerIOReadTest implements Serializable {
SpannerConfig spannerConfig = getSpannerConfig();
- PCollection<Struct> one =
- pipeline.apply(
- "read q",
- SpannerIO.read()
- .withSpannerConfig(spannerConfig)
- .withTable("users")
- .withColumns("id", "name")
- .withTimestampBound(timestampBound));
+ pipeline.apply(
+ "read q",
+ SpannerIO.read()
+ .withSpannerConfig(spannerConfig)
+ .withTable("users")
+ .withColumns("id", "name")
+ .withTimestampBound(timestampBound));
FakeBatchTransactionId id = new FakeBatchTransactionId("runReadTest");
when(mockBatchTx.getBatchTransactionId()).thenReturn(id);
diff --git
a/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/spanner/SpannerIOWriteTest.java
b/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/spanner/SpannerIOWriteTest.java
index aa275f9..ea7ca68 100644
---
a/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/spanner/SpannerIOWriteTest.java
+++
b/sdks/java/io/google-cloud-platform/src/test/java/org/apache/beam/sdk/io/gcp/spanner/SpannerIOWriteTest.java
@@ -995,7 +995,6 @@ public class SpannerIOWriteTest implements Serializable {
g(range)
};
- long mutationSize = MutationSizeEstimator.sizeOf(m(1L));
BatchableMutationFilterFn testFn = new BatchableMutationFilterFn(null,
null, 1000, 1000, 3);
BatchableMutationFilterFn.ProcessContext mockProcessContext =
diff --git a/sdks/java/io/hadoop-common/build.gradle
b/sdks/java/io/hadoop-common/build.gradle
index 327e6f3..6c5cc45 100644
--- a/sdks/java/io/hadoop-common/build.gradle
+++ b/sdks/java/io/hadoop-common/build.gradle
@@ -19,7 +19,7 @@
import java.util.stream.Collectors
plugins { id 'org.apache.beam.module' }
-applyJavaNature(suppressUnusedVariable: true, automaticModuleName:
'org.apache.beam.sdk.io.hadoop.common')
+applyJavaNature( automaticModuleName: 'org.apache.beam.sdk.io.hadoop.common')
description = "Apache Beam :: SDKs :: Java :: IO :: Hadoop Common"
ext.summary = "Library to add shared Hadoop classes among Beam IOs."
diff --git a/sdks/java/io/hadoop-file-system/build.gradle
b/sdks/java/io/hadoop-file-system/build.gradle
index 65aff2a..940dcb7 100644
--- a/sdks/java/io/hadoop-file-system/build.gradle
+++ b/sdks/java/io/hadoop-file-system/build.gradle
@@ -20,7 +20,6 @@ import java.util.stream.Collectors
plugins { id 'org.apache.beam.module' }
applyJavaNature(
- suppressUnusedVariable: true,
automaticModuleName: 'org.apache.beam.sdk.io.hdfs')
description = "Apache Beam :: SDKs :: Java :: IO :: Hadoop File System"
diff --git a/sdks/java/io/hadoop-format/build.gradle
b/sdks/java/io/hadoop-format/build.gradle
index d5301ad..d9f2c98 100644
--- a/sdks/java/io/hadoop-format/build.gradle
+++ b/sdks/java/io/hadoop-format/build.gradle
@@ -21,7 +21,6 @@ import java.util.stream.Collectors
plugins { id 'org.apache.beam.module' }
applyJavaNature(
- suppressUnusedVariable: true,
automaticModuleName: 'org.apache.beam.sdk.io.hadoop.format',
)
provideIntegrationTestingDependencies()
diff --git
a/sdks/java/io/hadoop-format/src/test/java/org/apache/beam/sdk/io/hadoop/format/HadoopFormatIOIT.java
b/sdks/java/io/hadoop-format/src/test/java/org/apache/beam/sdk/io/hadoop/format/HadoopFormatIOIT.java
index 822adb1..8973f58 100644
---
a/sdks/java/io/hadoop-format/src/test/java/org/apache/beam/sdk/io/hadoop/format/HadoopFormatIOIT.java
+++
b/sdks/java/io/hadoop-format/src/test/java/org/apache/beam/sdk/io/hadoop/format/HadoopFormatIOIT.java
@@ -102,8 +102,6 @@ public class HadoopFormatIOIT {
private static Integer numberOfRows;
private static String tableName;
private static SerializableConfiguration hadoopConfiguration;
- private static String bigQueryDataset;
- private static String bigQueryTable;
private static InfluxDBSettings settings;
private static HadoopFormatIOITOptions options;
@@ -133,8 +131,6 @@ public class HadoopFormatIOIT {
dataSource = DatabaseTestHelper.getPostgresDataSource(options);
numberOfRows = options.getNumberOfRecords();
tableName = DatabaseTestHelper.getTestTableName("HadoopFormatIOIT");
- bigQueryDataset = options.getBigQueryDataset();
- bigQueryTable = options.getBigQueryTable();
if (!options.isWithTestcontainers()) {
settings =
InfluxDBSettings.builder()
diff --git
a/sdks/java/io/hadoop-format/src/test/java/org/apache/beam/sdk/io/hadoop/format/HadoopFormatIOReadTest.java
b/sdks/java/io/hadoop-format/src/test/java/org/apache/beam/sdk/io/hadoop/format/HadoopFormatIOReadTest.java
index 8fa446f..bfad990 100644
---
a/sdks/java/io/hadoop-format/src/test/java/org/apache/beam/sdk/io/hadoop/format/HadoopFormatIOReadTest.java
+++
b/sdks/java/io/hadoop-format/src/test/java/org/apache/beam/sdk/io/hadoop/format/HadoopFormatIOReadTest.java
@@ -48,7 +48,6 @@ import org.apache.beam.sdk.testing.TestPipeline;
import org.apache.beam.sdk.transforms.SimpleFunction;
import org.apache.beam.sdk.transforms.display.DisplayData;
import org.apache.beam.sdk.values.KV;
-import org.apache.beam.sdk.values.PBegin;
import org.apache.beam.sdk.values.PCollection;
import org.apache.beam.sdk.values.Row;
import org.apache.hadoop.conf.Configuration;
@@ -84,8 +83,6 @@ public class HadoopFormatIOReadTest {
@Rule public final transient TestPipeline p = TestPipeline.create();
@Rule public ExpectedException thrown = ExpectedException.none();
- private PBegin input = PBegin.in(p);
-
@BeforeClass
public static void setUp() {
serConf = loadTestConfiguration(EmployeeInputFormat.class, Text.class,
Employee.class);
diff --git a/sdks/java/io/hbase/build.gradle b/sdks/java/io/hbase/build.gradle
index 2de5621..6dc9bb4 100644
--- a/sdks/java/io/hbase/build.gradle
+++ b/sdks/java/io/hbase/build.gradle
@@ -17,7 +17,7 @@
*/
plugins { id 'org.apache.beam.module' }
-applyJavaNature(suppressUnusedVariable: true, automaticModuleName:
'org.apache.beam.sdk.io.hbase')
+applyJavaNature( automaticModuleName: 'org.apache.beam.sdk.io.hbase')
provideIntegrationTestingDependencies()
enableJavaPerformanceTesting()
diff --git
a/sdks/java/io/hbase/src/main/java/org/apache/beam/sdk/io/hbase/HBaseIO.java
b/sdks/java/io/hbase/src/main/java/org/apache/beam/sdk/io/hbase/HBaseIO.java
index af925f5..04e0add 100644
--- a/sdks/java/io/hbase/src/main/java/org/apache/beam/sdk/io/hbase/HBaseIO.java
+++ b/sdks/java/io/hbase/src/main/java/org/apache/beam/sdk/io/hbase/HBaseIO.java
@@ -716,7 +716,6 @@ public class HBaseIO {
HBaseWriterFn(Write write) {
checkNotNull(write.tableId, "tableId");
checkNotNull(write.configuration, "configuration");
- this.write = write;
}
@Setup
@@ -760,7 +759,6 @@ public class HBaseIO {
builder.delegate(Write.this);
}
- private final Write write;
private long recordsWritten;
private transient Connection connection;
diff --git a/sdks/java/io/hcatalog/build.gradle
b/sdks/java/io/hcatalog/build.gradle
index 1fc3a16..67eb3d5 100644
--- a/sdks/java/io/hcatalog/build.gradle
+++ b/sdks/java/io/hcatalog/build.gradle
@@ -19,7 +19,7 @@
import java.util.stream.Collectors
plugins { id 'org.apache.beam.module' }
-applyJavaNature(suppressUnusedVariable: true, automaticModuleName:
'org.apache.beam.sdk.io.hcatalog')
+applyJavaNature( automaticModuleName: 'org.apache.beam.sdk.io.hcatalog')
description = "Apache Beam :: SDKs :: Java :: IO :: HCatalog"
ext.summary = "IO to read and write for HCatalog source."
diff --git a/sdks/java/io/influxdb/build.gradle
b/sdks/java/io/influxdb/build.gradle
index 1f84d02..706c51f 100644
--- a/sdks/java/io/influxdb/build.gradle
+++ b/sdks/java/io/influxdb/build.gradle
@@ -17,7 +17,7 @@
*/
plugins { id 'org.apache.beam.module' }
-applyJavaNature(suppressUnusedVariable: true, automaticModuleName:
'org.apache.beam.sdk.io.influxdb')
+applyJavaNature( automaticModuleName: 'org.apache.beam.sdk.io.influxdb')
provideIntegrationTestingDependencies()
enableJavaPerformanceTesting()
@@ -29,7 +29,6 @@ dependencies {
compile library.java.influxdb_library
compile "com.squareup.okhttp3:okhttp:4.6.0"
compile library.java.vendored_guava_26_0_jre
- compile library.java.slf4j_api
testCompile library.java.junit
testCompile library.java.powermock
testCompile library.java.powermock_mockito
diff --git
a/sdks/java/io/influxdb/src/main/java/org/apache/beam/sdk/io/influxdb/InfluxDbIO.java
b/sdks/java/io/influxdb/src/main/java/org/apache/beam/sdk/io/influxdb/InfluxDbIO.java
index 779f35e..b9cc3e4 100644
---
a/sdks/java/io/influxdb/src/main/java/org/apache/beam/sdk/io/influxdb/InfluxDbIO.java
+++
b/sdks/java/io/influxdb/src/main/java/org/apache/beam/sdk/io/influxdb/InfluxDbIO.java
@@ -56,8 +56,6 @@ import org.influxdb.dto.Query;
import org.influxdb.dto.QueryResult;
import org.influxdb.dto.QueryResult.Result;
import org.influxdb.dto.QueryResult.Series;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
/**
* IO to read and write from InfluxDB.
@@ -103,7 +101,6 @@ import org.slf4j.LoggerFactory;
"nullness" // TODO(https://issues.apache.org/jira/browse/BEAM-10402)
})
public class InfluxDbIO {
- private static final Logger LOG = LoggerFactory.getLogger(InfluxDbIO.class);
private static final String DEFAULT_RETENTION_POLICY = "autogen";
diff --git a/sdks/java/io/jdbc/build.gradle b/sdks/java/io/jdbc/build.gradle
index 18809bb..ca8616b 100644
--- a/sdks/java/io/jdbc/build.gradle
+++ b/sdks/java/io/jdbc/build.gradle
@@ -18,7 +18,6 @@
plugins { id 'org.apache.beam.module' }
applyJavaNature(
- suppressUnusedVariable: true,
automaticModuleName: 'org.apache.beam.sdk.io.jdbc',
)
provideIntegrationTestingDependencies()
diff --git
a/sdks/java/io/jdbc/src/test/java/org/apache/beam/sdk/io/jdbc/JdbcIOIT.java
b/sdks/java/io/jdbc/src/test/java/org/apache/beam/sdk/io/jdbc/JdbcIOIT.java
index 8ea7086..8cebbbd 100644
--- a/sdks/java/io/jdbc/src/test/java/org/apache/beam/sdk/io/jdbc/JdbcIOIT.java
+++ b/sdks/java/io/jdbc/src/test/java/org/apache/beam/sdk/io/jdbc/JdbcIOIT.java
@@ -88,8 +88,6 @@ public class JdbcIOIT {
private static int numberOfRows;
private static PGSimpleDataSource dataSource;
private static String tableName;
- private static String bigQueryDataset;
- private static String bigQueryTable;
private static Long tableSize;
private static InfluxDBSettings settings;
@Rule public TestPipeline pipelineWrite = TestPipeline.create();
@@ -100,8 +98,6 @@ public class JdbcIOIT {
PostgresIOTestPipelineOptions options =
readIOTestPipelineOptions(PostgresIOTestPipelineOptions.class);
- bigQueryDataset = options.getBigQueryDataset();
- bigQueryTable = options.getBigQueryTable();
numberOfRows = options.getNumberOfRecords();
dataSource = DatabaseTestHelper.getPostgresDataSource(options);
tableName = DatabaseTestHelper.getTestTableName("IT");
diff --git
a/sdks/java/io/jdbc/src/test/java/org/apache/beam/sdk/io/jdbc/JdbcIOTest.java
b/sdks/java/io/jdbc/src/test/java/org/apache/beam/sdk/io/jdbc/JdbcIOTest.java
index 9d1d75b..60d8cc7 100644
---
a/sdks/java/io/jdbc/src/test/java/org/apache/beam/sdk/io/jdbc/JdbcIOTest.java
+++
b/sdks/java/io/jdbc/src/test/java/org/apache/beam/sdk/io/jdbc/JdbcIOTest.java
@@ -94,13 +94,11 @@ import org.junit.Test;
import org.junit.rules.ExpectedException;
import org.junit.runner.RunWith;
import org.junit.runners.JUnit4;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
/** Test on the JdbcIO. */
@RunWith(JUnit4.class)
public class JdbcIOTest implements Serializable {
- private static final Logger LOG = LoggerFactory.getLogger(JdbcIOTest.class);
+
private static final DataSourceConfiguration DATA_SOURCE_CONFIGURATION =
DataSourceConfiguration.create(
"org.apache.derby.jdbc.EmbeddedDriver",
"jdbc:derby:memory:testDB;create=true");
@@ -462,17 +460,16 @@ public class JdbcIOTest implements Serializable {
public void testIfNumPartitionsIsZero() {
thrown.expect(IllegalArgumentException.class);
thrown.expectMessage("numPartitions can not be less than 1");
- PCollection<TestRow> rows =
- pipeline.apply(
- JdbcIO.<TestRow>readWithPartitions()
- .withDataSourceConfiguration(DATA_SOURCE_CONFIGURATION)
- .withRowMapper(new JdbcTestHelper.CreateTestRowOfNameAndId())
- .withCoder(SerializableCoder.of(TestRow.class))
- .withTable(READ_TABLE_NAME)
- .withNumPartitions(0)
- .withPartitionColumn("id")
- .withLowerBound(0L)
- .withUpperBound(1000L));
+ pipeline.apply(
+ JdbcIO.<TestRow>readWithPartitions()
+ .withDataSourceConfiguration(DATA_SOURCE_CONFIGURATION)
+ .withRowMapper(new JdbcTestHelper.CreateTestRowOfNameAndId())
+ .withCoder(SerializableCoder.of(TestRow.class))
+ .withTable(READ_TABLE_NAME)
+ .withNumPartitions(0)
+ .withPartitionColumn("id")
+ .withLowerBound(0L)
+ .withUpperBound(1000L));
pipeline.run();
}
@@ -956,7 +953,6 @@ public class JdbcIOTest implements Serializable {
ArrayList<Row> data = new ArrayList<>();
for (int i = 0; i < rowsToAdd; i++) {
- List<Object> fields = new ArrayList<>();
Row row =
schema.getFields().stream()
diff --git a/sdks/java/io/jms/build.gradle b/sdks/java/io/jms/build.gradle
index fee6cc3..3f9d0c7 100644
--- a/sdks/java/io/jms/build.gradle
+++ b/sdks/java/io/jms/build.gradle
@@ -18,7 +18,6 @@
plugins { id 'org.apache.beam.module' }
applyJavaNature(
- suppressUnusedVariable: true,
automaticModuleName: 'org.apache.beam.sdk.io.jms',
)
diff --git a/sdks/java/io/kafka/build.gradle b/sdks/java/io/kafka/build.gradle
index 8ee0fc5..7a4ca21 100644
--- a/sdks/java/io/kafka/build.gradle
+++ b/sdks/java/io/kafka/build.gradle
@@ -20,7 +20,6 @@ import java.util.stream.Collectors
plugins { id 'org.apache.beam.module' }
applyJavaNature(
- suppressUnusedVariable: true,
automaticModuleName: 'org.apache.beam.sdk.io.kafka',
mavenRepositories: [
[id: 'io.confluent', url: 'https://packages.confluent.io/maven/']
diff --git
a/sdks/java/io/kafka/src/main/java/org/apache/beam/sdk/io/kafka/KafkaExactlyOnceSink.java
b/sdks/java/io/kafka/src/main/java/org/apache/beam/sdk/io/kafka/KafkaExactlyOnceSink.java
index d4589a2..75c8270 100644
---
a/sdks/java/io/kafka/src/main/java/org/apache/beam/sdk/io/kafka/KafkaExactlyOnceSink.java
+++
b/sdks/java/io/kafka/src/main/java/org/apache/beam/sdk/io/kafka/KafkaExactlyOnceSink.java
@@ -90,7 +90,8 @@ import org.slf4j.LoggerFactory;
*/
@SuppressWarnings({
"rawtypes", // TODO(https://issues.apache.org/jira/browse/BEAM-10556)
- "nullness" // TODO(https://issues.apache.org/jira/browse/BEAM-10402)
+ "nullness", // TODO(https://issues.apache.org/jira/browse/BEAM-10402)
+ "unused" // TODO(BEAM-13271): Remove when new version of errorprone is
released (2.11.0)
})
class KafkaExactlyOnceSink<K, V>
extends PTransform<PCollection<ProducerRecord<K, V>>, PCollection<Void>> {
@@ -253,6 +254,7 @@ class KafkaExactlyOnceSink<K, V>
// started with same groupId used for storing state on Kafka side,
including the case where
// a job is restarted with same groupId, but the metadata from previous
run was not cleared.
// Better to be safe and error out with a clear message.
+
@StateId(WRITER_ID)
private final StateSpec<ValueState<String>> writerIdSpec =
StateSpecs.value();
diff --git
a/sdks/java/io/kafka/src/main/java/org/apache/beam/sdk/io/kafka/WatchKafkaTopicPartitionDoFn.java
b/sdks/java/io/kafka/src/main/java/org/apache/beam/sdk/io/kafka/WatchKafkaTopicPartitionDoFn.java
index 46cd823..c4e6272 100644
---
a/sdks/java/io/kafka/src/main/java/org/apache/beam/sdk/io/kafka/WatchKafkaTopicPartitionDoFn.java
+++
b/sdks/java/io/kafka/src/main/java/org/apache/beam/sdk/io/kafka/WatchKafkaTopicPartitionDoFn.java
@@ -49,7 +49,10 @@ import org.joda.time.Instant;
*
https://docs.google.com/document/d/1FU3GxVRetHPLVizP3Mdv6mP5tpjZ3fd99qNjUI5DT5k/edit#
for more
* details.
*/
-@SuppressWarnings({"nullness"})
+@SuppressWarnings({
+ "nullness",
+ "unused" // TODO(BEAM-13271): Remove when new version of errorprone is
released (2.11.0)
+})
@Experimental
class WatchKafkaTopicPartitionDoFn extends DoFn<KV<byte[], byte[]>,
KafkaSourceDescriptor> {
diff --git
a/sdks/java/io/kafka/src/test/java/org/apache/beam/sdk/io/kafka/KafkaIOExternalTest.java
b/sdks/java/io/kafka/src/test/java/org/apache/beam/sdk/io/kafka/KafkaIOExternalTest.java
index c699885..89f64b1 100644
---
a/sdks/java/io/kafka/src/test/java/org/apache/beam/sdk/io/kafka/KafkaIOExternalTest.java
+++
b/sdks/java/io/kafka/src/test/java/org/apache/beam/sdk/io/kafka/KafkaIOExternalTest.java
@@ -255,8 +255,7 @@ public class KafkaIOExternalTest {
RunnerApi.PTransform kafkaReadComposite =
result.getComponents().getTransformsOrThrow(transform.getSubtransforms(0));
- RunnerApi.PTransform kafkaComposite =
-
result.getComponents().getTransformsOrThrow(kafkaReadComposite.getSubtransforms(0));
+
result.getComponents().getTransformsOrThrow(kafkaReadComposite.getSubtransforms(0));
verifyKafkaReadComposite(
result.getComponents().getTransformsOrThrow(kafkaReadComposite.getSubtransforms(0)),
diff --git a/sdks/java/io/kinesis/build.gradle
b/sdks/java/io/kinesis/build.gradle
index 0085df6..cc5334f 100644
--- a/sdks/java/io/kinesis/build.gradle
+++ b/sdks/java/io/kinesis/build.gradle
@@ -17,7 +17,7 @@
*/
plugins { id 'org.apache.beam.module' }
-applyJavaNature(suppressUnusedVariable: true, automaticModuleName:
'org.apache.beam.sdk.io.kinesis')
+applyJavaNature( automaticModuleName: 'org.apache.beam.sdk.io.kinesis')
provideIntegrationTestingDependencies()
enableJavaPerformanceTesting()
diff --git a/sdks/java/io/kinesis/expansion-service/build.gradle
b/sdks/java/io/kinesis/expansion-service/build.gradle
index 5974901..c186e63 100644
--- a/sdks/java/io/kinesis/expansion-service/build.gradle
+++ b/sdks/java/io/kinesis/expansion-service/build.gradle
@@ -21,7 +21,6 @@ apply plugin: 'application'
mainClassName = "org.apache.beam.sdk.expansion.service.ExpansionService"
applyJavaNature(
- suppressUnusedVariable: true,
automaticModuleName: 'org.apache.beam.sdk.io.kinesis.expansion.service',
exportJavadoc: false,
validateShadowJar: false,
diff --git
a/sdks/java/io/kinesis/src/test/java/org/apache/beam/sdk/io/kinesis/KinesisIOReadTest.java
b/sdks/java/io/kinesis/src/test/java/org/apache/beam/sdk/io/kinesis/KinesisIOReadTest.java
index 1f7c493..fdacc62 100644
---
a/sdks/java/io/kinesis/src/test/java/org/apache/beam/sdk/io/kinesis/KinesisIOReadTest.java
+++
b/sdks/java/io/kinesis/src/test/java/org/apache/beam/sdk/io/kinesis/KinesisIOReadTest.java
@@ -68,7 +68,6 @@ public class KinesisIOReadTest {
public void testReadWithBasicCredentialsAndCustomEndpoint() {
String customEndpoint = "localhost:9999";
Regions region = Regions.US_WEST_1;
- BasicAWSCredentials credentials = new BasicAWSCredentials(ACCESS_KEY_ID,
SECRET_ACCESS_KEY);
Read<KinesisRecord> read =
KinesisIO.read()
@@ -102,7 +101,6 @@ public class KinesisIOReadTest {
public void testReadWithBasicCredentialsAndVerificationDisabled() {
String customEndpoint = "localhost:9999";
Regions region = Regions.US_WEST_1;
- BasicAWSCredentials credentials = new BasicAWSCredentials(ACCESS_KEY_ID,
SECRET_ACCESS_KEY);
Read<KinesisRecord> read =
KinesisIO.read()
diff --git
a/sdks/java/io/kinesis/src/test/java/org/apache/beam/sdk/io/kinesis/serde/AwsModuleTest.java
b/sdks/java/io/kinesis/src/test/java/org/apache/beam/sdk/io/kinesis/serde/AwsModuleTest.java
index 8ceda61..e58825e 100644
---
a/sdks/java/io/kinesis/src/test/java/org/apache/beam/sdk/io/kinesis/serde/AwsModuleTest.java
+++
b/sdks/java/io/kinesis/src/test/java/org/apache/beam/sdk/io/kinesis/serde/AwsModuleTest.java
@@ -99,8 +99,6 @@ public class AwsModuleTest {
credentialsProvider =
new AWSStaticCredentialsProvider(
new BasicSessionCredentials(ACCESS_KEY_ID, SECRET_ACCESS_KEY,
SESSION_TOKEN));
- serializedCredentialsProvider = serialize(credentialsProvider);
- deserializedCredentialsProvider =
deserializeCredentialsProvider(serializedCredentialsProvider);
checkStaticSessionCredentials(credentialsProvider);
}
diff --git a/sdks/java/io/kudu/build.gradle b/sdks/java/io/kudu/build.gradle
index 6509211..7f28d5e 100644
--- a/sdks/java/io/kudu/build.gradle
+++ b/sdks/java/io/kudu/build.gradle
@@ -18,7 +18,6 @@
plugins { id 'org.apache.beam.module' }
applyJavaNature(
- suppressUnusedVariable: true,
classesTriggerCheckerBugs: [
'KuduTestUtils': 'TODO: file a bug report',
'KuduIOIT': 'TODO: file a bug report',
diff --git
a/sdks/java/io/kudu/src/main/java/org/apache/beam/sdk/io/kudu/KuduIO.java
b/sdks/java/io/kudu/src/main/java/org/apache/beam/sdk/io/kudu/KuduIO.java
index 6ba383c..a8d794e 100644
--- a/sdks/java/io/kudu/src/main/java/org/apache/beam/sdk/io/kudu/KuduIO.java
+++ b/sdks/java/io/kudu/src/main/java/org/apache/beam/sdk/io/kudu/KuduIO.java
@@ -50,8 +50,6 @@ import org.apache.kudu.client.KuduPredicate;
import org.apache.kudu.client.Operation;
import org.apache.kudu.client.RowResult;
import org.checkerframework.checker.nullness.qual.Nullable;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
/**
* A bounded source and sink for Kudu.
@@ -126,7 +124,6 @@ import org.slf4j.LoggerFactory;
"nullness" // TODO(https://issues.apache.org/jira/browse/BEAM-10402)
})
public class KuduIO {
- private static final Logger LOG = LoggerFactory.getLogger(KuduIO.class);
private KuduIO() {}
diff --git
a/sdks/java/io/kudu/src/test/java/org/apache/beam/sdk/io/kudu/KuduIOIT.java
b/sdks/java/io/kudu/src/test/java/org/apache/beam/sdk/io/kudu/KuduIOIT.java
index a50209d..1add8f2 100644
--- a/sdks/java/io/kudu/src/test/java/org/apache/beam/sdk/io/kudu/KuduIOIT.java
+++ b/sdks/java/io/kudu/src/test/java/org/apache/beam/sdk/io/kudu/KuduIOIT.java
@@ -52,8 +52,6 @@ import org.junit.Test;
import org.junit.rules.ExpectedException;
import org.junit.runner.RunWith;
import org.junit.runners.JUnit4;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
/**
* A test of {@link org.apache.beam.sdk.io.kudu.KuduIO} on an independent Kudu
instance.
@@ -90,7 +88,6 @@ import org.slf4j.LoggerFactory;
*/
@RunWith(JUnit4.class)
public class KuduIOIT {
- private static final Logger LOG = LoggerFactory.getLogger(KuduIOIT.class);
/** KuduIOIT options. */
public interface KuduPipelineOptions extends IOTestPipelineOptions {
diff --git
a/sdks/java/io/kudu/src/test/java/org/apache/beam/sdk/io/kudu/KuduIOTest.java
b/sdks/java/io/kudu/src/test/java/org/apache/beam/sdk/io/kudu/KuduIOTest.java
index 86d60e0..66f908c 100644
---
a/sdks/java/io/kudu/src/test/java/org/apache/beam/sdk/io/kudu/KuduIOTest.java
+++
b/sdks/java/io/kudu/src/test/java/org/apache/beam/sdk/io/kudu/KuduIOTest.java
@@ -66,7 +66,6 @@ import org.slf4j.LoggerFactory;
*/
@RunWith(JUnit4.class)
public class KuduIOTest {
- private static final Logger LOG = LoggerFactory.getLogger(KuduIOTest.class);
@Rule public final TestPipeline writePipeline = TestPipeline.create();
@Rule public final TestPipeline readPipeline = TestPipeline.create();
diff --git a/sdks/java/io/mongodb/build.gradle
b/sdks/java/io/mongodb/build.gradle
index e8aa0e3..1090a2c 100644
--- a/sdks/java/io/mongodb/build.gradle
+++ b/sdks/java/io/mongodb/build.gradle
@@ -17,7 +17,7 @@
*/
plugins { id 'org.apache.beam.module' }
-applyJavaNature(suppressUnusedVariable: true, automaticModuleName:
'org.apache.beam.sdk.io.mongodb')
+applyJavaNature( automaticModuleName: 'org.apache.beam.sdk.io.mongodb')
provideIntegrationTestingDependencies()
enableJavaPerformanceTesting()
diff --git
a/sdks/java/io/mongodb/src/main/java/org/apache/beam/sdk/io/mongodb/MongoDbIO.java
b/sdks/java/io/mongodb/src/main/java/org/apache/beam/sdk/io/mongodb/MongoDbIO.java
index 8c7f03b..c70bfae 100644
---
a/sdks/java/io/mongodb/src/main/java/org/apache/beam/sdk/io/mongodb/MongoDbIO.java
+++
b/sdks/java/io/mongodb/src/main/java/org/apache/beam/sdk/io/mongodb/MongoDbIO.java
@@ -628,7 +628,6 @@ public class MongoDbIO {
ObjectId lowestBound = null; // lower boundary (previous split in the
iteration)
for (int i = 0; i < splitKeys.size(); i++) {
ObjectId splitKey = splitKeys.get(i).getObjectId("_id");
- String rangeFilter;
if (i == 0) {
aggregates.add(Aggregates.match(Filters.lte("_id", splitKey)));
if (splitKeys.size() == 1) {
diff --git
a/sdks/java/io/mongodb/src/test/java/org/apache/beam/sdk/io/mongodb/MongoDBIOIT.java
b/sdks/java/io/mongodb/src/test/java/org/apache/beam/sdk/io/mongodb/MongoDBIOIT.java
index cd83275..6b696a7 100644
---
a/sdks/java/io/mongodb/src/test/java/org/apache/beam/sdk/io/mongodb/MongoDBIOIT.java
+++
b/sdks/java/io/mongodb/src/test/java/org/apache/beam/sdk/io/mongodb/MongoDBIOIT.java
@@ -81,8 +81,6 @@ import org.junit.runners.JUnit4;
public class MongoDBIOIT {
private static final String NAMESPACE = MongoDBIOIT.class.getName();
- private static String bigQueryDataset;
- private static String bigQueryTable;
private static String mongoUrl;
private static MongoClient mongoClient;
private static InfluxDBSettings settings;
@@ -128,8 +126,6 @@ public class MongoDBIOIT {
PipelineOptionsFactory.register(MongoDBPipelineOptions.class);
options =
TestPipeline.testingPipelineOptions().as(MongoDBPipelineOptions.class);
collection = String.format("test_%s", new Date().getTime());
- bigQueryDataset = options.getBigQueryDataset();
- bigQueryTable = options.getBigQueryTable();
mongoUrl =
String.format("mongodb://%s:%s", options.getMongoDBHostName(),
options.getMongoDBPort());
mongoClient = MongoClients.create(mongoUrl);
diff --git a/sdks/java/io/mqtt/build.gradle b/sdks/java/io/mqtt/build.gradle
index 0468bd2..f6e9925 100644
--- a/sdks/java/io/mqtt/build.gradle
+++ b/sdks/java/io/mqtt/build.gradle
@@ -17,7 +17,7 @@
*/
plugins { id 'org.apache.beam.module' }
-applyJavaNature(suppressUnusedVariable: true, automaticModuleName:
'org.apache.beam.sdk.io.mqtt')
+applyJavaNature( automaticModuleName: 'org.apache.beam.sdk.io.mqtt')
description = "Apache Beam :: SDKs :: Java :: IO :: MQTT"
ext.summary = "IO to read and write to a MQTT broker."
diff --git a/sdks/java/io/parquet/build.gradle
b/sdks/java/io/parquet/build.gradle
index 92582e5..8e9f679 100644
--- a/sdks/java/io/parquet/build.gradle
+++ b/sdks/java/io/parquet/build.gradle
@@ -20,7 +20,6 @@ import java.util.stream.Collectors
plugins { id 'org.apache.beam.module' }
applyJavaNature(
- suppressUnusedVariable: true,
automaticModuleName: 'org.apache.beam.sdk.io.parquet',
)
diff --git
a/sdks/java/io/parquet/src/test/java/org/apache/beam/sdk/io/parquet/ParquetIOTest.java
b/sdks/java/io/parquet/src/test/java/org/apache/beam/sdk/io/parquet/ParquetIOTest.java
index 332e896..d03f6a3 100644
---
a/sdks/java/io/parquet/src/test/java/org/apache/beam/sdk/io/parquet/ParquetIOTest.java
+++
b/sdks/java/io/parquet/src/test/java/org/apache/beam/sdk/io/parquet/ParquetIOTest.java
@@ -386,7 +386,6 @@ public class ParquetIOTest implements Serializable {
ArrayList<GenericRecord> data = new ArrayList<>();
GenericRecordBuilder builder = new
GenericRecordBuilder(REQUESTED_ENCODER_SCHEMA);
for (int i = 0; i < count; i++) {
- int index = i % SCIENTISTS.length;
GenericRecord record = builder.set("id",
Integer.toString(i)).set("name", null).build();
data.add(record);
}
diff --git a/sdks/java/io/rabbitmq/build.gradle
b/sdks/java/io/rabbitmq/build.gradle
index 08c5a79..0f9a9ac 100644
--- a/sdks/java/io/rabbitmq/build.gradle
+++ b/sdks/java/io/rabbitmq/build.gradle
@@ -17,7 +17,7 @@
*/
plugins { id 'org.apache.beam.module' }
-applyJavaNature(suppressUnusedVariable: true, automaticModuleName:
'org.apache.beam.sdk.io.rabbitmq')
+applyJavaNature( automaticModuleName: 'org.apache.beam.sdk.io.rabbitmq')
description = "Apache Beam :: SDKs :: Java :: IO :: RabbitMQ"
ext.summary = "IO to read and write to a RabbitMQ broker."
diff --git
a/sdks/java/io/rabbitmq/src/test/java/org/apache/beam/sdk/io/rabbitmq/RabbitMqIOTest.java
b/sdks/java/io/rabbitmq/src/test/java/org/apache/beam/sdk/io/rabbitmq/RabbitMqIOTest.java
index c7b3c0e..2ceae33 100644
---
a/sdks/java/io/rabbitmq/src/test/java/org/apache/beam/sdk/io/rabbitmq/RabbitMqIOTest.java
+++
b/sdks/java/io/rabbitmq/src/test/java/org/apache/beam/sdk/io/rabbitmq/RabbitMqIOTest.java
@@ -63,13 +63,10 @@ import org.junit.rules.TemporaryFolder;
import org.junit.rules.Timeout;
import org.junit.runner.RunWith;
import org.junit.runners.JUnit4;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
/** Test of {@link RabbitMqIO}. */
@RunWith(JUnit4.class)
public class RabbitMqIOTest implements Serializable {
- private static final Logger LOG =
LoggerFactory.getLogger(RabbitMqIOTest.class);
private static int port;
private static String defaultPort;
diff --git a/sdks/java/io/redis/build.gradle b/sdks/java/io/redis/build.gradle
index f7092fb..0f668e8 100644
--- a/sdks/java/io/redis/build.gradle
+++ b/sdks/java/io/redis/build.gradle
@@ -17,7 +17,7 @@
*/
plugins { id 'org.apache.beam.module' }
-applyJavaNature(suppressUnusedVariable: true, automaticModuleName:
'org.apache.beam.sdk.io.redis')
+applyJavaNature( automaticModuleName: 'org.apache.beam.sdk.io.redis')
description = "Apache Beam :: SDKs :: Java :: IO :: Redis"
ext.summary ="IO to read and write on a Redis keystore."
diff --git a/sdks/java/io/snowflake/build.gradle
b/sdks/java/io/snowflake/build.gradle
index 3fe128e..c3c5d06 100644
--- a/sdks/java/io/snowflake/build.gradle
+++ b/sdks/java/io/snowflake/build.gradle
@@ -17,7 +17,6 @@
*/
plugins { id 'org.apache.beam.module' }
applyJavaNature(
- suppressUnusedVariable: true,
automaticModuleName: 'org.apache.beam.sdk.io.snowflake',
)
provideIntegrationTestingDependencies()
diff --git a/sdks/java/io/snowflake/expansion-service/build.gradle
b/sdks/java/io/snowflake/expansion-service/build.gradle
index 23a332e..247a622 100644
--- a/sdks/java/io/snowflake/expansion-service/build.gradle
+++ b/sdks/java/io/snowflake/expansion-service/build.gradle
@@ -21,7 +21,6 @@ apply plugin: 'application'
mainClassName = "org.apache.beam.sdk.expansion.service.ExpansionService"
applyJavaNature(
- suppressUnusedVariable: true,
automaticModuleName: 'org.apache.beam.sdk.io.expansion.service',
exportJavadoc: false,
validateShadowJar: false,
diff --git
a/sdks/java/io/snowflake/src/main/java/org/apache/beam/sdk/io/snowflake/SnowflakeIO.java
b/sdks/java/io/snowflake/src/main/java/org/apache/beam/sdk/io/snowflake/SnowflakeIO.java
index 05064bf..1071997 100644
---
a/sdks/java/io/snowflake/src/main/java/org/apache/beam/sdk/io/snowflake/SnowflakeIO.java
+++
b/sdks/java/io/snowflake/src/main/java/org/apache/beam/sdk/io/snowflake/SnowflakeIO.java
@@ -1284,7 +1284,6 @@ public class SnowflakeIO {
private final SnowflakeService snowflakeService;
private transient SimpleIngestManager ingestManager;
- private transient DataSource dataSource;
ArrayList<String> trackedFilesNames;
StreamToTableFn(
@@ -1303,7 +1302,7 @@ public class SnowflakeIO {
@Setup
public void setup() throws Exception {
- dataSource = dataSourceProviderFn.apply(null);
+ dataSourceProviderFn.apply(null);
DataSourceProviderFromDataSourceConfiguration
dataSourceProviderFromDataSourceConfiguration =
(DataSourceProviderFromDataSourceConfiguration)
this.dataSourceProviderFn;
diff --git
a/sdks/java/io/snowflake/src/main/java/org/apache/beam/sdk/io/snowflake/services/SnowflakeBatchServiceImpl.java
b/sdks/java/io/snowflake/src/main/java/org/apache/beam/sdk/io/snowflake/services/SnowflakeBatchServiceImpl.java
index 2927829..2700d8b 100644
---
a/sdks/java/io/snowflake/src/main/java/org/apache/beam/sdk/io/snowflake/services/SnowflakeBatchServiceImpl.java
+++
b/sdks/java/io/snowflake/src/main/java/org/apache/beam/sdk/io/snowflake/services/SnowflakeBatchServiceImpl.java
@@ -33,8 +33,6 @@ import
org.apache.beam.sdk.io.snowflake.data.SnowflakeTableSchema;
import org.apache.beam.sdk.io.snowflake.enums.CreateDisposition;
import org.apache.beam.sdk.io.snowflake.enums.WriteDisposition;
import org.apache.beam.sdk.transforms.SerializableFunction;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
/** Implemenation of {@link SnowflakeService} used in production. */
@SuppressWarnings({
@@ -42,7 +40,6 @@ import org.slf4j.LoggerFactory;
"nullness" // TODO(https://issues.apache.org/jira/browse/BEAM-10402)
})
public class SnowflakeBatchServiceImpl implements
SnowflakeService<SnowflakeBatchServiceConfig> {
- private static final Logger LOG =
LoggerFactory.getLogger(SnowflakeBatchServiceImpl.class);
private static final String SNOWFLAKE_GCS_PREFIX = "gcs://";
private static final String GCS_PREFIX = "gs://";
diff --git
a/sdks/java/io/snowflake/src/main/java/org/apache/beam/sdk/io/snowflake/services/SnowflakeStreamingServiceImpl.java
b/sdks/java/io/snowflake/src/main/java/org/apache/beam/sdk/io/snowflake/services/SnowflakeStreamingServiceImpl.java
index da0c328..2634a17 100644
---
a/sdks/java/io/snowflake/src/main/java/org/apache/beam/sdk/io/snowflake/services/SnowflakeStreamingServiceImpl.java
+++
b/sdks/java/io/snowflake/src/main/java/org/apache/beam/sdk/io/snowflake/services/SnowflakeStreamingServiceImpl.java
@@ -24,8 +24,6 @@ import java.util.Set;
import java.util.stream.Collectors;
import net.snowflake.ingest.SimpleIngestManager;
import net.snowflake.ingest.connection.IngestResponseException;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
/** Implemenation of {@link SnowflakeService} used in production. */
@SuppressWarnings({
@@ -33,7 +31,7 @@ import org.slf4j.LoggerFactory;
})
public class SnowflakeStreamingServiceImpl
implements SnowflakeService<SnowflakeStreamingServiceConfig> {
- private static final Logger LOG =
LoggerFactory.getLogger(SnowflakeStreamingServiceImpl.class);
+
private transient SimpleIngestManager ingestManager;
/** Writing data to Snowflake in streaming mode. */
diff --git
a/sdks/java/io/snowflake/src/test/java/org/apache/beam/sdk/io/snowflake/test/unit/DataSourceConfigurationTest.java
b/sdks/java/io/snowflake/src/test/java/org/apache/beam/sdk/io/snowflake/test/unit/DataSourceConfigurationTest.java
index 6bac967..3170206 100644
---
a/sdks/java/io/snowflake/src/test/java/org/apache/beam/sdk/io/snowflake/test/unit/DataSourceConfigurationTest.java
+++
b/sdks/java/io/snowflake/src/test/java/org/apache/beam/sdk/io/snowflake/test/unit/DataSourceConfigurationTest.java
@@ -194,8 +194,6 @@ public class DataSourceConfigurationTest {
.withServerName(SERVER_NAME)
.withKeyPairPathAuth(USERNAME, privateKeyPath, keyPassphrase);
- DataSource dataSource = configuration.buildDatasource();
-
assertEquals(USERNAME, configuration.getUsername().get());
// TODO assertEquals(privateKeyPath, configuration.getPrivateKeyPath());
assertEquals(keyPassphrase, configuration.getPrivateKeyPassphrase().get());
@@ -239,8 +237,6 @@ public class DataSourceConfigurationTest {
.withServerName(SERVER_NAME)
.withKeyPairRawAuth(USERNAME, rawPrivateKey, keyPassphrase);
- DataSource dataSource = configuration.buildDatasource();
-
assertEquals(USERNAME, configuration.getUsername().get());
assertEquals(rawPrivateKey, configuration.getRawPrivateKey().get());
assertEquals(keyPassphrase, configuration.getPrivateKeyPassphrase().get());
diff --git
a/sdks/java/io/snowflake/src/test/java/org/apache/beam/sdk/io/snowflake/test/unit/write/SchemaDispositionTest.java
b/sdks/java/io/snowflake/src/test/java/org/apache/beam/sdk/io/snowflake/test/unit/write/SchemaDispositionTest.java
index d150f44..89edc91 100644
---
a/sdks/java/io/snowflake/src/test/java/org/apache/beam/sdk/io/snowflake/test/unit/write/SchemaDispositionTest.java
+++
b/sdks/java/io/snowflake/src/test/java/org/apache/beam/sdk/io/snowflake/test/unit/write/SchemaDispositionTest.java
@@ -58,7 +58,6 @@ import org.junit.runners.JUnit4;
"rawtypes", // TODO(https://issues.apache.org/jira/browse/BEAM-10556)
})
public class SchemaDispositionTest {
- private static final String FAKE_TABLE = "FAKE_TABLE";
private static final String BUCKET_NAME = "BUCKET/";
@Rule public final transient TestPipeline pipeline = TestPipeline.create();
diff --git
a/sdks/java/io/snowflake/src/test/java/org/apache/beam/sdk/io/snowflake/test/unit/write/StreamingWriteTest.java
b/sdks/java/io/snowflake/src/test/java/org/apache/beam/sdk/io/snowflake/test/unit/write/StreamingWriteTest.java
index 0ee2928..7c68fc2 100644
---
a/sdks/java/io/snowflake/src/test/java/org/apache/beam/sdk/io/snowflake/test/unit/write/StreamingWriteTest.java
+++
b/sdks/java/io/snowflake/src/test/java/org/apache/beam/sdk/io/snowflake/test/unit/write/StreamingWriteTest.java
@@ -56,15 +56,13 @@ import org.junit.Test;
import org.junit.rules.ExpectedException;
import org.junit.runner.RunWith;
import org.junit.runners.JUnit4;
-import org.slf4j.Logger;
-import org.slf4j.LoggerFactory;
@RunWith(JUnit4.class)
@SuppressWarnings({
"rawtypes", // TODO(https://issues.apache.org/jira/browse/BEAM-10556)
})
public class StreamingWriteTest {
- private static final Logger LOG =
LoggerFactory.getLogger(StreamingWriteTest.class);
+
private static final String FAKE_TABLE = "TEST_TABLE";
private static final String STAGING_BUCKET_NAME = "BUCKET/";
private static final String STORAGE_INTEGRATION_NAME = "STORAGE_INTEGRATION";
diff --git a/sdks/java/io/solr/build.gradle b/sdks/java/io/solr/build.gradle
index a09191f..efdfde2 100644
--- a/sdks/java/io/solr/build.gradle
+++ b/sdks/java/io/solr/build.gradle
@@ -18,7 +18,6 @@
plugins { id 'org.apache.beam.module' }
applyJavaNature(
- suppressUnusedVariable: true,
automaticModuleName: 'org.apache.beam.sdk.io.solr',
)
diff --git a/sdks/java/io/splunk/build.gradle b/sdks/java/io/splunk/build.gradle
index 872793a..d8ea800 100644
--- a/sdks/java/io/splunk/build.gradle
+++ b/sdks/java/io/splunk/build.gradle
@@ -19,7 +19,7 @@
plugins {
id 'org.apache.beam.module'
}
-applyJavaNature(suppressUnusedVariable: true, automaticModuleName:
'org.apache.beam.sdk.io.splunk')
+applyJavaNature( automaticModuleName: 'org.apache.beam.sdk.io.splunk')
description = "Apache Beam :: SDKs :: Java :: IO :: Splunk"
ext.summary = "IO to write events to Splunk Http Event Collector (HEC)"
diff --git
a/sdks/java/io/splunk/src/main/java/org/apache/beam/sdk/io/splunk/SplunkEventWriter.java
b/sdks/java/io/splunk/src/main/java/org/apache/beam/sdk/io/splunk/SplunkEventWriter.java
index 7e41f15..51ded28 100644
---
a/sdks/java/io/splunk/src/main/java/org/apache/beam/sdk/io/splunk/SplunkEventWriter.java
+++
b/sdks/java/io/splunk/src/main/java/org/apache/beam/sdk/io/splunk/SplunkEventWriter.java
@@ -55,7 +55,8 @@ import org.slf4j.LoggerFactory;
/** A {@link DoFn} to write {@link SplunkEvent}s to Splunk's HEC endpoint. */
@AutoValue
@SuppressWarnings({
- "nullness" // TODO(https://issues.apache.org/jira/browse/BEAM-10402)
+ "nullness", // TODO(https://issues.apache.org/jira/browse/BEAM-10402)
+ "unused" // TODO(BEAM-13271): Remove when new version of errorprone is
released (2.11.0)
})
abstract class SplunkEventWriter extends DoFn<KV<Integer, SplunkEvent>,
SplunkWriteError> {
diff --git a/sdks/java/io/synthetic/build.gradle
b/sdks/java/io/synthetic/build.gradle
index bb428e6..65d58ce 100644
--- a/sdks/java/io/synthetic/build.gradle
+++ b/sdks/java/io/synthetic/build.gradle
@@ -18,7 +18,6 @@
plugins { id 'org.apache.beam.module' }
applyJavaNature(
- suppressUnusedVariable: true,
exportJavadoc: false,
automaticModuleName: 'org.apache.beam.sdk.io.synthetic',
)
diff --git a/sdks/java/io/thrift/build.gradle b/sdks/java/io/thrift/build.gradle
index 8e6f8dc..e012c7e 100644
--- a/sdks/java/io/thrift/build.gradle
+++ b/sdks/java/io/thrift/build.gradle
@@ -21,7 +21,6 @@ plugins {
// id "org.jruyi.thrift" version "0.4.1"
}
applyJavaNature(
- suppressUnusedVariable: true,
automaticModuleName: 'org.apache.beam.sdk.io.thrift',
generatedClassPatterns: [
/^org\.apache\.beam\.sdk\.io\.thrift\.payloads.*/,
diff --git
a/sdks/java/io/thrift/src/test/java/org/apache/beam/sdk/io/thrift/ThriftIOTest.java
b/sdks/java/io/thrift/src/test/java/org/apache/beam/sdk/io/thrift/ThriftIOTest.java
index afdc9c3..0b01fb3 100644
---
a/sdks/java/io/thrift/src/test/java/org/apache/beam/sdk/io/thrift/ThriftIOTest.java
+++
b/sdks/java/io/thrift/src/test/java/org/apache/beam/sdk/io/thrift/ThriftIOTest.java
@@ -36,7 +36,6 @@ import org.apache.thrift.protocol.TBinaryProtocol;
import org.apache.thrift.protocol.TCompactProtocol;
import org.apache.thrift.protocol.TJSONProtocol;
import org.apache.thrift.protocol.TProtocolFactory;
-import org.apache.thrift.protocol.TSimpleJSONProtocol;
import org.junit.Before;
import org.junit.Rule;
import org.junit.Test;
@@ -51,13 +50,12 @@ public class ThriftIOTest implements Serializable {
private static final String RESOURCE_DIR = "ThriftIOTest/";
private static final String THRIFT_DIR =
Resources.getResource(RESOURCE_DIR).getPath();
- private static final String ALL_THRIFT_STRING =
- Resources.getResource(RESOURCE_DIR).getPath() + "*";
+
private static final TestThriftStruct TEST_THRIFT_STRUCT = new
TestThriftStruct();
private static List<TestThriftStruct> testThriftStructs;
private final TProtocolFactory tBinaryProtoFactory = new
TBinaryProtocol.Factory();
private final TProtocolFactory tJsonProtocolFactory = new
TJSONProtocol.Factory();
- private final TProtocolFactory tSimpleJsonProtocolFactory = new
TSimpleJSONProtocol.Factory();
+
private final TProtocolFactory tCompactProtocolFactory = new
TCompactProtocol.Factory();
@Rule public transient TestPipeline mainPipeline = TestPipeline.create();
@Rule public transient TestPipeline readPipeline = TestPipeline.create();
diff --git a/sdks/java/io/tika/build.gradle b/sdks/java/io/tika/build.gradle
index 5b13e47..c6eab51 100644
--- a/sdks/java/io/tika/build.gradle
+++ b/sdks/java/io/tika/build.gradle
@@ -17,7 +17,7 @@
*/
plugins { id 'org.apache.beam.module' }
-applyJavaNature(suppressUnusedVariable: true, automaticModuleName:
'org.apache.beam.sdk.io.tika')
+applyJavaNature( automaticModuleName: 'org.apache.beam.sdk.io.tika')
description = "Apache Beam :: SDKs :: Java :: IO :: Tika"
ext.summary = "Tika Input to parse files."
diff --git a/sdks/java/io/xml/build.gradle b/sdks/java/io/xml/build.gradle
index 6548793..22e5b76 100644
--- a/sdks/java/io/xml/build.gradle
+++ b/sdks/java/io/xml/build.gradle
@@ -17,7 +17,7 @@
*/
plugins { id 'org.apache.beam.module' }
-applyJavaNature(suppressUnusedVariable: true, automaticModuleName:
'org.apache.beam.sdk.io.xml')
+applyJavaNature( automaticModuleName: 'org.apache.beam.sdk.io.xml')
description = "Apache Beam :: SDKs :: Java :: IO :: XML"
ext.summary = "IO to read and write XML files."