This is an automated email from the ASF dual-hosted git repository.
zichaowang pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/shardingsphere.git
The following commit(s) were added to refs/heads/master by this push:
new a8174aed869 Fix code inspection for kernel module (#31209)
a8174aed869 is described below
commit a8174aed86929ef9a8204a8d8672c92426fa4938
Author: Liang Zhang <[email protected]>
AuthorDate: Sun May 12 04:34:50 2024 +0800
Fix code inspection for kernel module (#31209)
---
.../core/channel/memory/MemoryPipelineChannel.java | 2 +-
.../ConsistencyCheckJobItemProgressContext.java | 2 +-
.../DataConsistencyCheckUtils.java | 2 +-
.../result/TableDataConsistencyCheckResult.java | 5 +--
.../table/MatchingTableInventoryChecker.java | 2 +-
.../importer/sink/type/PipelineDataSourceSink.java | 2 +-
.../type/pk/type/UnsupportedKeyIngestPosition.java | 2 +-
.../pipeline/core/job/id/PipelineJobIdUtils.java | 2 +-
.../job/progress/TransmissionJobItemProgress.java | 2 +-
.../persist/PipelineJobProgressPersistContext.java | 2 +-
.../persist/PipelineJobProgressPersistService.java | 6 +--
.../YamlJobItemInventoryTasksProgressSwapper.java | 2 +-
.../core/job/service/TransmissionJobManager.java | 4 +-
.../StandardPipelineTableMetaDataLoader.java | 2 +-
.../datasource/PipelineJobDataSourcePreparer.java | 2 +-
.../inventory/InventoryRecordsCountCalculator.java | 2 +-
.../preparer/inventory/InventoryTaskSplitter.java | 6 +--
.../job/PipelineJobCheckGovernanceRepository.java | 2 +-
.../sqlbuilder/sql/PipelineImportSQLBuilder.java | 2 +-
.../core/util/IntervalToRangeIterator.java | 8 ++--
.../core/util/ShardingColumnsExtractor.java | 2 +-
.../core/channel/MultiplexPipelineChannelTest.java | 2 +-
.../ConsistencyCheckDataBuilder.java | 8 ++--
.../DataConsistencyCheckUtilsTest.java | 4 +-
.../CRC32SingleTableInventoryCalculatorTest.java | 2 +-
.../datanode/JobDataNodeLineConvertUtilsTest.java | 6 +--
.../type/IntegerPrimaryKeyIngestPositionTest.java | 2 +-
.../progress/PipelineJobProgressDetectorTest.java | 50 ++++++----------------
.../progress/TransmissionJobItemProgressTest.java | 4 +-
.../metadata/node/PipelineMetaDataNodeTest.java | 3 +-
.../mysql/ingest/GlobalTableMapEventMapping.java | 2 +-
.../pipeline/mysql/ingest/client/MySQLClient.java | 2 +-
.../value/impl/MySQLUnsignedBigintHandler.java | 2 +-
.../ingest/MySQLIngestPositionManagerTest.java | 2 +-
.../mysql/ingest/binlog/BinlogPositionTest.java | 2 +-
.../mysql/ingest/client/MySQLClientTest.java | 2 +-
.../opengauss/ingest/OpenGaussWALDumper.java | 2 +-
.../ddlgenerator/AbstractPostgreSQLDDLAdapter.java | 12 +++---
.../PostgreSQLColumnPropertiesAppender.java | 2 +-
.../ddlgenerator/PostgreSQLColumnType.java | 2 +-
.../postgresql/ingest/PostgreSQLWALDumper.java | 4 +-
.../ingest/wal/WALEventConverterTest.java | 2 +-
.../query/ShowStreamingJobStatusExecutor.java | 2 +-
.../query/ShowMigrationJobStatusExecutor.java | 2 +-
.../cdc/client/util/ProtobufAnyValueConverter.java | 2 +-
.../data/pipeline/cdc/CDCJobType.java | 2 +-
.../data/pipeline/cdc/api/CDCJobAPI.java | 2 +-
.../pipeline/cdc/context/CDCJobItemContext.java | 4 +-
.../pipeline/cdc/core/importer/CDCImporter.java | 2 +-
.../pipeline/cdc/handler/CDCBackendHandler.java | 2 +-
.../data/pipeline/cdc/util/CDCDataNodeUtils.java | 2 +-
.../pipeline/cdc/util/CDCSchemaTableUtils.java | 6 +--
.../pipeline/cdc/util/ColumnValueConvertUtils.java | 8 ++--
.../pipeline/cdc/util/CDCSchemaTableUtilsTest.java | 7 ++-
.../cdc/util/ColumnValueConvertUtilsTest.java | 2 +-
.../cdc/util/DataRecordResultConvertUtilsTest.java | 2 +-
.../api/ConsistencyCheckJobAPI.java | 8 ++--
.../scenario/migration/api/MigrationJobAPI.java | 6 +--
.../migration/context/MigrationJobItemContext.java | 6 +--
.../migration/preparer/MigrationJobPreparer.java | 2 +-
.../type/tso/provider/RedisTSOProvider.java | 2 +-
.../shardingsphere/logging/util/LoggingUtils.java | 2 +-
.../database/DatabaseMetaDataPersistService.java | 4 +-
.../single/datanode/SingleTableDataNodeLoader.java | 4 +-
.../single/decider/SingleSQLFederationDecider.java | 2 +-
.../SingleRuleConfigurationDecorator.java | 8 ++--
.../single/util/SingleTableLoadUtils.java | 2 +-
.../config/pojo/YamlSingleRuleConfiguration.java | 2 +-
.../handler/update/LoadSingleTableExecutor.java | 2 +-
.../resultset/SQLFederationResultSetMetaData.java | 2 +-
.../core/SQLFederationDistSQLStatementVisitor.java | 2 +-
.../enumerable/EnumerableScanExecutor.java | 2 +-
.../executor/enumerator/JDBCRowEnumerator.java | 2 +-
.../expression/impl/RowExpressionConverter.java | 2 +-
.../optimizer/operator/logical/LogicalScan.java | 2 +-
.../planner/util/SQLFederationPlannerUtils.java | 2 +-
.../optimizer/it/SQLStatementCompilerIT.java | 2 +-
.../parser/rule/SQLParserRuleTest.java | 2 +-
.../rule/builder/SQLParserRuleBuilderTest.java | 6 +--
...LParserCacheOptionConfigurationSwapperTest.java | 2 +-
.../YamlSQLParserRuleConfigurationSwapperTest.java | 2 +-
.../query/ShowSQLParserRuleExecutorTest.java | 2 +-
.../core/SQLParserDistSQLStatementVisitor.java | 2 +-
.../transaction/ConnectionTransaction.java | 4 +-
.../transaction/xa/fixture/DataSourceUtils.java | 4 +-
.../manager/switcher/ResourceSwitchManager.java | 2 +-
.../mode/metadata/MetaDataContexts.java | 2 +-
.../manager/cluster/ClusterModeContextManager.java | 2 +-
.../compute/service/ComputeNodeStatusService.java | 2 +-
.../ResourceMetaDataChangedSubscriberTest.java | 2 +-
.../subscriber/StateChangedSubscriberTest.java | 2 +-
.../cluster/lock/holder/DistributedLockHolder.java | 3 +-
.../listener/SessionConnectionListener.java | 2 +-
.../standalone/StandaloneModeContextManager.java | 2 +-
94 files changed, 151 insertions(+), 181 deletions(-)
diff --git
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/channel/memory/MemoryPipelineChannel.java
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/channel/memory/MemoryPipelineChannel.java
index a2a6212fc0e..7061d515ad2 100644
---
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/channel/memory/MemoryPipelineChannel.java
+++
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/channel/memory/MemoryPipelineChannel.java
@@ -57,7 +57,7 @@ public final class MemoryPipelineChannel implements
PipelineChannel {
long startMillis = System.currentTimeMillis();
int recordsCount = 0;
do {
- List<Record> records = queue.poll(Math.max(0, timeoutMillis -
(System.currentTimeMillis() - startMillis)), TimeUnit.MILLISECONDS);
+ List<Record> records = queue.poll(Math.max(0L, timeoutMillis -
(System.currentTimeMillis() - startMillis)), TimeUnit.MILLISECONDS);
if (null == records || records.isEmpty()) {
continue;
}
diff --git
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/ConsistencyCheckJobItemProgressContext.java
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/ConsistencyCheckJobItemProgressContext.java
index c493dc9afb9..dba540c9f70 100644
---
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/ConsistencyCheckJobItemProgressContext.java
+++
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/ConsistencyCheckJobItemProgressContext.java
@@ -48,7 +48,7 @@ public final class ConsistencyCheckJobItemProgressContext
implements PipelineJob
private volatile long recordsCount;
- private final AtomicLong checkedRecordsCount = new AtomicLong(0);
+ private final AtomicLong checkedRecordsCount = new AtomicLong(0L);
private final long checkBeginTimeMillis = System.currentTimeMillis();
diff --git
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/DataConsistencyCheckUtils.java
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/DataConsistencyCheckUtils.java
index 0aa0844db66..f78eb54f972 100644
---
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/DataConsistencyCheckUtils.java
+++
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/DataConsistencyCheckUtils.java
@@ -126,6 +126,6 @@ public final class DataConsistencyCheckUtils {
decimalTwo = another;
}
}
- return decimalOne.equals(decimalTwo);
+ return 0 == decimalOne.compareTo(decimalTwo);
}
}
diff --git
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/result/TableDataConsistencyCheckResult.java
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/result/TableDataConsistencyCheckResult.java
index 8d73495befa..ebf462baea7 100644
---
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/result/TableDataConsistencyCheckResult.java
+++
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/result/TableDataConsistencyCheckResult.java
@@ -56,9 +56,6 @@ public final class TableDataConsistencyCheckResult {
* @return matched or not
*/
public boolean isMatched() {
- if (null != ignoredType) {
- return false;
- }
- return matched;
+ return null == ignoredType && matched;
}
}
diff --git
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/table/MatchingTableInventoryChecker.java
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/table/MatchingTableInventoryChecker.java
index abd1e1ecc04..f9faf20e2ce 100644
---
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/table/MatchingTableInventoryChecker.java
+++
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/table/MatchingTableInventoryChecker.java
@@ -62,7 +62,7 @@ public abstract class MatchingTableInventoryChecker
implements TableInventoryChe
@Override
public TableDataConsistencyCheckResult checkSingleTableInventoryData() {
ThreadFactory threadFactory =
ExecutorThreadFactoryBuilder.build("job-" + getJobIdDigest(param.getJobId()) +
"-matching-check-%d");
- ThreadPoolExecutor executor = new ThreadPoolExecutor(2, 2, 60,
TimeUnit.SECONDS, new ArrayBlockingQueue<>(2), threadFactory);
+ ThreadPoolExecutor executor = new ThreadPoolExecutor(2, 2, 60L,
TimeUnit.SECONDS, new ArrayBlockingQueue<>(2), threadFactory);
try {
return checkSingleTableInventoryData(param, executor);
} finally {
diff --git
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/importer/sink/type/PipelineDataSourceSink.java
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/importer/sink/type/PipelineDataSourceSink.java
index 1cabadec3a8..8fdf6cfb94d 100644
---
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/importer/sink/type/PipelineDataSourceSink.java
+++
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/importer/sink/type/PipelineDataSourceSink.java
@@ -101,7 +101,7 @@ public final class PipelineDataSourceSink implements
PipelineSink {
if (i == importerConfig.getRetryTimes()) {
throw new PipelineImporterJobWriteException(ex);
}
- Thread.sleep(Math.min(5 * 60 * 1000L, 1000L << i));
+ Thread.sleep(Math.min(5L * 60L * 1000L, 1000L << i));
}
}
}
diff --git
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/ingest/position/type/pk/type/UnsupportedKeyIngestPosition.java
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/ingest/position/type/pk/type/UnsupportedKeyIngestPosition.java
index e589193eb45..d8563ac1e41 100644
---
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/ingest/position/type/pk/type/UnsupportedKeyIngestPosition.java
+++
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/ingest/position/type/pk/type/UnsupportedKeyIngestPosition.java
@@ -36,7 +36,7 @@ public final class UnsupportedKeyIngestPosition implements
PrimaryKeyIngestPosit
@Override
public Void convert(final String value) {
- throw new UnsupportedOperationException();
+ throw new UnsupportedOperationException("");
}
@Override
diff --git
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/id/PipelineJobIdUtils.java
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/id/PipelineJobIdUtils.java
index 8972531cca2..459193309f3 100644
---
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/id/PipelineJobIdUtils.java
+++
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/id/PipelineJobIdUtils.java
@@ -101,7 +101,7 @@ public final class PipelineJobIdUtils {
* @return ElasticJob configuration POJO
*/
public static JobConfigurationPOJO getElasticJobConfigurationPOJO(final
String jobId) {
- JobConfigurationPOJO result =
PipelineAPIFactory.getJobConfigurationAPI(PipelineJobIdUtils.parseContextKey(jobId)).getJobConfiguration(jobId);
+ JobConfigurationPOJO result =
PipelineAPIFactory.getJobConfigurationAPI(parseContextKey(jobId)).getJobConfiguration(jobId);
ShardingSpherePreconditions.checkNotNull(result, () -> new
PipelineJobNotFoundException(jobId));
return result;
}
diff --git
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/TransmissionJobItemProgress.java
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/TransmissionJobItemProgress.java
index 8d38f79fe0b..56c18fa144c 100644
---
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/TransmissionJobItemProgress.java
+++
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/TransmissionJobItemProgress.java
@@ -70,7 +70,7 @@ public final class TransmissionJobItemProgress implements
PipelineJobItemProgres
}
private JobItemInventoryTasksProgress getInventoryTasksProgress(final
Collection<PipelineTask> inventoryTasks) {
- Map<String, InventoryTaskProgress> inventoryTaskProgressMap = new
HashMap<>();
+ Map<String, InventoryTaskProgress> inventoryTaskProgressMap = new
HashMap<>(inventoryTasks.size(), 1F);
for (PipelineTask each : inventoryTasks) {
inventoryTaskProgressMap.put(each.getTaskId(),
(InventoryTaskProgress) each.getTaskProgress());
}
diff --git
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/persist/PipelineJobProgressPersistContext.java
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/persist/PipelineJobProgressPersistContext.java
index beafba57ba3..839950933e0 100644
---
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/persist/PipelineJobProgressPersistContext.java
+++
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/persist/PipelineJobProgressPersistContext.java
@@ -34,7 +34,7 @@ public final class PipelineJobProgressPersistContext {
private final int shardingItem;
- private final AtomicLong unhandledEventCount = new AtomicLong(0);
+ private final AtomicLong unhandledEventCount = new AtomicLong(0L);
private final AtomicBoolean firstExceptionLogged = new
AtomicBoolean(false);
}
diff --git
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/persist/PipelineJobProgressPersistService.java
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/persist/PipelineJobProgressPersistService.java
index e6bc68e0547..461efafcc9c 100644
---
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/persist/PipelineJobProgressPersistService.java
+++
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/persist/PipelineJobProgressPersistService.java
@@ -53,7 +53,7 @@ public final class PipelineJobProgressPersistService {
private static final long DELAY_SECONDS = 1L;
static {
- JOB_PERSIST_EXECUTOR.scheduleWithFixedDelay(new
PersistJobContextRunnable(), 0, DELAY_SECONDS, TimeUnit.SECONDS);
+ JOB_PERSIST_EXECUTOR.scheduleWithFixedDelay(new
PersistJobContextRunnable(), 0L, DELAY_SECONDS, TimeUnit.SECONDS);
}
/**
@@ -126,8 +126,8 @@ public final class PipelineJobProgressPersistService {
private static void persist0(final String jobId, final int
shardingItem, final PipelineJobProgressPersistContext persistContext) {
long currentUnhandledEventCount =
persistContext.getUnhandledEventCount().get();
- ShardingSpherePreconditions.checkState(currentUnhandledEventCount
>= 0, () -> new IllegalStateException("Current unhandled event count must be
greater than or equal to 0"));
- if (0 == currentUnhandledEventCount) {
+ ShardingSpherePreconditions.checkState(currentUnhandledEventCount
>= 0L, () -> new IllegalStateException("Current unhandled event count must be
greater than or equal to 0"));
+ if (0L == currentUnhandledEventCount) {
return;
}
Optional<PipelineJobItemContext> jobItemContext =
PipelineJobRegistry.getItemContext(jobId, shardingItem);
diff --git
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/yaml/swapper/YamlJobItemInventoryTasksProgressSwapper.java
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/yaml/swapper/YamlJobItemInventoryTasksProgressSwapper.java
index 91486949172..e3c464b8f66 100644
---
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/yaml/swapper/YamlJobItemInventoryTasksProgressSwapper.java
+++
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/progress/yaml/swapper/YamlJobItemInventoryTasksProgressSwapper.java
@@ -75,7 +75,7 @@ public final class YamlJobItemInventoryTasksProgressSwapper {
if (null == yamlProgress) {
return new JobItemInventoryTasksProgress(Collections.emptyMap());
}
- Map<String, InventoryTaskProgress> taskProgressMap = new
LinkedHashMap<>();
+ Map<String, InventoryTaskProgress> taskProgressMap = new
LinkedHashMap<>(yamlProgress.getFinished().length +
yamlProgress.getUnfinished().size(), 1F);
taskProgressMap.putAll(Arrays.stream(yamlProgress.getFinished()).collect(Collectors.toMap(key
-> key, value -> new InventoryTaskProgress(new IngestFinishedPosition()))));
taskProgressMap.putAll(yamlProgress.getUnfinished().entrySet().stream().collect(Collectors.toMap(Entry::getKey,
getInventoryTaskProgressFunction())));
return new JobItemInventoryTasksProgress(taskProgressMap);
diff --git
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/service/TransmissionJobManager.java
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/service/TransmissionJobManager.java
index 9dcd2cd4676..df81d4312cd 100644
---
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/service/TransmissionJobManager.java
+++
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/job/service/TransmissionJobManager.java
@@ -81,8 +81,8 @@ public final class TransmissionJobManager {
if (JobStatus.EXECUTE_INCREMENTAL_TASK == jobItemProgress.getStatus()
|| JobStatus.FINISHED == jobItemProgress.getStatus()) {
return 100;
}
- if (0 != jobItemProgress.getProcessedRecordsCount() && 0 !=
jobItemProgress.getInventoryRecordsCount()) {
- return (int) Math.min(100,
jobItemProgress.getProcessedRecordsCount() * 100 /
jobItemProgress.getInventoryRecordsCount());
+ if (0L != jobItemProgress.getProcessedRecordsCount() && 0L !=
jobItemProgress.getInventoryRecordsCount()) {
+ return (int) Math.min(100L,
jobItemProgress.getProcessedRecordsCount() * 100L /
jobItemProgress.getInventoryRecordsCount());
}
return 0;
}
diff --git
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/metadata/loader/StandardPipelineTableMetaDataLoader.java
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/metadata/loader/StandardPipelineTableMetaDataLoader.java
index 659777adb18..ac97df73d35 100644
---
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/metadata/loader/StandardPipelineTableMetaDataLoader.java
+++
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/metadata/loader/StandardPipelineTableMetaDataLoader.java
@@ -89,7 +89,7 @@ public final class StandardPipelineTableMetaDataLoader
implements PipelineTableM
tableNames.add(tableName);
}
}
- Map<CaseInsensitiveIdentifier, PipelineTableMetaData> result = new
LinkedHashMap<>();
+ Map<CaseInsensitiveIdentifier, PipelineTableMetaData> result = new
LinkedHashMap<>(tableNames.size(), 1F);
for (String each : tableNames) {
Set<String> primaryKeys = loadPrimaryKeys(connection, schemaName,
each);
Map<String, Collection<String>> uniqueKeys =
loadUniqueIndexesOfTable(connection, schemaName, each);
diff --git
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/preparer/datasource/PipelineJobDataSourcePreparer.java
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/preparer/datasource/PipelineJobDataSourcePreparer.java
index b62427ccfea..20f71add9c0 100644
---
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/preparer/datasource/PipelineJobDataSourcePreparer.java
+++
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/preparer/datasource/PipelineJobDataSourcePreparer.java
@@ -69,7 +69,7 @@ public final class PipelineJobDataSourcePreparer {
}
String defaultSchema =
dialectDatabaseMetaData.getDefaultSchema().orElse(null);
PipelinePrepareSQLBuilder pipelineSQLBuilder = new
PipelinePrepareSQLBuilder(targetDatabaseType);
- Collection<String> createdSchemaNames = new HashSet<>();
+ Collection<String> createdSchemaNames = new
HashSet<>(param.getCreateTableConfigurations().size(), 1F);
for (CreateTableConfiguration each :
param.getCreateTableConfigurations()) {
String targetSchemaName =
each.getTargetName().getSchemaName().toString();
if (null == targetSchemaName ||
targetSchemaName.equalsIgnoreCase(defaultSchema) ||
createdSchemaNames.contains(targetSchemaName)) {
diff --git
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/preparer/inventory/InventoryRecordsCountCalculator.java
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/preparer/inventory/InventoryRecordsCountCalculator.java
index 86c461ad110..0544ce606e5 100644
---
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/preparer/inventory/InventoryRecordsCountCalculator.java
+++
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/preparer/inventory/InventoryRecordsCountCalculator.java
@@ -60,7 +60,7 @@ public final class InventoryRecordsCountCalculator {
if (sql.isPresent()) {
DatabaseType databaseType =
TypedSPILoader.getService(DatabaseType.class,
dataSource.getDatabaseType().getType());
long result = getEstimatedCount(databaseType, dataSource,
sql.get());
- return result > 0 ? result : getCount(dataSource,
pipelineSQLBuilder.buildCountSQL(schemaName, actualTableName));
+ return result > 0L ? result : getCount(dataSource,
pipelineSQLBuilder.buildCountSQL(schemaName, actualTableName));
}
return getCount(dataSource,
pipelineSQLBuilder.buildCountSQL(schemaName, actualTableName));
} catch (final SQLException ex) {
diff --git
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/preparer/inventory/InventoryTaskSplitter.java
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/preparer/inventory/InventoryTaskSplitter.java
index cfe3c9e87d6..88a2f5536a5 100644
---
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/preparer/inventory/InventoryTaskSplitter.java
+++
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/preparer/inventory/InventoryTaskSplitter.java
@@ -183,13 +183,13 @@ public final class InventoryTaskSplitter {
private Collection<IngestPosition>
getPositionByIntegerUniqueKeyRange(final InventoryDumperContext dumperContext,
final long tableRecordsCount,
final TransmissionJobItemContext jobItemContext, final
PipelineDataSourceWrapper dataSource) {
- if (0 == tableRecordsCount) {
- return Collections.singletonList(new
IntegerPrimaryKeyIngestPosition(0, 0));
+ if (0L == tableRecordsCount) {
+ return Collections.singletonList(new
IntegerPrimaryKeyIngestPosition(0L, 0L));
}
Collection<IngestPosition> result = new LinkedList<>();
Range<Long> uniqueKeyValuesRange =
getUniqueKeyValuesRange(jobItemContext, dataSource, dumperContext);
int shardingSize =
jobItemContext.getJobProcessContext().getProcessConfiguration().getRead().getShardingSize();
- long splitCount = tableRecordsCount / shardingSize +
(tableRecordsCount % shardingSize > 0 ? 1 : 0);
+ long splitCount = tableRecordsCount / shardingSize +
(tableRecordsCount % shardingSize > 0L ? 1 : 0);
long interval = (uniqueKeyValuesRange.getMaximum() -
uniqueKeyValuesRange.getMinimum()) / splitCount;
IntervalToRangeIterator rangeIterator = new
IntervalToRangeIterator(uniqueKeyValuesRange.getMinimum(),
uniqueKeyValuesRange.getMaximum(), interval);
while (rangeIterator.hasNext()) {
diff --git
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/registrycenter/repository/job/PipelineJobCheckGovernanceRepository.java
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/registrycenter/repository/job/PipelineJobCheckGovernanceRepository.java
index 4875f179837..acaf3779139 100644
---
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/registrycenter/repository/job/PipelineJobCheckGovernanceRepository.java
+++
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/registrycenter/repository/job/PipelineJobCheckGovernanceRepository.java
@@ -125,7 +125,7 @@ public final class PipelineJobCheckGovernanceRepository {
if (null == checkResultMap) {
return;
}
- Map<String, String> yamlCheckResultMap = new LinkedHashMap<>();
+ Map<String, String> yamlCheckResultMap = new
LinkedHashMap<>(checkResultMap.size(), 1F);
for (Entry<String, TableDataConsistencyCheckResult> entry :
checkResultMap.entrySet()) {
YamlTableDataConsistencyCheckResult yamlCheckResult = new
YamlTableDataConsistencyCheckResultSwapper().swapToYamlConfiguration(entry.getValue());
yamlCheckResultMap.put(entry.getKey(),
YamlEngine.marshal(yamlCheckResult));
diff --git
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/sqlbuilder/sql/PipelineImportSQLBuilder.java
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/sqlbuilder/sql/PipelineImportSQLBuilder.java
index 735d6d266d3..d7c7e7861de 100644
---
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/sqlbuilder/sql/PipelineImportSQLBuilder.java
+++
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/sqlbuilder/sql/PipelineImportSQLBuilder.java
@@ -51,7 +51,7 @@ public final class PipelineImportSQLBuilder {
public PipelineImportSQLBuilder(final DatabaseType databaseType) {
dialectSQLBuilder =
DatabaseTypedSPILoader.getService(DialectPipelineSQLBuilder.class,
databaseType);
sqlSegmentBuilder = new PipelineSQLSegmentBuilder(databaseType);
- sqlCache =
Caffeine.newBuilder().initialCapacity(16).maximumSize(1024).build();
+ sqlCache =
Caffeine.newBuilder().initialCapacity(16).maximumSize(1024L).build();
}
/**
diff --git
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/util/IntervalToRangeIterator.java
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/util/IntervalToRangeIterator.java
index 2ac010aaaa7..ab8bbbfb7fa 100644
---
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/util/IntervalToRangeIterator.java
+++
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/util/IntervalToRangeIterator.java
@@ -40,12 +40,12 @@ public final class IntervalToRangeIterator implements
Iterator<Range<Long>> {
if (minimum > maximum) {
throw new IllegalArgumentException("minimum greater than maximum");
}
- if (interval < 0) {
+ if (interval < 0L) {
throw new IllegalArgumentException("interval is less than zero");
}
this.maximum = maximum;
this.interval = interval;
- this.current = minimum;
+ current = minimum;
}
@Override
@@ -56,11 +56,11 @@ public final class IntervalToRangeIterator implements
Iterator<Range<Long>> {
@Override
public Range<Long> next() {
if (!hasNext()) {
- throw new NoSuchElementException();
+ throw new NoSuchElementException("");
}
long upperLimit = Math.min(maximum, current + interval);
Range<Long> result = Range.between(current, upperLimit);
- current = upperLimit + 1;
+ current = upperLimit + 1L;
return result;
}
}
diff --git
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/util/ShardingColumnsExtractor.java
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/util/ShardingColumnsExtractor.java
index 0d1773b5fd6..276b8581c70 100644
---
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/util/ShardingColumnsExtractor.java
+++
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/util/ShardingColumnsExtractor.java
@@ -55,7 +55,7 @@ public final class ShardingColumnsExtractor {
}
Set<String> defaultDatabaseShardingColumns =
extractShardingColumns(shardingRuleConfig.get().getDefaultDatabaseShardingStrategy());
Set<String> defaultTableShardingColumns =
extractShardingColumns(shardingRuleConfig.get().getDefaultTableShardingStrategy());
- Map<CaseInsensitiveIdentifier, Set<String>> result = new
ConcurrentHashMap<>();
+ Map<CaseInsensitiveIdentifier, Set<String>> result = new
ConcurrentHashMap<>(shardingRuleConfig.get().getTables().size(), 1F);
for (ShardingTableRuleConfiguration each :
shardingRuleConfig.get().getTables()) {
CaseInsensitiveIdentifier logicTableName = new
CaseInsensitiveIdentifier(each.getLogicTable());
if (!logicTableNames.contains(logicTableName)) {
diff --git
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/channel/MultiplexPipelineChannelTest.java
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/channel/MultiplexPipelineChannelTest.java
index 22e1292d10e..1d7f852b89b 100644
---
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/channel/MultiplexPipelineChannelTest.java
+++
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/channel/MultiplexPipelineChannelTest.java
@@ -84,7 +84,7 @@ class MultiplexPipelineChannelTest {
MultiplexPipelineChannel channel = new
MultiplexPipelineChannel(CHANNEL_NUMBER,
TypedSPILoader.getService(PipelineChannelCreator.class, "MEMORY"), 10000,
ackCallback);
fetchWithMultiThreads(channel, countDownLatch);
channel.push(Arrays.asList(records));
- boolean awaitResult = countDownLatch.await(10, TimeUnit.SECONDS);
+ boolean awaitResult = countDownLatch.await(10L, TimeUnit.SECONDS);
assertTrue(awaitResult, "await failed");
}
diff --git
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/ConsistencyCheckDataBuilder.java
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/ConsistencyCheckDataBuilder.java
index 9284bdff825..8174bc01255 100644
---
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/ConsistencyCheckDataBuilder.java
+++
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/ConsistencyCheckDataBuilder.java
@@ -33,7 +33,7 @@ public final class ConsistencyCheckDataBuilder {
* @return built record
*/
public static Map<String, Object> buildFixedFullTypeRecord(final int id) {
- Map<String, Object> result = new LinkedHashMap<>();
+ Map<String, Object> result = new LinkedHashMap<>(15, 1F);
result.put("id", id);
result.put("c_bool", true);
result.put("c_int1", Byte.MAX_VALUE);
@@ -97,13 +97,13 @@ public final class ConsistencyCheckDataBuilder {
return value + "-";
}
if (value instanceof Time) {
- return new Time(((Time) value).getTime() - 1);
+ return new Time(((Time) value).getTime() - 1L);
}
if (value instanceof Date) {
- return new Date(((Date) value).getTime() - 1);
+ return new Date(((Date) value).getTime() - 1L);
}
if (value instanceof Timestamp) {
- return new Timestamp(((Timestamp) value).getTime() - 1);
+ return new Timestamp(((Timestamp) value).getTime() - 1L);
}
if (value instanceof int[]) {
int[] result = ((int[]) value).clone();
diff --git
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/DataConsistencyCheckUtilsTest.java
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/DataConsistencyCheckUtilsTest.java
index 01676117059..7e2de6daa18 100644
---
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/DataConsistencyCheckUtilsTest.java
+++
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/DataConsistencyCheckUtilsTest.java
@@ -38,8 +38,8 @@ class DataConsistencyCheckUtilsTest {
@Test
void assertIsBigDecimalEquals() {
- BigDecimal one = BigDecimal.valueOf(3322, 1);
- BigDecimal another = BigDecimal.valueOf(33220, 2);
+ BigDecimal one = BigDecimal.valueOf(3322L, 1);
+ BigDecimal another = BigDecimal.valueOf(33220L, 2);
assertTrue(DataConsistencyCheckUtils.isBigDecimalEquals(one, another));
}
}
diff --git
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/table/calculator/CRC32SingleTableInventoryCalculatorTest.java
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/table/calculator/CRC32SingleTableInventoryCalculatorTest.java
index 04635829c2e..cf5b661cb60 100644
---
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/table/calculator/CRC32SingleTableInventoryCalculatorTest.java
+++
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/consistencycheck/table/calculator/CRC32SingleTableInventoryCalculatorTest.java
@@ -93,7 +93,7 @@ class CRC32SingleTableInventoryCalculatorTest {
@Test
void assertCalculateFailed() throws SQLException {
- when(connection.prepareStatement(anyString())).thenThrow(new
SQLException());
+ when(connection.prepareStatement(anyString())).thenThrow(new
SQLException(""));
assertThrows(PipelineTableDataConsistencyCheckLoadingFailedException.class, ()
-> new CRC32SingleTableInventoryCalculator().calculate(parameter));
}
}
diff --git
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/datanode/JobDataNodeLineConvertUtilsTest.java
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/datanode/JobDataNodeLineConvertUtilsTest.java
index 33750199d46..2ae0b5792bb 100644
---
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/datanode/JobDataNodeLineConvertUtilsTest.java
+++
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/datanode/JobDataNodeLineConvertUtilsTest.java
@@ -35,7 +35,7 @@ class JobDataNodeLineConvertUtilsTest {
@Test
void assertConvertDataNodesToLines() {
- Map<String, List<DataNode>> mockDataNodes = new LinkedHashMap<>();
+ Map<String, List<DataNode>> mockDataNodes = new LinkedHashMap<>(2, 1F);
List<DataNode> dataNodes = Arrays.asList(new DataNode("ds_0",
"t_order_0"), new DataNode("ds_0", "t_order_1"));
List<DataNode> itemDataNodes = Collections.singletonList(new
DataNode("ds_0", "t_order_item_0"));
mockDataNodes.put("t_order", dataNodes);
@@ -51,10 +51,8 @@ class JobDataNodeLineConvertUtilsTest {
@Test
void assertConvertDataNodesToLinesWithMultipleDataSource() {
- Map<String, List<DataNode>> mockDataNodes = new LinkedHashMap<>();
List<DataNode> dataNodes = Arrays.asList(new DataNode("ds_0",
"t_order_0"), new DataNode("ds_0", "t_order_2"), new DataNode("ds_1",
"t_order_1"), new DataNode("ds_1", "t_order_3"));
- mockDataNodes.put("t_order", dataNodes);
- List<JobDataNodeLine> jobDataNodeLines =
JobDataNodeLineConvertUtils.convertDataNodesToLines(mockDataNodes);
+ List<JobDataNodeLine> jobDataNodeLines =
JobDataNodeLineConvertUtils.convertDataNodesToLines(Collections.singletonMap("t_order",
dataNodes));
assertThat(jobDataNodeLines.size(), is(2));
JobDataNodeEntry jobDataNodeEntry =
jobDataNodeLines.get(0).getEntries().iterator().next();
assertThat(jobDataNodeEntry.getDataNodes().stream().map(DataNode::getTableName).collect(Collectors.toList()),
is(Arrays.asList("t_order_0", "t_order_2")));
diff --git
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/ingest/position/type/pk/type/IntegerPrimaryKeyIngestPositionTest.java
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/ingest/position/type/pk/type/IntegerPrimaryKeyIngestPositionTest.java
index 327d472ea9f..9153d853bd2 100644
---
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/ingest/position/type/pk/type/IntegerPrimaryKeyIngestPositionTest.java
+++
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/ingest/position/type/pk/type/IntegerPrimaryKeyIngestPositionTest.java
@@ -34,6 +34,6 @@ class IntegerPrimaryKeyIngestPositionTest {
@Test
void assertToString() {
- assertThat(new IntegerPrimaryKeyIngestPosition(1, 100).toString(),
is("i,1,100"));
+ assertThat(new IntegerPrimaryKeyIngestPosition(1L, 100L).toString(),
is("i,1,100"));
}
}
diff --git
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/job/progress/PipelineJobProgressDetectorTest.java
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/job/progress/PipelineJobProgressDetectorTest.java
index fdbe2d2a88d..9a3bfdd5971 100644
---
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/job/progress/PipelineJobProgressDetectorTest.java
+++
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/job/progress/PipelineJobProgressDetectorTest.java
@@ -24,15 +24,11 @@ import
org.apache.shardingsphere.data.pipeline.core.ingest.position.IngestPositi
import
org.apache.shardingsphere.data.pipeline.core.ingest.position.type.finished.IngestFinishedPosition;
import
org.apache.shardingsphere.data.pipeline.core.ingest.position.type.placeholder.IngestPlaceholderPosition;
import org.apache.shardingsphere.data.pipeline.core.task.InventoryTask;
-import org.apache.shardingsphere.data.pipeline.core.task.PipelineTask;
import
org.apache.shardingsphere.data.pipeline.core.task.progress.InventoryTaskProgress;
import org.junit.jupiter.api.Test;
-import java.util.ArrayList;
-import java.util.Collection;
-import java.util.HashMap;
-import java.util.List;
-import java.util.Map;
+import java.util.Arrays;
+import java.util.Collections;
import java.util.concurrent.atomic.AtomicReference;
import static org.junit.jupiter.api.Assertions.assertFalse;
@@ -43,7 +39,7 @@ class PipelineJobProgressDetectorTest {
@Test
void assertAllInventoryTasksAreFinishedWhenCollectionIsEmpty() {
- assertTrue(PipelineJobProgressDetector.isAllInventoryTasksFinished(new
ArrayList<>()));
+
assertTrue(PipelineJobProgressDetector.isAllInventoryTasksFinished(Collections.emptyList()));
}
@Test
@@ -52,10 +48,7 @@ class PipelineJobProgressDetectorTest {
AtomicReference<IngestPosition> unfinishedPosition = new
AtomicReference<>(new IngestPlaceholderPosition());
InventoryTask actualTask1 = new InventoryTask("foo_id_1",
mock(ExecuteEngine.class), mock(ExecuteEngine.class), mock(Dumper.class),
mock(Importer.class), finishedPosition);
InventoryTask actualTask2 = new InventoryTask("foo_id_2",
mock(ExecuteEngine.class), mock(ExecuteEngine.class), mock(Dumper.class),
mock(Importer.class), unfinishedPosition);
- Collection<PipelineTask> inventoryTaskArrayList = new ArrayList<>();
- inventoryTaskArrayList.add(actualTask1);
- inventoryTaskArrayList.add(actualTask2);
-
assertFalse(PipelineJobProgressDetector.isAllInventoryTasksFinished(inventoryTaskArrayList));
+
assertFalse(PipelineJobProgressDetector.isAllInventoryTasksFinished(Arrays.asList(actualTask1,
actualTask2)));
}
@Test
@@ -63,59 +56,42 @@ class PipelineJobProgressDetectorTest {
AtomicReference<IngestPosition> finishedPosition = new
AtomicReference<>(new IngestFinishedPosition());
InventoryTask actualTask1 = new InventoryTask("foo_id_1",
mock(ExecuteEngine.class), mock(ExecuteEngine.class), mock(Dumper.class),
mock(Importer.class), finishedPosition);
InventoryTask actualTask2 = new InventoryTask("foo_id_2",
mock(ExecuteEngine.class), mock(ExecuteEngine.class), mock(Dumper.class),
mock(Importer.class), finishedPosition);
- Collection<PipelineTask> inventoryTaskArrayList = new ArrayList<>();
- inventoryTaskArrayList.add(actualTask1);
- inventoryTaskArrayList.add(actualTask2);
-
assertTrue(PipelineJobProgressDetector.isAllInventoryTasksFinished(inventoryTaskArrayList));
+
assertTrue(PipelineJobProgressDetector.isAllInventoryTasksFinished(Arrays.asList(actualTask1,
actualTask2)));
}
@Test
void assertIsInventoryFinishedWhenCollectionElementIsNull() {
TransmissionJobItemProgress jobItemProgress = null;
- Collection<TransmissionJobItemProgress> jobItemProgresses = new
ArrayList<>();
- jobItemProgresses.add(jobItemProgress);
- assertFalse(PipelineJobProgressDetector.isInventoryFinished(1,
jobItemProgresses));
+ assertFalse(PipelineJobProgressDetector.isInventoryFinished(1,
Collections.singleton(jobItemProgress)));
}
@Test
void assertIsInventoryFinishedWhenJobCountDoesNotMatchJobItemProgresses() {
TransmissionJobItemProgress transmissionJobItemProgress = new
TransmissionJobItemProgress();
- List<TransmissionJobItemProgress> jobItemProgresses = new
ArrayList<>();
- jobItemProgresses.add(transmissionJobItemProgress);
- assertFalse(PipelineJobProgressDetector.isInventoryFinished(2,
jobItemProgresses));
+ assertFalse(PipelineJobProgressDetector.isInventoryFinished(2,
Collections.singleton(transmissionJobItemProgress)));
}
@Test
void assertIsInventoryFinishedWhenInventoryTaskProgressHasEmptyMap() {
- JobItemInventoryTasksProgress jobItemInventoryTasksProgress = new
JobItemInventoryTasksProgress(new HashMap<>());
+ JobItemInventoryTasksProgress jobItemInventoryTasksProgress = new
JobItemInventoryTasksProgress(Collections.emptyMap());
TransmissionJobItemProgress transmissionJobItemProgress = new
TransmissionJobItemProgress();
transmissionJobItemProgress.setInventory(jobItemInventoryTasksProgress);
- List<TransmissionJobItemProgress> jobItemProgresses = new
ArrayList<>();
- jobItemProgresses.add(transmissionJobItemProgress);
- assertFalse(PipelineJobProgressDetector.isInventoryFinished(1,
jobItemProgresses));
+ assertFalse(PipelineJobProgressDetector.isInventoryFinished(1,
Collections.singleton(transmissionJobItemProgress)));
}
@Test
void assertIsInventoryFinishedWhenNotAllInventoryTasksCompleted() {
- Map<String, InventoryTaskProgress> progresses = new HashMap<>();
- progresses.put("TEST", new InventoryTaskProgress(new
IngestPlaceholderPosition()));
- JobItemInventoryTasksProgress inventoryTasksProgress = new
JobItemInventoryTasksProgress(progresses);
+ JobItemInventoryTasksProgress inventoryTasksProgress = new
JobItemInventoryTasksProgress(Collections.singletonMap("TEST", new
InventoryTaskProgress(new IngestPlaceholderPosition())));
TransmissionJobItemProgress transmissionJobItemProgress = new
TransmissionJobItemProgress();
transmissionJobItemProgress.setInventory(inventoryTasksProgress);
- List<TransmissionJobItemProgress> jobItemProgresses = new
ArrayList<>();
- jobItemProgresses.add(transmissionJobItemProgress);
- assertFalse(PipelineJobProgressDetector.isInventoryFinished(1,
jobItemProgresses));
+ assertFalse(PipelineJobProgressDetector.isInventoryFinished(1,
Collections.singleton(transmissionJobItemProgress)));
}
@Test
void assertIsInventoryFinished() {
- Map<String, InventoryTaskProgress> progresses = new HashMap<>();
- progresses.put("TEST", new InventoryTaskProgress(new
IngestFinishedPosition()));
- JobItemInventoryTasksProgress inventoryTasksProgress = new
JobItemInventoryTasksProgress(progresses);
+ JobItemInventoryTasksProgress inventoryTasksProgress = new
JobItemInventoryTasksProgress(Collections.singletonMap("TEST", new
InventoryTaskProgress(new IngestFinishedPosition())));
TransmissionJobItemProgress transmissionJobItemProgress = new
TransmissionJobItemProgress();
transmissionJobItemProgress.setInventory(inventoryTasksProgress);
- List<TransmissionJobItemProgress> jobItemProgresses = new
ArrayList<>();
- jobItemProgresses.add(transmissionJobItemProgress);
- assertTrue(PipelineJobProgressDetector.isInventoryFinished(1,
jobItemProgresses));
+ assertTrue(PipelineJobProgressDetector.isInventoryFinished(1,
Collections.singleton(transmissionJobItemProgress)));
}
}
diff --git
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/job/progress/TransmissionJobItemProgressTest.java
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/job/progress/TransmissionJobItemProgressTest.java
index 0831af6e1d8..6588d0bf8de 100644
---
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/job/progress/TransmissionJobItemProgressTest.java
+++
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/job/progress/TransmissionJobItemProgressTest.java
@@ -83,8 +83,8 @@ class TransmissionJobItemProgressTest {
@Test
void assertGetProgressesCorrectly() {
- Map<String, InventoryTaskProgress> progresses = new HashMap<>();
- progresses.put("ds.order_item#0", new InventoryTaskProgress(new
IntegerPrimaryKeyIngestPosition(1, 100)));
+ Map<String, InventoryTaskProgress> progresses = new HashMap<>(4, 1F);
+ progresses.put("ds.order_item#0", new InventoryTaskProgress(new
IntegerPrimaryKeyIngestPosition(1L, 100L)));
progresses.put("ds.order_item#1", new InventoryTaskProgress(new
UnsupportedKeyIngestPosition()));
progresses.put("ds.order#0", new InventoryTaskProgress(new
IngestFinishedPosition()));
progresses.put("ds.test_order#0", new InventoryTaskProgress(new
StringPrimaryKeyIngestPosition("1", "100")));
diff --git
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/metadata/node/PipelineMetaDataNodeTest.java
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/metadata/node/PipelineMetaDataNodeTest.java
index 2f2dd47479b..7a0e9ce8467 100644
---
a/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/metadata/node/PipelineMetaDataNodeTest.java
+++
b/kernel/data-pipeline/core/src/test/java/org/apache/shardingsphere/data/pipeline/core/metadata/node/PipelineMetaDataNodeTest.java
@@ -17,7 +17,6 @@
package org.apache.shardingsphere.data.pipeline.core.metadata.node;
-import org.hamcrest.MatcherAssert;
import org.junit.jupiter.api.Test;
import static org.hamcrest.CoreMatchers.is;
@@ -37,7 +36,7 @@ class PipelineMetaDataNodeTest {
@Test
void assertGetMetaDataDataSourcesPath() {
-
MatcherAssert.assertThat(PipelineMetaDataNode.getMetaDataDataSourcesPath("FIXTURE"),
is(migrationMetaDataRootPath + "/data_sources"));
+ assertThat(PipelineMetaDataNode.getMetaDataDataSourcesPath("FIXTURE"),
is(migrationMetaDataRootPath + "/data_sources"));
}
@Test
diff --git
a/kernel/data-pipeline/dialect/mysql/src/main/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/GlobalTableMapEventMapping.java
b/kernel/data-pipeline/dialect/mysql/src/main/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/GlobalTableMapEventMapping.java
index 095bcc999a6..78a215e80da 100644
---
a/kernel/data-pipeline/dialect/mysql/src/main/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/GlobalTableMapEventMapping.java
+++
b/kernel/data-pipeline/dialect/mysql/src/main/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/GlobalTableMapEventMapping.java
@@ -40,6 +40,6 @@ public final class GlobalTableMapEventMapping {
* @return table map event map
*/
public static Map<Long, MySQLBinlogTableMapEventPacket>
getTableMapEventMap(final String databaseURL) {
- return TABLE_MAP_EVENT_MAPPING.computeIfAbsent(databaseURL, k -> new
ConcurrentHashMap<>());
+ return TABLE_MAP_EVENT_MAPPING.computeIfAbsent(databaseURL, key -> new
ConcurrentHashMap<>());
}
}
diff --git
a/kernel/data-pipeline/dialect/mysql/src/main/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/client/MySQLClient.java
b/kernel/data-pipeline/dialect/mysql/src/main/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/client/MySQLClient.java
index 9ea0b0809b0..df60c00de2f 100644
---
a/kernel/data-pipeline/dialect/mysql/src/main/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/client/MySQLClient.java
+++
b/kernel/data-pipeline/dialect/mysql/src/main/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/client/MySQLClient.java
@@ -361,7 +361,7 @@ public final class MySQLClient {
private void reconnect() throws ExecutionException,
InterruptedException, TimeoutException {
Optional<ChannelFuture> future = closeChannel();
if (future.isPresent()) {
- future.get().get(1, TimeUnit.SECONDS);
+ future.get().get(1L, TimeUnit.SECONDS);
}
if (reconnectTimes.incrementAndGet() > 3) {
log.warn("Exceeds the maximum number of retry times, last
binlog event:{}", lastBinlogEvent);
diff --git
a/kernel/data-pipeline/dialect/mysql/src/main/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/column/value/impl/MySQLUnsignedBigintHandler.java
b/kernel/data-pipeline/dialect/mysql/src/main/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/column/value/impl/MySQLUnsignedBigintHandler.java
index 1a88f402bac..6691cc0207e 100644
---
a/kernel/data-pipeline/dialect/mysql/src/main/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/column/value/impl/MySQLUnsignedBigintHandler.java
+++
b/kernel/data-pipeline/dialect/mysql/src/main/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/column/value/impl/MySQLUnsignedBigintHandler.java
@@ -35,7 +35,7 @@ public final class MySQLUnsignedBigintHandler implements
MySQLDataTypeHandler {
return null;
}
long longValue = (long) value;
- return 0 > longValue ?
BIGINT_MODULO.add(BigInteger.valueOf(longValue)) : longValue;
+ return 0L > longValue ?
BIGINT_MODULO.add(BigInteger.valueOf(longValue)) : longValue;
}
@Override
diff --git
a/kernel/data-pipeline/dialect/mysql/src/test/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/MySQLIngestPositionManagerTest.java
b/kernel/data-pipeline/dialect/mysql/src/test/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/MySQLIngestPositionManagerTest.java
index eca2f990241..a6107d23c0b 100644
---
a/kernel/data-pipeline/dialect/mysql/src/test/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/MySQLIngestPositionManagerTest.java
+++
b/kernel/data-pipeline/dialect/mysql/src/test/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/MySQLIngestPositionManagerTest.java
@@ -42,7 +42,7 @@ class MySQLIngestPositionManagerTest {
private static final long LOG_POSITION = 4L;
- private static final long SERVER_ID = 555555;
+ private static final long SERVER_ID = 555555L;
@Mock(extraInterfaces = AutoCloseable.class)
private DataSource dataSource;
diff --git
a/kernel/data-pipeline/dialect/mysql/src/test/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/binlog/BinlogPositionTest.java
b/kernel/data-pipeline/dialect/mysql/src/test/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/binlog/BinlogPositionTest.java
index 5d7a31939a8..31ae65d93bc 100644
---
a/kernel/data-pipeline/dialect/mysql/src/test/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/binlog/BinlogPositionTest.java
+++
b/kernel/data-pipeline/dialect/mysql/src/test/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/binlog/BinlogPositionTest.java
@@ -26,6 +26,6 @@ class BinlogPositionTest {
@Test
void assertToString() {
- assertThat(new BinlogPosition("mysql-bin.000001", 4, 0L).toString(),
is("mysql-bin.000001#4"));
+ assertThat(new BinlogPosition("mysql-bin.000001", 4L, 0L).toString(),
is("mysql-bin.000001#4"));
}
}
diff --git
a/kernel/data-pipeline/dialect/mysql/src/test/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/client/MySQLClientTest.java
b/kernel/data-pipeline/dialect/mysql/src/test/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/client/MySQLClientTest.java
index f117d8c680c..0485c7d65be 100644
---
a/kernel/data-pipeline/dialect/mysql/src/test/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/client/MySQLClientTest.java
+++
b/kernel/data-pipeline/dialect/mysql/src/test/java/org/apache/shardingsphere/data/pipeline/mysql/ingest/client/MySQLClientTest.java
@@ -100,7 +100,7 @@ class MySQLClientTest {
@Test
void assertExecuteUpdate() throws ReflectiveOperationException {
- MySQLOKPacket expected = new MySQLOKPacket(10, 0, 0);
+ MySQLOKPacket expected = new MySQLOKPacket(10L, 0L, 0);
Plugins.getMemberAccessor().set(MySQLOKPacket.class.getDeclaredField("affectedRows"),
expected, 10L);
mockChannelResponse(expected);
Plugins.getMemberAccessor().set(MySQLClient.class.getDeclaredField("channel"),
mysqlClient, channel);
diff --git
a/kernel/data-pipeline/dialect/opengauss/src/main/java/org/apache/shardingsphere/data/pipeline/opengauss/ingest/OpenGaussWALDumper.java
b/kernel/data-pipeline/dialect/opengauss/src/main/java/org/apache/shardingsphere/data/pipeline/opengauss/ingest/OpenGaussWALDumper.java
index bbed76a6b96..b50fcbc6eee 100644
---
a/kernel/data-pipeline/dialect/opengauss/src/main/java/org/apache/shardingsphere/data/pipeline/opengauss/ingest/OpenGaussWALDumper.java
+++
b/kernel/data-pipeline/dialect/opengauss/src/main/java/org/apache/shardingsphere/data/pipeline/opengauss/ingest/OpenGaussWALDumper.java
@@ -95,7 +95,7 @@ public final class OpenGaussWALDumper extends
AbstractPipelineLifecycleRunnable
this.channel = channel;
walEventConverter = new WALEventConverter(dumperContext,
metaDataLoader);
logicalReplication = new OpenGaussLogicalReplication();
- this.decodeWithTX = dumperContext.isDecodeWithTX();
+ decodeWithTX = dumperContext.isDecodeWithTX();
}
@SneakyThrows(InterruptedException.class)
diff --git
a/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ddlgenerator/AbstractPostgreSQLDDLAdapter.java
b/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ddlgenerator/AbstractPostgreSQLDDLAdapter.java
index 1c0e88a6ca0..de2ae2abc0b 100644
---
a/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ddlgenerator/AbstractPostgreSQLDDLAdapter.java
+++
b/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ddlgenerator/AbstractPostgreSQLDDLAdapter.java
@@ -65,8 +65,9 @@ public abstract class AbstractPostgreSQLDDLAdapter {
ResultSetMetaData metaData = resultSet.getMetaData();
Collection<Map<String, Object>> result = new LinkedList<>();
while (resultSet.next()) {
- Map<String, Object> row = new LinkedHashMap<>();
- for (int i = 1; i <= metaData.getColumnCount(); i++) {
+ int columnCount = metaData.getColumnCount();
+ Map<String, Object> row = new LinkedHashMap<>(columnCount);
+ for (int i = 1; i <= columnCount; i++) {
row.put(metaData.getColumnName(i), resultSet.getObject(i));
}
result.add(row);
@@ -86,9 +87,10 @@ public abstract class AbstractPostgreSQLDDLAdapter {
private Map<String, Object> getSingleRow(final ResultSet resultSet) throws
SQLException {
ResultSetMetaData metaData = resultSet.getMetaData();
- Map<String, Object> result = new LinkedHashMap<>();
+ int columnCount = metaData.getColumnCount();
+ Map<String, Object> result = new LinkedHashMap<>(columnCount);
if (resultSet.next()) {
- for (int i = 1; i <= metaData.getColumnCount(); i++) {
+ for (int i = 1; i <= columnCount; i++) {
result.put(metaData.getColumnName(i), resultSet.getObject(i));
}
}
@@ -102,7 +104,7 @@ public abstract class AbstractPostgreSQLDDLAdapter {
Collection<Map<String, String>> formatLabels = new LinkedList<>();
Collection<String> securityLabels = Arrays.stream((String[]) ((Array)
data.get("seclabels")).getArray()).collect(Collectors.toList());
for (String each : securityLabels) {
- Map<String, String> securityLabel = new LinkedHashMap<>();
+ Map<String, String> securityLabel = new LinkedHashMap<>(2, 1F);
securityLabel.put("provider", each.substring(0,
each.indexOf(SECURITY_LABEL_SPLIT)));
securityLabel.put("label",
each.substring(each.indexOf(SECURITY_LABEL_SPLIT) + 1));
formatLabels.add(securityLabel);
diff --git
a/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ddlgenerator/PostgreSQLColumnPropertiesAppender.java
b/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ddlgenerator/PostgreSQLColumnPropertiesAppender.java
index ea99972f0bd..4be24d07ca3 100644
---
a/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ddlgenerator/PostgreSQLColumnPropertiesAppender.java
+++
b/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ddlgenerator/PostgreSQLColumnPropertiesAppender.java
@@ -221,7 +221,7 @@ public final class PostgreSQLColumnPropertiesAppender
extends AbstractPostgreSQL
name = name.substring(1, name.length() - 1);
}
Integer typmod = (Integer) column.get("atttypmod");
- String length = -1 != typmod ? checkTypmod(typmod, name) : "";
+ String length = -1 == typmod ? "" : checkTypmod(typmod, name);
return getFullTypeValue(name, schema, length, numdims == 1 ? "[]" :
"");
}
diff --git
a/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ddlgenerator/PostgreSQLColumnType.java
b/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ddlgenerator/PostgreSQLColumnType.java
index 649a20e900b..5fbf94c41e3 100644
---
a/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ddlgenerator/PostgreSQLColumnType.java
+++
b/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ddlgenerator/PostgreSQLColumnType.java
@@ -44,7 +44,7 @@ public enum PostgreSQLColumnType {
* @return value of column type
*/
public static PostgreSQLColumnType valueOf(final Long elemoid) {
- if (0 == elemoid) {
+ if (0L == elemoid) {
return UNKNOWN;
}
for (PostgreSQLColumnType each : values()) {
diff --git
a/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ingest/PostgreSQLWALDumper.java
b/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ingest/PostgreSQLWALDumper.java
index 5f7f2adac97..2d1c5d9dff6 100644
---
a/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ingest/PostgreSQLWALDumper.java
+++
b/kernel/data-pipeline/dialect/postgresql/src/main/java/org/apache/shardingsphere/data/pipeline/postgresql/ingest/PostgreSQLWALDumper.java
@@ -85,7 +85,7 @@ public final class PostgreSQLWALDumper extends
AbstractPipelineLifecycleRunnable
this.channel = channel;
walEventConverter = new WALEventConverter(dumperContext,
metaDataLoader);
logicalReplication = new PostgreSQLLogicalReplication();
- this.decodeWithTX = dumperContext.isDecodeWithTX();
+ decodeWithTX = dumperContext.isDecodeWithTX();
}
@SneakyThrows(InterruptedException.class)
@@ -100,7 +100,7 @@ public final class PostgreSQLWALDumper extends
AbstractPipelineLifecycleRunnable
int times = reconnectTimes.incrementAndGet();
log.error("Connect failed, reconnect times={}", times, ex);
if (isRunning()) {
- Thread.sleep(5000);
+ Thread.sleep(5000L);
}
if (times >= 5) {
throw new IngestException(ex);
diff --git
a/kernel/data-pipeline/dialect/postgresql/src/test/java/org/apache/shardingsphere/data/pipeline/postgresql/ingest/wal/WALEventConverterTest.java
b/kernel/data-pipeline/dialect/postgresql/src/test/java/org/apache/shardingsphere/data/pipeline/postgresql/ingest/wal/WALEventConverterTest.java
index fb1782907da..8492f2756c7 100644
---
a/kernel/data-pipeline/dialect/postgresql/src/test/java/org/apache/shardingsphere/data/pipeline/postgresql/ingest/wal/WALEventConverterTest.java
+++
b/kernel/data-pipeline/dialect/postgresql/src/test/java/org/apache/shardingsphere/data/pipeline/postgresql/ingest/wal/WALEventConverterTest.java
@@ -153,7 +153,7 @@ class WALEventConverterTest {
@Test
void assertConvertCommitTXEvent() {
- CommitTXEvent commitTXEvent = new CommitTXEvent(1, 3468L);
+ CommitTXEvent commitTXEvent = new CommitTXEvent(1L, 3468L);
commitTXEvent.setLogSequenceNumber(new
PostgreSQLLogSequenceNumber(logSequenceNumber));
Record record = walEventConverter.convert(commitTXEvent);
assertInstanceOf(PlaceholderRecord.class, record);
diff --git
a/kernel/data-pipeline/distsql/handler/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/distsql/handler/query/ShowStreamingJobStatusExecutor.java
b/kernel/data-pipeline/distsql/handler/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/distsql/handler/query/ShowStreamingJobStatusExecutor.java
index ca408b4e353..be7253f94fe 100644
---
a/kernel/data-pipeline/distsql/handler/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/distsql/handler/query/ShowStreamingJobStatusExecutor.java
+++
b/kernel/data-pipeline/distsql/handler/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/distsql/handler/query/ShowStreamingJobStatusExecutor.java
@@ -67,7 +67,7 @@ public final class ShowStreamingJobStatusExecutor implements
DistSQLQueryExecuto
}
private static Optional<Long> getIncrementalIdleSeconds(final
TransmissionJobItemProgress jobItemProgress, final TransmissionJobItemInfo
transmissionJobItemInfo, final long currentTimeMillis) {
- if
(jobItemProgress.getIncremental().getIncrementalLatestActiveTimeMillis() > 0) {
+ if
(jobItemProgress.getIncremental().getIncrementalLatestActiveTimeMillis() > 0L) {
long latestActiveTimeMillis =
Math.max(transmissionJobItemInfo.getStartTimeMillis(),
jobItemProgress.getIncremental().getIncrementalLatestActiveTimeMillis());
return
Optional.of(TimeUnit.MILLISECONDS.toSeconds(currentTimeMillis -
latestActiveTimeMillis));
}
diff --git
a/kernel/data-pipeline/distsql/handler/src/main/java/org/apache/shardingsphere/data/pipeline/migration/distsql/handler/query/ShowMigrationJobStatusExecutor.java
b/kernel/data-pipeline/distsql/handler/src/main/java/org/apache/shardingsphere/data/pipeline/migration/distsql/handler/query/ShowMigrationJobStatusExecutor.java
index 83bf0bf39ee..bb6e3553924 100644
---
a/kernel/data-pipeline/distsql/handler/src/main/java/org/apache/shardingsphere/data/pipeline/migration/distsql/handler/query/ShowMigrationJobStatusExecutor.java
+++
b/kernel/data-pipeline/distsql/handler/src/main/java/org/apache/shardingsphere/data/pipeline/migration/distsql/handler/query/ShowMigrationJobStatusExecutor.java
@@ -60,7 +60,7 @@ public final class ShowMigrationJobStatusExecutor implements
DistSQLQueryExecuto
}
private Optional<Long> getIncrementalIdleSeconds(final
TransmissionJobItemProgress jobItemProgress, final TransmissionJobItemInfo
jobItemInfo, final long currentTimeMillis) {
- if
(jobItemProgress.getIncremental().getIncrementalLatestActiveTimeMillis() > 0) {
+ if
(jobItemProgress.getIncremental().getIncrementalLatestActiveTimeMillis() > 0L) {
long latestActiveTimeMillis =
Math.max(jobItemInfo.getStartTimeMillis(),
jobItemProgress.getIncremental().getIncrementalLatestActiveTimeMillis());
return
Optional.of(TimeUnit.MILLISECONDS.toSeconds(currentTimeMillis -
latestActiveTimeMillis));
}
diff --git
a/kernel/data-pipeline/scenario/cdc/client/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/client/util/ProtobufAnyValueConverter.java
b/kernel/data-pipeline/scenario/cdc/client/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/client/util/ProtobufAnyValueConverter.java
index e1eebc88028..0a563cb3b01 100644
---
a/kernel/data-pipeline/scenario/cdc/client/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/client/util/ProtobufAnyValueConverter.java
+++
b/kernel/data-pipeline/scenario/cdc/client/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/client/util/ProtobufAnyValueConverter.java
@@ -98,7 +98,7 @@ public final class ProtobufAnyValueConverter {
}
private static Timestamp converProtobufTimestamp(final
com.google.protobuf.Timestamp timestamp) {
- Timestamp result = new Timestamp(timestamp.getSeconds() * 1000);
+ Timestamp result = new Timestamp(timestamp.getSeconds() * 1000L);
result.setNanos(timestamp.getNanos());
return result;
}
diff --git
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/CDCJobType.java
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/CDCJobType.java
index 47fc27d3e94..eb97ea51f0c 100644
---
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/CDCJobType.java
+++
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/CDCJobType.java
@@ -72,7 +72,7 @@ public final class CDCJobType implements PipelineJobType {
@Override
public PipelineDataConsistencyChecker buildDataConsistencyChecker(final
PipelineJobConfiguration jobConfig, final TransmissionProcessContext
processContext,
final
ConsistencyCheckJobItemProgressContext progressContext) {
- throw new UnsupportedOperationException();
+ throw new UnsupportedOperationException("");
}
@Override
diff --git
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/api/CDCJobAPI.java
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/api/CDCJobAPI.java
index 29e65fbfce1..b91a975f7a9 100644
---
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/api/CDCJobAPI.java
+++
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/api/CDCJobAPI.java
@@ -168,7 +168,7 @@ public final class CDCJobAPI implements TransmissionJobAPI {
}
private ShardingSpherePipelineDataSourceConfiguration
getDataSourceConfiguration(final ShardingSphereDatabase database) {
- Map<String, Map<String, Object>> dataSourcePoolProps = new HashMap<>();
+ Map<String, Map<String, Object>> dataSourcePoolProps = new
HashMap<>(database.getResourceMetaData().getStorageUnits().size(), 1F);
for (Entry<String, StorageUnit> entry :
database.getResourceMetaData().getStorageUnits().entrySet()) {
dataSourcePoolProps.put(entry.getKey(),
dataSourceConfigSwapper.swapToMap(entry.getValue().getDataSourcePoolProperties()));
}
diff --git
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/context/CDCJobItemContext.java
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/context/CDCJobItemContext.java
index dfb8b23d5b0..b721cc44b7b 100644
---
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/context/CDCJobItemContext.java
+++
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/context/CDCJobItemContext.java
@@ -71,9 +71,9 @@ public final class CDCJobItemContext implements
TransmissionJobItemContext {
private final Collection<PipelineTask> incrementalTasks = new
LinkedList<>();
- private final AtomicLong processedRecordsCount = new AtomicLong(0);
+ private final AtomicLong processedRecordsCount = new AtomicLong(0L);
- private final AtomicLong inventoryRecordsCount = new AtomicLong(0);
+ private final AtomicLong inventoryRecordsCount = new AtomicLong(0L);
private final LazyInitializer<PipelineDataSourceWrapper>
sourceDataSourceLazyInitializer = new
LazyInitializer<PipelineDataSourceWrapper>() {
diff --git
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/core/importer/CDCImporter.java
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/core/importer/CDCImporter.java
index ff130efc1a6..2357fb41384 100644
---
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/core/importer/CDCImporter.java
+++
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/core/importer/CDCImporter.java
@@ -71,7 +71,7 @@ public final class CDCImporter extends
AbstractPipelineLifecycleRunnable impleme
private final PriorityQueue<CSNRecords> csnRecordsQueue = new
PriorityQueue<>(new CSNRecordsComparator());
- private final Cache<String, List<Pair<CDCChannelProgressPair,
CDCAckPosition>>> ackCache =
Caffeine.newBuilder().maximumSize(10000).expireAfterAccess(5,
TimeUnit.MINUTES).build();
+ private final Cache<String, List<Pair<CDCChannelProgressPair,
CDCAckPosition>>> ackCache =
Caffeine.newBuilder().maximumSize(10000L).expireAfterAccess(5L,
TimeUnit.MINUTES).build();
@Override
protected void runBlocking() {
diff --git
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/handler/CDCBackendHandler.java
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/handler/CDCBackendHandler.java
index 825e50df67a..4baf8622e45 100644
---
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/handler/CDCBackendHandler.java
+++
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/handler/CDCBackendHandler.java
@@ -110,7 +110,7 @@ public final class CDCBackendHandler {
schemaTableNameMap =
CDCSchemaTableUtils.parseTableExpressionWithSchema(database,
requestBody.getSourceSchemaTableList());
// TODO if different schema have same table names, table name may
be overwritten, because the table name at sharding rule not contain schema.
tableNames =
schemaTableNameMap.values().stream().flatMap(Collection::stream).collect(Collectors.toList());
- schemaTableNameMap.forEach((k, v) -> v.forEach(tableName ->
schemaTableNames.add(k.isEmpty() ? tableName : String.join(".", k,
tableName))));
+ schemaTableNameMap.forEach((key, value) -> value.forEach(tableName
-> schemaTableNames.add(key.isEmpty() ? tableName : String.join(".", key,
tableName))));
} else {
schemaTableNames.addAll(CDCSchemaTableUtils.parseTableExpressionWithoutSchema(database,
requestBody.getSourceSchemaTableList().stream().map(SchemaTable::getTable)
.collect(Collectors.toList())));
diff --git
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/util/CDCDataNodeUtils.java
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/util/CDCDataNodeUtils.java
index a7fab4f593e..2fc2d55845d 100644
---
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/util/CDCDataNodeUtils.java
+++
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/util/CDCDataNodeUtils.java
@@ -54,7 +54,7 @@ public final class CDCDataNodeUtils {
Optional<ShardingRule> shardingRule =
database.getRuleMetaData().findSingleRule(ShardingRule.class);
Optional<SingleRule> singleRule =
database.getRuleMetaData().findSingleRule(SingleRule.class);
Optional<BroadcastRule> broadcastRule =
database.getRuleMetaData().findSingleRule(BroadcastRule.class);
- Map<String, List<DataNode>> result = new HashMap<>();
+ Map<String, List<DataNode>> result = new HashMap<>(tableNames.size(),
1F);
// TODO support virtual data source name
for (String each : tableNames) {
if (singleRule.isPresent() &&
singleRule.get().getAttributes().getAttribute(DataNodeRuleAttribute.class).getAllDataNodes().containsKey(each))
{
diff --git
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/util/CDCSchemaTableUtils.java
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/util/CDCSchemaTableUtils.java
index aaed4565c58..be6fcb8150b 100644
---
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/util/CDCSchemaTableUtils.java
+++
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/util/CDCSchemaTableUtils.java
@@ -77,7 +77,7 @@ public final class CDCSchemaTableUtils {
}
private static Map<String, Set<String>>
parseTableExpressionWithAllTables(final ShardingSphereDatabase database, final
Collection<String> systemSchemas) {
- Map<String, Set<String>> result = new
HashMap<>(database.getSchemas().size(), 1);
+ Map<String, Set<String>> result = new
HashMap<>(database.getSchemas().size(), 1F);
for (Entry<String, ShardingSphereSchema> entry :
database.getSchemas().entrySet()) {
if (!systemSchemas.contains(entry.getKey())) {
entry.getValue().getAllTableNames().forEach(tableName ->
result.computeIfAbsent(entry.getKey(), ignored -> new
HashSet<>()).add(tableName));
@@ -88,7 +88,7 @@ public final class CDCSchemaTableUtils {
}
private static Map<String, Set<String>>
parseTableExpressionWithAllSchema(final ShardingSphereDatabase database, final
Collection<String> systemSchemas, final SchemaTable table) {
- Map<String, Set<String>> result = new HashMap<>();
+ Map<String, Set<String>> result = new
HashMap<>(database.getSchemas().size(), 1F);
for (Entry<String, ShardingSphereSchema> entry :
database.getSchemas().entrySet()) {
if (!systemSchemas.contains(entry.getKey())) {
entry.getValue().getAllTableNames().stream().filter(tableName
-> tableName.equals(table.getTable())).findFirst()
@@ -99,11 +99,11 @@ public final class CDCSchemaTableUtils {
}
private static Map<String, Set<String>>
parseTableExpressionWithAllTable(final ShardingSphereDatabase database, final
SchemaTable each) {
- Map<String, Set<String>> result = new HashMap<>();
DialectDatabaseMetaData dialectDatabaseMetaData = new
DatabaseTypeRegistry(database.getProtocolType()).getDialectDatabaseMetaData();
String schemaName = each.getSchema().isEmpty() ?
dialectDatabaseMetaData.getDefaultSchema().orElseThrow(() -> new
IllegalStateException("Default schema should exist.")) : each.getSchema();
ShardingSphereSchema schema = database.getSchema(schemaName);
ShardingSpherePreconditions.checkNotNull(schema, () -> new
SchemaNotFoundException(each.getSchema()));
+ Map<String, Set<String>> result = new
HashMap<>(schema.getAllTableNames().size(), 1F);
schema.getAllTableNames().forEach(tableName ->
result.computeIfAbsent(schemaName, ignored -> new HashSet<>()).add(tableName));
return result;
}
diff --git
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/util/ColumnValueConvertUtils.java
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/util/ColumnValueConvertUtils.java
index 1f61502d80a..0c200c8fc2f 100644
---
a/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/util/ColumnValueConvertUtils.java
+++
b/kernel/data-pipeline/scenario/cdc/core/src/main/java/org/apache/shardingsphere/data/pipeline/cdc/util/ColumnValueConvertUtils.java
@@ -135,11 +135,11 @@ public final class ColumnValueConvertUtils {
}
if (object instanceof Clob) {
Clob clob = (Clob) object;
- return StringValue.of(clob.getSubString(1, (int) clob.length()));
+ return StringValue.of(clob.getSubString(1L, (int) clob.length()));
}
if (object instanceof Blob) {
Blob blob = (Blob) object;
- return BytesValue.of(ByteString.copyFrom(blob.getBytes(1, (int)
blob.length())));
+ return BytesValue.of(ByteString.copyFrom(blob.getBytes(1L, (int)
blob.length())));
}
return StringValue.newBuilder().setValue(object.toString()).build();
}
@@ -147,9 +147,9 @@ public final class ColumnValueConvertUtils {
private static com.google.protobuf.Timestamp
converToProtobufTimestamp(final Date timestamp) {
if (timestamp instanceof Timestamp) {
Timestamp value = (Timestamp) timestamp;
- return
com.google.protobuf.Timestamp.newBuilder().setSeconds(value.getTime() /
1000).setNanos(value.getNanos()).build();
+ return
com.google.protobuf.Timestamp.newBuilder().setSeconds(value.getTime() /
1000L).setNanos(value.getNanos()).build();
}
long millis = timestamp.getTime();
- return com.google.protobuf.Timestamp.newBuilder().setSeconds(millis /
1000).setNanos((int) ((millis % 1000) * 1000000)).build();
+ return com.google.protobuf.Timestamp.newBuilder().setSeconds(millis /
1000L).setNanos((int) ((millis % 1000L) * 1000000L)).build();
}
}
diff --git
a/kernel/data-pipeline/scenario/cdc/core/src/test/java/org/apache/shardingsphere/data/pipeline/cdc/util/CDCSchemaTableUtilsTest.java
b/kernel/data-pipeline/scenario/cdc/core/src/test/java/org/apache/shardingsphere/data/pipeline/cdc/util/CDCSchemaTableUtilsTest.java
index 220be845acc..09ae021e0d7 100644
---
a/kernel/data-pipeline/scenario/cdc/core/src/test/java/org/apache/shardingsphere/data/pipeline/cdc/util/CDCSchemaTableUtilsTest.java
+++
b/kernel/data-pipeline/scenario/cdc/core/src/test/java/org/apache/shardingsphere/data/pipeline/cdc/util/CDCSchemaTableUtilsTest.java
@@ -42,13 +42,13 @@ class CDCSchemaTableUtilsTest {
@Test
void assertParseTableExpression() {
- Map<String, ShardingSphereSchema> schemas = new HashMap<>();
+ Map<String, ShardingSphereSchema> schemas = new HashMap<>(2, 1F);
schemas.put("public", mockedPublicSchema());
schemas.put("test", mockedTestSchema());
ShardingSphereDatabase database = new
ShardingSphereDatabase("sharding_db",
TypedSPILoader.getService(DatabaseType.class, "openGauss"), null, null,
schemas);
List<SchemaTable> schemaTables =
Arrays.asList(SchemaTable.newBuilder().setSchema("public").setTable("t_order").build(),
SchemaTable.newBuilder().setSchema("test").setTable("*").build());
- Map<String, Set<String>> expected = new HashMap<>();
+ Map<String, Set<String>> expected = new HashMap<>(2, 1F);
expected.put("test", new HashSet<>(Arrays.asList("t_order_item",
"t_order_item2")));
expected.put("public", Collections.singleton("t_order"));
Map<String, Set<String>> actual =
CDCSchemaTableUtils.parseTableExpressionWithSchema(database, schemaTables);
@@ -59,8 +59,7 @@ class CDCSchemaTableUtilsTest {
assertThat(actual, is(expected));
schemaTables =
Collections.singletonList(SchemaTable.newBuilder().setSchema("*").setTable("t_order").build());
actual = CDCSchemaTableUtils.parseTableExpressionWithSchema(database,
schemaTables);
- expected = new HashMap<>();
- expected.put("public", Collections.singleton("t_order"));
+ expected = Collections.singletonMap("public",
Collections.singleton("t_order"));
assertThat(actual, is(expected));
}
diff --git
a/kernel/data-pipeline/scenario/cdc/core/src/test/java/org/apache/shardingsphere/data/pipeline/cdc/util/ColumnValueConvertUtilsTest.java
b/kernel/data-pipeline/scenario/cdc/core/src/test/java/org/apache/shardingsphere/data/pipeline/cdc/util/ColumnValueConvertUtilsTest.java
index e9d33ef5d18..86974d9597b 100644
---
a/kernel/data-pipeline/scenario/cdc/core/src/test/java/org/apache/shardingsphere/data/pipeline/cdc/util/ColumnValueConvertUtilsTest.java
+++
b/kernel/data-pipeline/scenario/cdc/core/src/test/java/org/apache/shardingsphere/data/pipeline/cdc/util/ColumnValueConvertUtilsTest.java
@@ -110,7 +110,7 @@ class ColumnValueConvertUtilsTest {
@Test
void assertTimeConvert() {
- Time time = new Time(-3600 * 1000 - 1234);
+ Time time = new Time(-3600L * 1000L - 1234L);
int nanos = new Timestamp(time.getTime()).getNanos();
Int64Value actualMessage = (Int64Value)
ColumnValueConvertUtils.convertToProtobufMessage(time);
assertThat(LocalTime.ofNanoOfDay(actualMessage.getValue()),
is(time.toLocalTime().withNano(nanos)));
diff --git
a/kernel/data-pipeline/scenario/cdc/core/src/test/java/org/apache/shardingsphere/data/pipeline/cdc/util/DataRecordResultConvertUtilsTest.java
b/kernel/data-pipeline/scenario/cdc/core/src/test/java/org/apache/shardingsphere/data/pipeline/cdc/util/DataRecordResultConvertUtilsTest.java
index ffbb09e938b..0f3c39146d5 100644
---
a/kernel/data-pipeline/scenario/cdc/core/src/test/java/org/apache/shardingsphere/data/pipeline/cdc/util/DataRecordResultConvertUtilsTest.java
+++
b/kernel/data-pipeline/scenario/cdc/core/src/test/java/org/apache/shardingsphere/data/pipeline/cdc/util/DataRecordResultConvertUtilsTest.java
@@ -55,7 +55,7 @@ class DataRecordResultConvertUtilsTest {
@Test
void assertConvertDataRecordToRecord() throws
InvalidProtocolBufferException, SQLException {
- DataRecord dataRecord = new
DataRecord(PipelineSQLOperationType.INSERT, "t_order", new
IntegerPrimaryKeyIngestPosition(0, 1), 2);
+ DataRecord dataRecord = new
DataRecord(PipelineSQLOperationType.INSERT, "t_order", new
IntegerPrimaryKeyIngestPosition(0L, 1L), 2);
dataRecord.addColumn(new Column("order_id", BigInteger.ONE, false,
true));
dataRecord.addColumn(new Column("price", BigDecimal.valueOf(123),
false, false));
dataRecord.addColumn(new Column("user_id", Long.MAX_VALUE, false,
false));
diff --git
a/kernel/data-pipeline/scenario/consistencycheck/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/consistencycheck/api/ConsistencyCheckJobAPI.java
b/kernel/data-pipeline/scenario/consistencycheck/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/consistencycheck/api/ConsistencyCheckJobAPI.java
index c22d89843c2..5726948ca12 100644
---
a/kernel/data-pipeline/scenario/consistencycheck/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/consistencycheck/api/ConsistencyCheckJobAPI.java
+++
b/kernel/data-pipeline/scenario/consistencycheck/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/consistencycheck/api/ConsistencyCheckJobAPI.java
@@ -243,8 +243,8 @@ public final class ConsistencyCheckJobAPI {
result.setDurationSeconds(duration.getSeconds());
result.setCheckEndTime(DateTimeFormatterFactory.getLongMillsFormatter().format(checkEndTime));
result.setInventoryRemainingSeconds(0L);
- } else if (0 != recordsCount && 0 != checkedRecordsCount) {
- result.setInventoryFinishedPercentage((int) (checkedRecordsCount *
100 / recordsCount));
+ } else if (0L != recordsCount && 0L != checkedRecordsCount) {
+ result.setInventoryFinishedPercentage((int) (checkedRecordsCount *
100L / recordsCount));
LocalDateTime stopTime = jobConfigPOJO.isDisabled() ?
LocalDateTime.from(DateTimeFormatterFactory.getStandardFormatter().parse(jobConfigPOJO.getProps().getProperty("stop_time")))
: null;
long durationMillis = (null != stopTime ?
Timestamp.valueOf(stopTime).getTime() : System.currentTimeMillis()) -
jobItemProgress.getCheckBeginTimeMillis();
@@ -252,8 +252,8 @@ public final class ConsistencyCheckJobAPI {
if (null != stopTime) {
result.setCheckEndTime(jobConfigPOJO.getProps().getProperty("stop_time"));
}
- long remainingMills = Math.max(0, (long) ((recordsCount -
checkedRecordsCount) * 1.0D / checkedRecordsCount * durationMillis));
- result.setInventoryRemainingSeconds(remainingMills / 1000);
+ long remainingMills = Math.max(0L, (long) ((recordsCount -
checkedRecordsCount) * 1.0D / checkedRecordsCount * durationMillis));
+ result.setInventoryRemainingSeconds(remainingMills / 1000L);
}
}
diff --git
a/kernel/data-pipeline/scenario/migration/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/migration/api/MigrationJobAPI.java
b/kernel/data-pipeline/scenario/migration/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/migration/api/MigrationJobAPI.java
index 5e2ff9914d0..ba4e11f69de 100644
---
a/kernel/data-pipeline/scenario/migration/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/migration/api/MigrationJobAPI.java
+++
b/kernel/data-pipeline/scenario/migration/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/migration/api/MigrationJobAPI.java
@@ -173,7 +173,7 @@ public final class MigrationJobAPI implements
TransmissionJobAPI {
}
private PipelineDataSourceConfiguration
buildTargetPipelineDataSourceConfiguration(final ShardingSphereDatabase
targetDatabase) {
- Map<String, Map<String, Object>> targetPoolProps = new HashMap<>();
+ Map<String, Map<String, Object>> targetPoolProps = new
HashMap<>(targetDatabase.getResourceMetaData().getStorageUnits().size(), 1F);
YamlDataSourceConfigurationSwapper dataSourceConfigSwapper = new
YamlDataSourceConfigurationSwapper();
for (Entry<String, StorageUnit> entry :
targetDatabase.getResourceMetaData().getStorageUnits().entrySet()) {
targetPoolProps.put(entry.getKey(),
dataSourceConfigSwapper.swapToMap(entry.getValue().getDataSourcePoolProperties()));
@@ -192,7 +192,7 @@ public final class MigrationJobAPI implements
TransmissionJobAPI {
}
private Map<String, String> buildTargetTableSchemaMap(final Map<String,
List<DataNode>> sourceDataNodes) {
- Map<String, String> result = new LinkedHashMap<>();
+ Map<String, String> result = new
LinkedHashMap<>(sourceDataNodes.size(), 1F);
sourceDataNodes.forEach((tableName, dataNodes) ->
result.put(tableName, dataNodes.get(0).getSchemaName()));
return result;
}
@@ -285,7 +285,7 @@ public final class MigrationJobAPI implements
TransmissionJobAPI {
}
private void refreshTableMetadata(final String jobId, final String
databaseName) {
- // TODO use origin database name now, wait reloadDatabaseMetaData fix
case-sensitive probelm
+ // TODO use origin database name now, wait reloadDatabaseMetaData fix
case-sensitive problem
ContextManager contextManager =
PipelineContextManager.getContext(PipelineJobIdUtils.parseContextKey(jobId)).getContextManager();
ShardingSphereDatabase database =
contextManager.getMetaDataContexts().getMetaData().getDatabase(databaseName);
contextManager.refreshTableMetaData(database);
diff --git
a/kernel/data-pipeline/scenario/migration/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/migration/context/MigrationJobItemContext.java
b/kernel/data-pipeline/scenario/migration/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/migration/context/MigrationJobItemContext.java
index a7caa8aa922..743ca8f47e4 100644
---
a/kernel/data-pipeline/scenario/migration/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/migration/context/MigrationJobItemContext.java
+++
b/kernel/data-pipeline/scenario/migration/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/migration/context/MigrationJobItemContext.java
@@ -67,9 +67,9 @@ public final class MigrationJobItemContext implements
TransmissionJobItemContext
private final Collection<PipelineTask> incrementalTasks = new
LinkedList<>();
- private final AtomicLong processedRecordsCount = new AtomicLong(0);
+ private final AtomicLong processedRecordsCount = new AtomicLong(0L);
- private final AtomicLong inventoryRecordsCount = new AtomicLong(0);
+ private final AtomicLong inventoryRecordsCount = new AtomicLong(0L);
private final MigrationJobConfiguration jobConfig;
@@ -98,7 +98,7 @@ public final class MigrationJobItemContext implements
TransmissionJobItemContext
this.jobConfig = jobConfig;
jobId = jobConfig.getJobId();
this.shardingItem = shardingItem;
- this.dataSourceName = taskConfig.getDataSourceName();
+ dataSourceName = taskConfig.getDataSourceName();
this.initProgress = initProgress;
if (null != initProgress) {
processedRecordsCount.set(initProgress.getProcessedRecordsCount());
diff --git
a/kernel/data-pipeline/scenario/migration/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/migration/preparer/MigrationJobPreparer.java
b/kernel/data-pipeline/scenario/migration/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/migration/preparer/MigrationJobPreparer.java
index b0c695fc383..a893e286c24 100644
---
a/kernel/data-pipeline/scenario/migration/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/migration/preparer/MigrationJobPreparer.java
+++
b/kernel/data-pipeline/scenario/migration/src/main/java/org/apache/shardingsphere/data/pipeline/scenario/migration/preparer/MigrationJobPreparer.java
@@ -130,7 +130,7 @@ public final class MigrationJobPreparer {
}
LockDefinition lockDefinition = new
GlobalLockDefinition(String.format(GlobalLockNames.PREPARE.getLockName(),
jobConfig.getJobId()));
long startTimeMillis = System.currentTimeMillis();
- if (lockContext.tryLock(lockDefinition, 600000)) {
+ if (lockContext.tryLock(lockDefinition, 600000L)) {
log.info("try lock success, jobId={}, shardingItem={}, cost {}
ms", jobId, jobItemContext.getShardingItem(), System.currentTimeMillis() -
startTimeMillis);
try {
JobOffsetInfo offsetInfo =
PipelineAPIFactory.getPipelineGovernanceFacade(PipelineJobIdUtils.parseContextKey(jobId)).getJobFacade().getOffset().load(jobId);
diff --git
a/kernel/global-clock/type/tso/provider/redis/src/main/java/org/apache/shardingsphere/globalclock/type/tso/provider/RedisTSOProvider.java
b/kernel/global-clock/type/tso/provider/redis/src/main/java/org/apache/shardingsphere/globalclock/type/tso/provider/RedisTSOProvider.java
index 48b69f7d39a..67838b40bd9 100644
---
a/kernel/global-clock/type/tso/provider/redis/src/main/java/org/apache/shardingsphere/globalclock/type/tso/provider/RedisTSOProvider.java
+++
b/kernel/global-clock/type/tso/provider/redis/src/main/java/org/apache/shardingsphere/globalclock/type/tso/provider/RedisTSOProvider.java
@@ -32,7 +32,7 @@ public final class RedisTSOProvider implements TSOProvider {
private static final String CSN_KEY = "csn";
- private static final long ERROR_CSN = 0;
+ private static final long ERROR_CSN = 0L;
private static final long INIT_CSN = Integer.MAX_VALUE;
diff --git
a/kernel/logging/core/src/main/java/org/apache/shardingsphere/logging/util/LoggingUtils.java
b/kernel/logging/core/src/main/java/org/apache/shardingsphere/logging/util/LoggingUtils.java
index 206e713952a..440283a7bc1 100644
---
a/kernel/logging/core/src/main/java/org/apache/shardingsphere/logging/util/LoggingUtils.java
+++
b/kernel/logging/core/src/main/java/org/apache/shardingsphere/logging/util/LoggingUtils.java
@@ -64,7 +64,7 @@ public final class LoggingUtils {
* @param props configuration properties
*/
public static void syncLoggingConfig(final LoggingRuleConfiguration
loggingRuleConfig, final ConfigurationProperties props) {
- LoggingUtils.getSQLLogger(loggingRuleConfig).ifPresent(option -> {
+ getSQLLogger(loggingRuleConfig).ifPresent(option -> {
Properties loggerProps = option.getProps();
syncPropsToLoggingRule(loggerProps, props);
syncLoggingRuleToProps(loggerProps, props);
diff --git
a/kernel/metadata/core/src/main/java/org/apache/shardingsphere/metadata/persist/service/database/DatabaseMetaDataPersistService.java
b/kernel/metadata/core/src/main/java/org/apache/shardingsphere/metadata/persist/service/database/DatabaseMetaDataPersistService.java
index 351629379eb..d3034bea2ee 100644
---
a/kernel/metadata/core/src/main/java/org/apache/shardingsphere/metadata/persist/service/database/DatabaseMetaDataPersistService.java
+++
b/kernel/metadata/core/src/main/java/org/apache/shardingsphere/metadata/persist/service/database/DatabaseMetaDataPersistService.java
@@ -48,8 +48,8 @@ public final class DatabaseMetaDataPersistService implements
DatabaseMetaDataBas
public DatabaseMetaDataPersistService(final PersistRepository repository,
final MetaDataVersionPersistService metaDataVersionPersistService) {
this.repository = repository;
this.metaDataVersionPersistService = metaDataVersionPersistService;
- this.tableMetaDataPersistService = new
TableMetaDataPersistService(repository, metaDataVersionPersistService);
- this.viewMetaDataPersistService = new
ViewMetaDataPersistService(repository, metaDataVersionPersistService);
+ tableMetaDataPersistService = new
TableMetaDataPersistService(repository, metaDataVersionPersistService);
+ viewMetaDataPersistService = new
ViewMetaDataPersistService(repository, metaDataVersionPersistService);
}
@Override
diff --git
a/kernel/single/core/src/main/java/org/apache/shardingsphere/single/datanode/SingleTableDataNodeLoader.java
b/kernel/single/core/src/main/java/org/apache/shardingsphere/single/datanode/SingleTableDataNodeLoader.java
index 42500e70f37..d8443e5d7ca 100644
---
a/kernel/single/core/src/main/java/org/apache/shardingsphere/single/datanode/SingleTableDataNodeLoader.java
+++
b/kernel/single/core/src/main/java/org/apache/shardingsphere/single/datanode/SingleTableDataNodeLoader.java
@@ -114,7 +114,7 @@ public final class SingleTableDataNodeLoader {
private static Map<String, Collection<DataNode>>
loadSpecifiedDataNodes(final Map<String, Collection<DataNode>> actualDataNodes,
final Collection<String> featureRequiredSingleTables,
final Map<String, Map<String, Collection<String>>> configuredTableMap) {
- Map<String, Collection<DataNode>> result = new ConcurrentHashMap<>();
+ Map<String, Collection<DataNode>> result = new
ConcurrentHashMap<>(actualDataNodes.size(), 1F);
for (Entry<String, Collection<DataNode>> entry :
actualDataNodes.entrySet()) {
Collection<DataNode> singleNode =
loadSpecifiedDataNode(entry.getValue(), featureRequiredSingleTables,
configuredTableMap);
if (!singleNode.isEmpty()) {
@@ -158,8 +158,8 @@ public final class SingleTableDataNodeLoader {
if (configuredTables.isEmpty()) {
return Collections.emptyMap();
}
- Map<String, Map<String, Collection<String>>> result = new
LinkedHashMap<>();
Collection<DataNode> dataNodes =
SingleTableLoadUtils.convertToDataNodes(databaseName, protocolType,
configuredTables);
+ Map<String, Map<String, Collection<String>>> result = new
LinkedHashMap<>(dataNodes.size(), 1F);
for (DataNode each : dataNodes) {
Map<String, Collection<String>> schemaTables =
result.getOrDefault(each.getDataSourceName(), new LinkedHashMap<>());
Collection<String> tables =
schemaTables.getOrDefault(each.getSchemaName(), new LinkedList<>());
diff --git
a/kernel/single/core/src/main/java/org/apache/shardingsphere/single/decider/SingleSQLFederationDecider.java
b/kernel/single/core/src/main/java/org/apache/shardingsphere/single/decider/SingleSQLFederationDecider.java
index 0799b25da5f..89963073110 100644
---
a/kernel/single/core/src/main/java/org/apache/shardingsphere/single/decider/SingleSQLFederationDecider.java
+++
b/kernel/single/core/src/main/java/org/apache/shardingsphere/single/decider/SingleSQLFederationDecider.java
@@ -66,7 +66,7 @@ public final class SingleSQLFederationDecider implements
SQLFederationDecider<Si
}
private Collection<DataNode> getTableDataNodes(final SingleRule rule,
final Collection<QualifiedTable> singleTables) {
- Collection<DataNode> result = new HashSet<>();
+ Collection<DataNode> result = new HashSet<>(singleTables.size(), 1F);
for (QualifiedTable each : singleTables) {
rule.getAttributes().getAttribute(MutableDataNodeRuleAttribute.class).findTableDataNode(each.getSchemaName(),
each.getTableName()).ifPresent(result::add);
}
diff --git
a/kernel/single/core/src/main/java/org/apache/shardingsphere/single/decorator/SingleRuleConfigurationDecorator.java
b/kernel/single/core/src/main/java/org/apache/shardingsphere/single/decorator/SingleRuleConfigurationDecorator.java
index 34182d0f86b..e9b2c81838f 100644
---
a/kernel/single/core/src/main/java/org/apache/shardingsphere/single/decorator/SingleRuleConfigurationDecorator.java
+++
b/kernel/single/core/src/main/java/org/apache/shardingsphere/single/decorator/SingleRuleConfigurationDecorator.java
@@ -82,8 +82,8 @@ public final class SingleRuleConfigurationDecorator
implements RuleConfiguration
private Collection<String> loadSpecifiedTables(final boolean
isSchemaSupportedDatabaseType, final Map<String, Collection<DataNode>>
actualDataNodes,
final
Collection<ShardingSphereRule> builtRules, final Collection<DataNode>
configuredDataNodes) {
- Collection<String> expandRequiredDataSources = new LinkedHashSet<>();
- Map<String, DataNode> expectedDataNodes = new LinkedHashMap<>();
+ Collection<String> expandRequiredDataSources = new
LinkedHashSet<>(configuredDataNodes.size(), 1F);
+ Map<String, DataNode> expectedDataNodes = new
LinkedHashMap<>(configuredDataNodes.size(), 1F);
for (DataNode each : configuredDataNodes) {
if (SingleTableConstants.ASTERISK.equals(each.getTableName())) {
expandRequiredDataSources.add(each.getDataSourceName());
@@ -101,7 +101,7 @@ public final class SingleRuleConfigurationDecorator
implements RuleConfiguration
private Collection<String> loadSpecifiedTablesWithExpand(final boolean
isSchemaSupportedDatabaseType, final Map<String, Collection<DataNode>>
actualDataNodes,
final
Collection<String> featureRequiredSingleTables, final Collection<String>
expandRequiredDataSources,
final Map<String,
DataNode> expectedDataNodes) {
- Collection<String> result = new LinkedHashSet<>();
+ Collection<String> result = new
LinkedHashSet<>(actualDataNodes.size(), 1F);
for (Entry<String, Collection<DataNode>> entry :
actualDataNodes.entrySet()) {
if (featureRequiredSingleTables.contains(entry.getKey())) {
continue;
@@ -125,7 +125,7 @@ public final class SingleRuleConfigurationDecorator
implements RuleConfiguration
private Collection<String> loadSpecifiedTablesWithoutExpand(final boolean
isSchemaSupportedDatabaseType, final Map<String, Collection<DataNode>>
actualDataNodes,
final
Collection<DataNode> configuredDataNodes) {
- Collection<String> result = new LinkedHashSet<>();
+ Collection<String> result = new
LinkedHashSet<>(configuredDataNodes.size(), 1F);
for (DataNode each : configuredDataNodes) {
ShardingSpherePreconditions.checkContainsKey(actualDataNodes,
each.getTableName(), () -> new
SingleTableNotFoundException(getTableNodeString(isSchemaSupportedDatabaseType,
each)));
DataNode actualDataNode =
actualDataNodes.get(each.getTableName()).iterator().next();
diff --git
a/kernel/single/core/src/main/java/org/apache/shardingsphere/single/util/SingleTableLoadUtils.java
b/kernel/single/core/src/main/java/org/apache/shardingsphere/single/util/SingleTableLoadUtils.java
index dc0309e4570..3a14dd9b0ed 100644
---
a/kernel/single/core/src/main/java/org/apache/shardingsphere/single/util/SingleTableLoadUtils.java
+++
b/kernel/single/core/src/main/java/org/apache/shardingsphere/single/util/SingleTableLoadUtils.java
@@ -143,7 +143,7 @@ public final class SingleTableLoadUtils {
* @return data nodes
*/
public static Collection<DataNode> convertToDataNodes(final String
databaseName, final DatabaseType databaseType, final Collection<String> tables)
{
- Collection<DataNode> result = new LinkedHashSet<>();
+ Collection<DataNode> result = new LinkedHashSet<>(tables.size(), 1F);
for (String each : tables) {
result.add(new DataNode(databaseName, databaseType, each));
}
diff --git
a/kernel/single/core/src/main/java/org/apache/shardingsphere/single/yaml/config/pojo/YamlSingleRuleConfiguration.java
b/kernel/single/core/src/main/java/org/apache/shardingsphere/single/yaml/config/pojo/YamlSingleRuleConfiguration.java
index 3ce2a9ab09a..96e765ea2a6 100644
---
a/kernel/single/core/src/main/java/org/apache/shardingsphere/single/yaml/config/pojo/YamlSingleRuleConfiguration.java
+++
b/kernel/single/core/src/main/java/org/apache/shardingsphere/single/yaml/config/pojo/YamlSingleRuleConfiguration.java
@@ -30,7 +30,7 @@ import java.util.LinkedList;
/**
* Single rule configuration for YAML.
*/
-@RepositoryTupleEntity(value = "single")
+@RepositoryTupleEntity("single")
@Getter
@Setter
public final class YamlSingleRuleConfiguration implements
YamlRuleConfiguration {
diff --git
a/kernel/single/distsql/handler/src/main/java/org/apache/shardingsphere/single/distsql/handler/update/LoadSingleTableExecutor.java
b/kernel/single/distsql/handler/src/main/java/org/apache/shardingsphere/single/distsql/handler/update/LoadSingleTableExecutor.java
index 5e9263ba320..216839ac96d 100644
---
a/kernel/single/distsql/handler/src/main/java/org/apache/shardingsphere/single/distsql/handler/update/LoadSingleTableExecutor.java
+++
b/kernel/single/distsql/handler/src/main/java/org/apache/shardingsphere/single/distsql/handler/update/LoadSingleTableExecutor.java
@@ -131,7 +131,7 @@ public final class LoadSingleTableExecutor implements
DatabaseRuleCreateExecutor
}
private Map<String, Map<String, Collection<String>>>
getActualTableNodes(final Collection<String> requiredDataSources, final
Map<String, DataSource> aggregateDataSourceMap) {
- Map<String, Map<String, Collection<String>>> result = new
LinkedHashMap<>();
+ Map<String, Map<String, Collection<String>>> result = new
LinkedHashMap<>(requiredDataSources.size(), 1F);
for (String each : requiredDataSources) {
DataSource dataSource = aggregateDataSourceMap.get(each);
Map<String, Collection<String>> schemaTableNames =
SingleTableDataNodeLoader.loadSchemaTableNames(database.getName(),
DatabaseTypeEngine.getStorageType(dataSource), dataSource, each);
diff --git
a/kernel/sql-federation/core/src/main/java/org/apache/shardingsphere/sqlfederation/resultset/SQLFederationResultSetMetaData.java
b/kernel/sql-federation/core/src/main/java/org/apache/shardingsphere/sqlfederation/resultset/SQLFederationResultSetMetaData.java
index d6a847fc057..0330f32d4dd 100644
---
a/kernel/sql-federation/core/src/main/java/org/apache/shardingsphere/sqlfederation/resultset/SQLFederationResultSetMetaData.java
+++
b/kernel/sql-federation/core/src/main/java/org/apache/shardingsphere/sqlfederation/resultset/SQLFederationResultSetMetaData.java
@@ -54,7 +54,7 @@ public final class SQLFederationResultSetMetaData extends
WrapperAdapter impleme
public SQLFederationResultSetMetaData(final Schema sqlFederationSchema,
final SelectStatementContext selectStatementContext, final RelDataType
resultColumnType,
final Map<Integer, String>
indexAndColumnLabels, final SQLFederationColumnTypeConverter
columnTypeConverter) {
this.sqlFederationSchema = sqlFederationSchema;
- this.relDataTypeFactory = new JavaTypeFactoryImpl();
+ relDataTypeFactory = new JavaTypeFactoryImpl();
this.selectStatementContext = selectStatementContext;
this.resultColumnType = resultColumnType;
this.indexAndColumnLabels = indexAndColumnLabels;
diff --git
a/kernel/sql-federation/distsql/parser/src/main/java/org/apache/shardingsphere/sqlfederation/distsql/parser/core/SQLFederationDistSQLStatementVisitor.java
b/kernel/sql-federation/distsql/parser/src/main/java/org/apache/shardingsphere/sqlfederation/distsql/parser/core/SQLFederationDistSQLStatementVisitor.java
index 5909767f1cf..633ed53c809 100644
---
a/kernel/sql-federation/distsql/parser/src/main/java/org/apache/shardingsphere/sqlfederation/distsql/parser/core/SQLFederationDistSQLStatementVisitor.java
+++
b/kernel/sql-federation/distsql/parser/src/main/java/org/apache/shardingsphere/sqlfederation/distsql/parser/core/SQLFederationDistSQLStatementVisitor.java
@@ -42,7 +42,7 @@ public final class SQLFederationDistSQLStatementVisitor
extends SQLFederationDis
@Override
public ASTNode visitAlterSQLFederationRule(final
AlterSQLFederationRuleContext ctx) {
- return super.visit(ctx.sqlFederationRuleDefinition());
+ return visit(ctx.sqlFederationRuleDefinition());
}
@Override
diff --git
a/kernel/sql-federation/executor/src/main/java/org/apache/shardingsphere/sqlfederation/executor/enumerable/EnumerableScanExecutor.java
b/kernel/sql-federation/executor/src/main/java/org/apache/shardingsphere/sqlfederation/executor/enumerable/EnumerableScanExecutor.java
index cce7fc4e64d..f0921d47722 100644
---
a/kernel/sql-federation/executor/src/main/java/org/apache/shardingsphere/sqlfederation/executor/enumerable/EnumerableScanExecutor.java
+++
b/kernel/sql-federation/executor/src/main/java/org/apache/shardingsphere/sqlfederation/executor/enumerable/EnumerableScanExecutor.java
@@ -217,7 +217,7 @@ public final class EnumerableScanExecutor implements
ScanExecutor {
if (null == paramIndexes) {
return Collections.emptyList();
}
- List<Object> result = new ArrayList<>();
+ List<Object> result = new ArrayList<>(paramIndexes.length);
for (int each : paramIndexes) {
result.add(federationContext.getQueryContext().getParameters().get(each));
}
diff --git
a/kernel/sql-federation/executor/src/main/java/org/apache/shardingsphere/sqlfederation/executor/enumerator/JDBCRowEnumerator.java
b/kernel/sql-federation/executor/src/main/java/org/apache/shardingsphere/sqlfederation/executor/enumerator/JDBCRowEnumerator.java
index 033d1979648..bb26814c93b 100644
---
a/kernel/sql-federation/executor/src/main/java/org/apache/shardingsphere/sqlfederation/executor/enumerator/JDBCRowEnumerator.java
+++
b/kernel/sql-federation/executor/src/main/java/org/apache/shardingsphere/sqlfederation/executor/enumerator/JDBCRowEnumerator.java
@@ -69,7 +69,7 @@ public final class JDBCRowEnumerator implements
Enumerator<Object> {
for (int i = 0; i < metaData.getColumnCount(); i++) {
rowValues[i] = getValue(i);
}
- this.currentRow = 1 == metaData.getColumnCount() ? rowValues[0] :
rowValues;
+ currentRow = 1 == metaData.getColumnCount() ? rowValues[0] : rowValues;
}
private Object getValue(final int index) throws SQLException {
diff --git
a/kernel/sql-federation/optimizer/src/main/java/org/apache/shardingsphere/sqlfederation/optimizer/converter/segment/expression/impl/RowExpressionConverter.java
b/kernel/sql-federation/optimizer/src/main/java/org/apache/shardingsphere/sqlfederation/optimizer/converter/segment/expression/impl/RowExpressionConverter.java
index b2859fb3201..2134bce1fc0 100644
---
a/kernel/sql-federation/optimizer/src/main/java/org/apache/shardingsphere/sqlfederation/optimizer/converter/segment/expression/impl/RowExpressionConverter.java
+++
b/kernel/sql-federation/optimizer/src/main/java/org/apache/shardingsphere/sqlfederation/optimizer/converter/segment/expression/impl/RowExpressionConverter.java
@@ -43,7 +43,7 @@ public final class RowExpressionConverter {
* @return sql node
*/
public static Optional<SqlNode> convert(final RowExpression segment) {
- List<SqlNode> sqlNodes = new ArrayList<>();
+ List<SqlNode> sqlNodes = new ArrayList<>(segment.getItems().size());
for (ExpressionSegment each : segment.getItems()) {
ExpressionConverter.convert(each).ifPresent(sqlNodes::add);
}
diff --git
a/kernel/sql-federation/optimizer/src/main/java/org/apache/shardingsphere/sqlfederation/optimizer/operator/logical/LogicalScan.java
b/kernel/sql-federation/optimizer/src/main/java/org/apache/shardingsphere/sqlfederation/optimizer/operator/logical/LogicalScan.java
index 05b24844403..eaf4c188dc6 100644
---
a/kernel/sql-federation/optimizer/src/main/java/org/apache/shardingsphere/sqlfederation/optimizer/operator/logical/LogicalScan.java
+++
b/kernel/sql-federation/optimizer/src/main/java/org/apache/shardingsphere/sqlfederation/optimizer/operator/logical/LogicalScan.java
@@ -48,7 +48,7 @@ public final class LogicalScan extends TableScan {
}
private void resetRowType(final RelNode relNode) {
- this.rowType = relNode.getRowType();
+ rowType = relNode.getRowType();
}
/**
diff --git
a/kernel/sql-federation/optimizer/src/main/java/org/apache/shardingsphere/sqlfederation/optimizer/planner/util/SQLFederationPlannerUtils.java
b/kernel/sql-federation/optimizer/src/main/java/org/apache/shardingsphere/sqlfederation/optimizer/planner/util/SQLFederationPlannerUtils.java
index 6a7596ab65e..8ffc3f3bab3 100644
---
a/kernel/sql-federation/optimizer/src/main/java/org/apache/shardingsphere/sqlfederation/optimizer/planner/util/SQLFederationPlannerUtils.java
+++
b/kernel/sql-federation/optimizer/src/main/java/org/apache/shardingsphere/sqlfederation/optimizer/planner/util/SQLFederationPlannerUtils.java
@@ -259,6 +259,6 @@ public final class SQLFederationPlannerUtils {
* @return rel opt cluster
*/
public static RelOptCluster createRelOptCluster(final RelDataTypeFactory
relDataTypeFactory) {
- return
RelOptCluster.create(SQLFederationPlannerUtils.createVolcanoPlanner(), new
RexBuilder(relDataTypeFactory));
+ return RelOptCluster.create(createVolcanoPlanner(), new
RexBuilder(relDataTypeFactory));
}
}
diff --git
a/kernel/sql-federation/optimizer/src/test/java/org/apache/shardingsphere/sqlfederation/optimizer/it/SQLStatementCompilerIT.java
b/kernel/sql-federation/optimizer/src/test/java/org/apache/shardingsphere/sqlfederation/optimizer/it/SQLStatementCompilerIT.java
index f7bcbd4aaa8..04b3ba3978e 100644
---
a/kernel/sql-federation/optimizer/src/test/java/org/apache/shardingsphere/sqlfederation/optimizer/it/SQLStatementCompilerIT.java
+++
b/kernel/sql-federation/optimizer/src/test/java/org/apache/shardingsphere/sqlfederation/optimizer/it/SQLStatementCompilerIT.java
@@ -68,7 +68,7 @@ class SQLStatementCompilerIT {
@BeforeEach
void init() {
- Map<String, ShardingSphereTable> tables = new HashMap<>();
+ Map<String, ShardingSphereTable> tables = new HashMap<>(12, 1F);
tables.put("t_order_federate", createOrderFederationTableMetaData());
tables.put("t_user_info", createUserInfoTableMetaData());
tables.put("t_order", createTOrderTableMetaData());
diff --git
a/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/rule/SQLParserRuleTest.java
b/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/rule/SQLParserRuleTest.java
index 55c1b1fc99f..d8897224dbc 100644
---
a/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/rule/SQLParserRuleTest.java
+++
b/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/rule/SQLParserRuleTest.java
@@ -34,7 +34,7 @@ class SQLParserRuleTest {
@BeforeEach
void setup() {
- sqlParserRule = new SQLParserRule(new SQLParserRuleConfiguration(new
CacheOption(2, 4), new CacheOption(3, 7)));
+ sqlParserRule = new SQLParserRule(new SQLParserRuleConfiguration(new
CacheOption(2, 4L), new CacheOption(3, 7L)));
}
@Test
diff --git
a/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/rule/builder/SQLParserRuleBuilderTest.java
b/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/rule/builder/SQLParserRuleBuilderTest.java
index 29746123292..3477814873f 100644
---
a/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/rule/builder/SQLParserRuleBuilderTest.java
+++
b/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/rule/builder/SQLParserRuleBuilderTest.java
@@ -24,7 +24,7 @@ import org.apache.shardingsphere.parser.rule.SQLParserRule;
import org.apache.shardingsphere.sql.parser.api.CacheOption;
import org.junit.jupiter.api.Test;
-import java.util.HashMap;
+import java.util.Collections;
import java.util.Properties;
import static org.hamcrest.MatcherAssert.assertThat;
@@ -34,8 +34,8 @@ class SQLParserRuleBuilderTest {
@Test
void assertBuild() {
- SQLParserRuleConfiguration ruleConfig = new
SQLParserRuleConfiguration(new CacheOption(2, 5), new CacheOption(4, 7));
- SQLParserRule actualResult = new
SQLParserRuleBuilder().build(ruleConfig, new HashMap<>(), new
ConfigurationProperties(new Properties()));
+ SQLParserRuleConfiguration ruleConfig = new
SQLParserRuleConfiguration(new CacheOption(2, 5L), new CacheOption(4, 7L));
+ SQLParserRule actualResult = new
SQLParserRuleBuilder().build(ruleConfig, Collections.emptyMap(), new
ConfigurationProperties(new Properties()));
assertThat(actualResult.getConfiguration(), is(ruleConfig));
assertThat(actualResult.getSqlStatementCache().getInitialCapacity(),
is(4));
assertThat(actualResult.getSqlStatementCache().getMaximumSize(),
is(7L));
diff --git
a/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/yaml/swapper/YamlSQLParserCacheOptionConfigurationSwapperTest.java
b/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/yaml/swapper/YamlSQLParserCacheOptionConfigurationSwapperTest.java
index ec18c970232..2199069f743 100644
---
a/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/yaml/swapper/YamlSQLParserCacheOptionConfigurationSwapperTest.java
+++
b/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/yaml/swapper/YamlSQLParserCacheOptionConfigurationSwapperTest.java
@@ -28,7 +28,7 @@ class YamlSQLParserCacheOptionConfigurationSwapperTest {
@Test
void assertSwapToYamlConfiguration() {
- YamlSQLParserCacheOptionRuleConfiguration actual = new
YamlSQLParserCacheOptionConfigurationSwapper().swapToYamlConfiguration(new
CacheOption(2, 5));
+ YamlSQLParserCacheOptionRuleConfiguration actual = new
YamlSQLParserCacheOptionConfigurationSwapper().swapToYamlConfiguration(new
CacheOption(2, 5L));
assertThat(actual.getInitialCapacity(), is(2));
assertThat(actual.getMaximumSize(), is(5L));
}
diff --git
a/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/yaml/swapper/YamlSQLParserRuleConfigurationSwapperTest.java
b/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/yaml/swapper/YamlSQLParserRuleConfigurationSwapperTest.java
index 7f395ea8f27..0d4d6078f6b 100644
---
a/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/yaml/swapper/YamlSQLParserRuleConfigurationSwapperTest.java
+++
b/kernel/sql-parser/core/src/test/java/org/apache/shardingsphere/parser/yaml/swapper/YamlSQLParserRuleConfigurationSwapperTest.java
@@ -31,7 +31,7 @@ class YamlSQLParserRuleConfigurationSwapperTest {
@Test
void assertSwapToYamlConfiguration() {
YamlSQLParserRuleConfiguration actual =
- new
YamlSQLParserRuleConfigurationSwapper().swapToYamlConfiguration(new
SQLParserRuleConfiguration(new CacheOption(2, 5), new CacheOption(4, 7)));
+ new
YamlSQLParserRuleConfigurationSwapper().swapToYamlConfiguration(new
SQLParserRuleConfiguration(new CacheOption(2, 5L), new CacheOption(4, 7L)));
assertThat(actual.getParseTreeCache().getInitialCapacity(), is(2));
assertThat(actual.getParseTreeCache().getMaximumSize(), is(5L));
assertThat(actual.getSqlStatementCache().getInitialCapacity(), is(4));
diff --git
a/kernel/sql-parser/distsql/handler/src/test/java/org/apache/shardingsphere/parser/distsql/handler/query/ShowSQLParserRuleExecutorTest.java
b/kernel/sql-parser/distsql/handler/src/test/java/org/apache/shardingsphere/parser/distsql/handler/query/ShowSQLParserRuleExecutorTest.java
index 2764e4158ec..7d737c4e046 100644
---
a/kernel/sql-parser/distsql/handler/src/test/java/org/apache/shardingsphere/parser/distsql/handler/query/ShowSQLParserRuleExecutorTest.java
+++
b/kernel/sql-parser/distsql/handler/src/test/java/org/apache/shardingsphere/parser/distsql/handler/query/ShowSQLParserRuleExecutorTest.java
@@ -51,7 +51,7 @@ class ShowSQLParserRuleExecutorTest {
private ContextManager mockContextManager() {
ContextManager result = mock(ContextManager.class, RETURNS_DEEP_STUBS);
SQLParserRule rule = mock(SQLParserRule.class);
- when(rule.getConfiguration()).thenReturn(new
SQLParserRuleConfiguration(new CacheOption(128, 1024), new CacheOption(2000,
65535)));
+ when(rule.getConfiguration()).thenReturn(new
SQLParserRuleConfiguration(new CacheOption(128, 1024L), new CacheOption(2000,
65535L)));
when(result.getMetaDataContexts().getMetaData().getGlobalRuleMetaData().findSingleRule(SQLParserRule.class)).thenReturn(Optional.of(rule));
return result;
}
diff --git
a/kernel/sql-parser/distsql/parser/src/main/java/org/apache/shardingsphere/parser/distsql/parser/core/SQLParserDistSQLStatementVisitor.java
b/kernel/sql-parser/distsql/parser/src/main/java/org/apache/shardingsphere/parser/distsql/parser/core/SQLParserDistSQLStatementVisitor.java
index 348a9592c5f..3d4cb0e03ca 100644
---
a/kernel/sql-parser/distsql/parser/src/main/java/org/apache/shardingsphere/parser/distsql/parser/core/SQLParserDistSQLStatementVisitor.java
+++
b/kernel/sql-parser/distsql/parser/src/main/java/org/apache/shardingsphere/parser/distsql/parser/core/SQLParserDistSQLStatementVisitor.java
@@ -42,7 +42,7 @@ public final class SQLParserDistSQLStatementVisitor extends
SQLParserDistSQLStat
@Override
public ASTNode visitAlterSQLParserRule(final AlterSQLParserRuleContext
ctx) {
- return super.visit(ctx.sqlParserRuleDefinition());
+ return visit(ctx.sqlParserRuleDefinition());
}
@Override
diff --git
a/kernel/transaction/core/src/main/java/org/apache/shardingsphere/transaction/ConnectionTransaction.java
b/kernel/transaction/core/src/main/java/org/apache/shardingsphere/transaction/ConnectionTransaction.java
index 5adf0cc7b07..55425f3153d 100644
---
a/kernel/transaction/core/src/main/java/org/apache/shardingsphere/transaction/ConnectionTransaction.java
+++
b/kernel/transaction/core/src/main/java/org/apache/shardingsphere/transaction/ConnectionTransaction.java
@@ -40,9 +40,9 @@ public final class ConnectionTransaction {
private final TransactionConnectionContext transactionContext;
public ConnectionTransaction(final TransactionRule rule, final
TransactionConnectionContext transactionContext) {
- this.transactionType =
transactionContext.getTransactionType().isPresent() ?
TransactionType.valueOf(transactionContext.getTransactionType().get()) :
rule.getDefaultType();
+ transactionType = transactionContext.getTransactionType().isPresent()
? TransactionType.valueOf(transactionContext.getTransactionType().get()) :
rule.getDefaultType();
this.transactionContext = transactionContext;
- transactionManager = TransactionType.LOCAL == this.transactionType ?
null : rule.getResource().getTransactionManager(rule.getDefaultType());
+ transactionManager = TransactionType.LOCAL == transactionType ? null :
rule.getResource().getTransactionManager(rule.getDefaultType());
}
/**
diff --git
a/kernel/transaction/type/xa/core/src/test/java/org/apache/shardingsphere/transaction/xa/fixture/DataSourceUtils.java
b/kernel/transaction/type/xa/core/src/test/java/org/apache/shardingsphere/transaction/xa/fixture/DataSourceUtils.java
index 4986f7dd3db..ae60d42255c 100644
---
a/kernel/transaction/type/xa/core/src/test/java/org/apache/shardingsphere/transaction/xa/fixture/DataSourceUtils.java
+++
b/kernel/transaction/type/xa/core/src/test/java/org/apache/shardingsphere/transaction/xa/fixture/DataSourceUtils.java
@@ -61,8 +61,8 @@ public final class DataSourceUtils {
result.setPassword("root");
result.setMaximumPoolSize(10);
result.setMinimumIdle(2);
- result.setConnectionTimeout(15 * 1000L);
- result.setIdleTimeout(40 * 1000L);
+ result.setConnectionTimeout(15L * 1000L);
+ result.setIdleTimeout(40L * 1000L);
return result;
}
diff --git
a/mode/core/src/main/java/org/apache/shardingsphere/mode/manager/switcher/ResourceSwitchManager.java
b/mode/core/src/main/java/org/apache/shardingsphere/mode/manager/switcher/ResourceSwitchManager.java
index ae29706d9a1..2431cd90d56 100644
---
a/mode/core/src/main/java/org/apache/shardingsphere/mode/manager/switcher/ResourceSwitchManager.java
+++
b/mode/core/src/main/java/org/apache/shardingsphere/mode/manager/switcher/ResourceSwitchManager.java
@@ -117,7 +117,7 @@ public final class ResourceSwitchManager {
}
private Map<StorageNode, DataSource> getToBeRemovedStaleDataSource(final
ResourceMetaData resourceMetaData, final Collection<String> storageUnitNames) {
- Map<StorageNode, DataSource> result = new LinkedHashMap<>();
+ Map<StorageNode, DataSource> result = new
LinkedHashMap<>(storageUnitNames.size(), 1F);
for (String each : storageUnitNames) {
if (!resourceMetaData.getStorageUnits().containsKey(each)) {
return Collections.emptyMap();
diff --git
a/mode/core/src/main/java/org/apache/shardingsphere/mode/metadata/MetaDataContexts.java
b/mode/core/src/main/java/org/apache/shardingsphere/mode/metadata/MetaDataContexts.java
index dc0e794aa50..c6cdd7a8c59 100644
---
a/mode/core/src/main/java/org/apache/shardingsphere/mode/metadata/MetaDataContexts.java
+++
b/mode/core/src/main/java/org/apache/shardingsphere/mode/metadata/MetaDataContexts.java
@@ -53,7 +53,7 @@ public final class MetaDataContexts implements AutoCloseable {
public MetaDataContexts(final MetaDataBasedPersistService persistService,
final ShardingSphereMetaData metaData) {
this.persistService = persistService;
this.metaData = metaData;
- this.statistics = initStatistics(metaData);
+ statistics = initStatistics(metaData);
}
private ShardingSphereStatistics initStatistics(final
ShardingSphereMetaData metaData) {
diff --git
a/mode/type/cluster/core/src/main/java/org/apache/shardingsphere/mode/manager/cluster/ClusterModeContextManager.java
b/mode/type/cluster/core/src/main/java/org/apache/shardingsphere/mode/manager/cluster/ClusterModeContextManager.java
index 80cfe7ebf0e..2efacde544d 100644
---
a/mode/type/cluster/core/src/main/java/org/apache/shardingsphere/mode/manager/cluster/ClusterModeContextManager.java
+++
b/mode/type/cluster/core/src/main/java/org/apache/shardingsphere/mode/manager/cluster/ClusterModeContextManager.java
@@ -109,7 +109,7 @@ public final class ClusterModeContextManager implements
ModeContextManager, Cont
}
private Map<String, DataSourcePoolProperties>
getToBeDroppedDataSourcePoolProperties(final Map<String,
DataSourcePoolProperties> propsMap, final Collection<String>
toBeDroppedResourceNames) {
- Map<String, DataSourcePoolProperties> result = new LinkedHashMap<>();
+ Map<String, DataSourcePoolProperties> result = new
LinkedHashMap<>(toBeDroppedResourceNames.size(), 1F);
for (String each : toBeDroppedResourceNames) {
if (propsMap.containsKey(each)) {
result.put(each, propsMap.get(each));
diff --git
a/mode/type/cluster/core/src/main/java/org/apache/shardingsphere/mode/manager/cluster/coordinator/registry/status/compute/service/ComputeNodeStatusService.java
b/mode/type/cluster/core/src/main/java/org/apache/shardingsphere/mode/manager/cluster/coordinator/registry/status/compute/service/ComputeNodeStatusService.java
index 199ad54bb7c..96488f881f1 100644
---
a/mode/type/cluster/core/src/main/java/org/apache/shardingsphere/mode/manager/cluster/coordinator/registry/status/compute/service/ComputeNodeStatusService.java
+++
b/mode/type/cluster/core/src/main/java/org/apache/shardingsphere/mode/manager/cluster/coordinator/registry/status/compute/service/ComputeNodeStatusService.java
@@ -172,8 +172,8 @@ public final class ComputeNodeStatusService {
* @return assigned worker ids
*/
public Collection<Integer> getAssignedWorkerIds() {
- Collection<Integer> result = new LinkedHashSet<>();
Collection<String> childrenKeys =
repository.getChildrenKeys(ComputeNode.getInstanceWorkerIdRootNodePath());
+ Collection<Integer> result = new LinkedHashSet<>(childrenKeys.size(),
1F);
for (String each : childrenKeys) {
String workerId =
repository.getDirectly(ComputeNode.getInstanceWorkerIdNodePath(each));
if (null != workerId) {
diff --git
a/mode/type/cluster/core/src/test/java/org/apache/shardingsphere/mode/manager/cluster/coordinator/subscriber/ResourceMetaDataChangedSubscriberTest.java
b/mode/type/cluster/core/src/test/java/org/apache/shardingsphere/mode/manager/cluster/coordinator/subscriber/ResourceMetaDataChangedSubscriberTest.java
index 6ca87662d91..4ec220a0359 100644
---
a/mode/type/cluster/core/src/test/java/org/apache/shardingsphere/mode/manager/cluster/coordinator/subscriber/ResourceMetaDataChangedSubscriberTest.java
+++
b/mode/type/cluster/core/src/test/java/org/apache/shardingsphere/mode/manager/cluster/coordinator/subscriber/ResourceMetaDataChangedSubscriberTest.java
@@ -127,7 +127,7 @@ class ResourceMetaDataChangedSubscriberTest {
@Test
void assertRenewForSchemaAdded() {
subscriber.renew(new SchemaAddedEvent("db", "foo_schema"));
-
verify(contextManager.getMetaDataContexts().getMetaData().getDatabase("db")).addSchema(argThat(argument
-> argument.equals("foo_schema")), any(ShardingSphereSchema.class));
+
verify(contextManager.getMetaDataContexts().getMetaData().getDatabase("db")).addSchema(argThat("foo_schema"::equals),
any(ShardingSphereSchema.class));
}
@Test
diff --git
a/mode/type/cluster/core/src/test/java/org/apache/shardingsphere/mode/manager/cluster/coordinator/subscriber/StateChangedSubscriberTest.java
b/mode/type/cluster/core/src/test/java/org/apache/shardingsphere/mode/manager/cluster/coordinator/subscriber/StateChangedSubscriberTest.java
index 6526dd265ef..71ee8ac0641 100644
---
a/mode/type/cluster/core/src/test/java/org/apache/shardingsphere/mode/manager/cluster/coordinator/subscriber/StateChangedSubscriberTest.java
+++
b/mode/type/cluster/core/src/test/java/org/apache/shardingsphere/mode/manager/cluster/coordinator/subscriber/StateChangedSubscriberTest.java
@@ -117,7 +117,7 @@ class StateChangedSubscriberTest {
subscriber.renew(event);
verify(ruleAttribute).updateStatus(
argThat(qualifiedDataSource ->
Objects.equals(event.getQualifiedDataSource(), qualifiedDataSource)),
- argThat(dataSourceState ->
Objects.equals(event.getStatus().getStatus(), dataSourceState)));
+ argThat(dataSourceState -> event.getStatus().getStatus() ==
dataSourceState));
}
@Test
diff --git
a/mode/type/cluster/repository/api/src/main/java/org/apache/shardingsphere/mode/repository/cluster/lock/holder/DistributedLockHolder.java
b/mode/type/cluster/repository/api/src/main/java/org/apache/shardingsphere/mode/repository/cluster/lock/holder/DistributedLockHolder.java
index 00c169180f6..4c1f99faada 100644
---
a/mode/type/cluster/repository/api/src/main/java/org/apache/shardingsphere/mode/repository/cluster/lock/holder/DistributedLockHolder.java
+++
b/mode/type/cluster/repository/api/src/main/java/org/apache/shardingsphere/mode/repository/cluster/lock/holder/DistributedLockHolder.java
@@ -36,14 +36,13 @@ public final class DistributedLockHolder {
private final TypedProperties<?> props;
- private final Map<String, DistributedLock> locks;
+ private final Map<String, DistributedLock> locks = new
ConcurrentHashMap<>();
@SuppressWarnings("unchecked")
public DistributedLockHolder(final String type, final Object client, final
TypedProperties<?> props) {
creator = TypedSPILoader.getService(DistributedLockCreator.class,
type);
this.client = client;
this.props = props;
- locks = new ConcurrentHashMap<>();
}
/**
diff --git
a/mode/type/cluster/repository/provider/zookeeper/src/main/java/org/apache/shardingsphere/mode/repository/cluster/zookeeper/listener/SessionConnectionListener.java
b/mode/type/cluster/repository/provider/zookeeper/src/main/java/org/apache/shardingsphere/mode/repository/cluster/zookeeper/listener/SessionConnectionListener.java
index a4288ae55cf..a03783f2be8 100644
---
a/mode/type/cluster/repository/provider/zookeeper/src/main/java/org/apache/shardingsphere/mode/repository/cluster/zookeeper/listener/SessionConnectionListener.java
+++
b/mode/type/cluster/repository/provider/zookeeper/src/main/java/org/apache/shardingsphere/mode/repository/cluster/zookeeper/listener/SessionConnectionListener.java
@@ -39,7 +39,7 @@ import java.util.Properties;
@Slf4j
public final class SessionConnectionListener implements
ConnectionStateListener {
- private static final int RECONNECT_INTERVAL_SECONDS = 5;
+ private static final long RECONNECT_INTERVAL_SECONDS = 5L;
private final InstanceContext instanceContext;
diff --git
a/mode/type/standalone/core/src/main/java/org/apache/shardingsphere/mode/manager/standalone/StandaloneModeContextManager.java
b/mode/type/standalone/core/src/main/java/org/apache/shardingsphere/mode/manager/standalone/StandaloneModeContextManager.java
index 41042cf9ee3..4e9d42fc673 100644
---
a/mode/type/standalone/core/src/main/java/org/apache/shardingsphere/mode/manager/standalone/StandaloneModeContextManager.java
+++
b/mode/type/standalone/core/src/main/java/org/apache/shardingsphere/mode/manager/standalone/StandaloneModeContextManager.java
@@ -177,7 +177,7 @@ public final class StandaloneModeContextManager implements
ModeContextManager, C
@Override
public void dropSchema(final String databaseName, final Collection<String>
schemaNames) {
Collection<String> tobeRemovedTables = new LinkedHashSet<>();
- Collection<String> tobeRemovedSchemas = new LinkedHashSet<>();
+ Collection<String> tobeRemovedSchemas = new
LinkedHashSet<>(schemaNames.size(), 1F);
ShardingSphereMetaData metaData =
contextManager.getMetaDataContexts().getMetaData();
ShardingSphereDatabase database = metaData.getDatabase(databaseName);
for (String each : schemaNames) {