This is an automated email from the ASF dual-hosted git repository.

zhonghongsheng pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/shardingsphere.git


The following commit(s) were added to refs/heads/master by this push:
     new bc325021ba0 small fixes added for sonar (#25717)
bc325021ba0 is described below

commit bc325021ba01377f643850fcfb7973f117b1fe53
Author: Salman <[email protected]>
AuthorDate: Mon May 22 04:53:41 2023 +0200

    small fixes added for sonar (#25717)
    
    * small fixes added for sonar
    
    * fix reverted and a small issue fixed
---
 .../AbstractStreamingDataConsistencyCalculateAlgorithm.java         | 4 ++--
 .../algorithm/DataMatchDataConsistencyCalculateAlgorithm.java       | 6 +++---
 2 files changed, 5 insertions(+), 5 deletions(-)

diff --git 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/check/consistency/algorithm/AbstractStreamingDataConsistencyCalculateAlgorithm.java
 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/check/consistency/algorithm/AbstractStreamingDataConsistencyCalculateAlgorithm.java
index e0a93d76d3f..351810a6560 100644
--- 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/check/consistency/algorithm/AbstractStreamingDataConsistencyCalculateAlgorithm.java
+++ 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/check/consistency/algorithm/AbstractStreamingDataConsistencyCalculateAlgorithm.java
@@ -78,9 +78,9 @@ public abstract class 
AbstractStreamingDataConsistencyCalculateAlgorithm extends
         @Override
         public DataConsistencyCalculatedResult next() {
             calculateIfNecessary();
-            Optional<DataConsistencyCalculatedResult> nextResult = 
this.nextResult.get();
+            Optional<DataConsistencyCalculatedResult> nextCalculatedResult = 
this.nextResult.get();
             this.nextResult.set(null);
-            return nextResult.orElseThrow(NoSuchElementException::new);
+            return 
nextCalculatedResult.orElseThrow(NoSuchElementException::new);
         }
         
         private void calculateIfNecessary() {
diff --git 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/check/consistency/algorithm/DataMatchDataConsistencyCalculateAlgorithm.java
 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/check/consistency/algorithm/DataMatchDataConsistencyCalculateAlgorithm.java
index 01b30ae060f..75f71ddff57 100644
--- 
a/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/check/consistency/algorithm/DataMatchDataConsistencyCalculateAlgorithm.java
+++ 
b/kernel/data-pipeline/core/src/main/java/org/apache/shardingsphere/data/pipeline/core/check/consistency/algorithm/DataMatchDataConsistencyCalculateAlgorithm.java
@@ -106,11 +106,11 @@ public final class 
DataMatchDataConsistencyCalculateAlgorithm extends AbstractSt
                 ShardingSpherePreconditions.checkState(!isCanceling(), () -> 
new 
PipelineTableDataConsistencyCheckLoadingFailedException(param.getSchemaName(), 
param.getLogicTableName()));
                 ResultSetMetaData resultSetMetaData = resultSet.getMetaData();
                 int columnCount = resultSetMetaData.getColumnCount();
-                Collection<Object> record = new LinkedList<>();
+                Collection<Object> columnRecord = new LinkedList<>();
                 for (int columnIndex = 1; columnIndex <= columnCount; 
columnIndex++) {
-                    record.add(columnValueReader.readValue(resultSet, 
resultSetMetaData, columnIndex));
+                    columnRecord.add(columnValueReader.readValue(resultSet, 
resultSetMetaData, columnIndex));
                 }
-                records.add(record);
+                records.add(columnRecord);
                 maxUniqueKeyValue = columnValueReader.readValue(resultSet, 
resultSetMetaData, param.getUniqueKey().getOrdinalPosition());
                 if (records.size() == chunkSize) {
                     break;

Reply via email to