This is an automated email from the ASF dual-hosted git repository.

sunnianjun pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/shardingsphere.git


The following commit(s) were added to refs/heads/master by this push:
     new 01b5f54e81c Rename SOURCE_TABLE_NAME and TARGET_TABLE_NAME (#24602)
01b5f54e81c is described below

commit 01b5f54e81c218da82d8de3176ea12feb4aaea51
Author: Liang Zhang <[email protected]>
AuthorDate: Wed Mar 15 06:15:59 2023 +0800

    Rename SOURCE_TABLE_NAME and TARGET_TABLE_NAME (#24602)
---
 .../test/e2e/data/pipeline/cases/cdc/CDCE2EIT.java | 16 ++++++++--------
 .../general/MySQLMigrationGeneralE2EIT.java        | 14 +++++++-------
 .../general/PostgreSQLMigrationGeneralE2EIT.java   | 22 +++++++++++-----------
 .../migration/general/RulesMigrationE2EIT.java     | 12 ++++++------
 .../primarykey/IndexesMigrationE2EIT.java          | 12 ++++++------
 .../primarykey/MariaDBMigrationE2EIT.java          | 12 ++++++------
 .../primarykey/TextPrimaryKeyMigrationE2EIT.java   | 12 ++++++------
 7 files changed, 50 insertions(+), 50 deletions(-)

diff --git 
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/cdc/CDCE2EIT.java
 
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/cdc/CDCE2EIT.java
index 6326cfc5abf..3cb539f26b3 100644
--- 
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/cdc/CDCE2EIT.java
+++ 
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/cdc/CDCE2EIT.java
@@ -89,7 +89,7 @@ public final class CDCE2EIT {
             + "KEY_GENERATE_STRATEGY(COLUMN=order_id,TYPE(NAME='snowflake'))"
             + ")", PipelineContainerComposer.DS_0, 
PipelineContainerComposer.DS_1);
     
-    private static final String SOURCE_TABLE_ORDER_NAME = "t_order";
+    private static final String SOURCE_TABLE_NAME = "t_order";
     
     private final PipelineContainerComposer containerComposer;
     
@@ -135,7 +135,7 @@ public final class CDCE2EIT {
         DataSource jdbcDataSource = 
containerComposer.generateShardingSphereDataSourceFromProxy();
         Pair<List<Object[]>, List<Object[]>> dataPair = 
PipelineCaseHelper.generateFullInsertData(containerComposer.getDatabaseType(), 
20);
         log.info("init data begin: {}", LocalDateTime.now());
-        DataSourceExecuteUtil.execute(jdbcDataSource, 
containerComposer.getExtraSQLCommand().getFullInsertOrder(SOURCE_TABLE_ORDER_NAME),
 dataPair.getLeft());
+        DataSourceExecuteUtil.execute(jdbcDataSource, 
containerComposer.getExtraSQLCommand().getFullInsertOrder(SOURCE_TABLE_NAME), 
dataPair.getLeft());
         log.info("init data end: {}", LocalDateTime.now());
         try (
                 Connection connection = 
DriverManager.getConnection(containerComposer.getActualJdbcUrlTemplate(PipelineContainerComposer.DS_4,
 false),
@@ -146,7 +146,7 @@ public final class CDCE2EIT {
         Awaitility.await().atMost(10, TimeUnit.SECONDS).pollInterval(1, 
TimeUnit.SECONDS).until(() -> !containerComposer.queryForListWithLog("SHOW 
STREAMING LIST").isEmpty());
         String jobId = containerComposer.queryForListWithLog("SHOW STREAMING 
LIST").get(0).get("id").toString();
         containerComposer.waitIncrementTaskFinished(String.format("SHOW 
STREAMING STATUS '%s'", jobId));
-        containerComposer.startIncrementTask(new 
E2EIncrementalTask(jdbcDataSource, SOURCE_TABLE_ORDER_NAME, new 
SnowflakeKeyGenerateAlgorithm(), containerComposer.getDatabaseType(), 20));
+        containerComposer.startIncrementTask(new 
E2EIncrementalTask(jdbcDataSource, SOURCE_TABLE_NAME, new 
SnowflakeKeyGenerateAlgorithm(), containerComposer.getDatabaseType(), 20));
         containerComposer.getIncreaseTaskThread().join(10000L);
         List<Map<String, Object>> actualProxyList;
         try (Connection connection = jdbcDataSource.getConnection()) {
@@ -155,8 +155,8 @@ public final class CDCE2EIT {
         }
         Awaitility.await().atMost(20, TimeUnit.SECONDS).pollInterval(2, 
TimeUnit.SECONDS).until(() -> 
listOrderRecords(getOrderTableNameWithSchema()).size() == 
actualProxyList.size());
         SchemaTableName schemaTableName = 
containerComposer.getDatabaseType().isSchemaAvailable()
-                ? new SchemaTableName(new 
SchemaName(PipelineContainerComposer.SCHEMA_NAME), new 
TableName(SOURCE_TABLE_ORDER_NAME))
-                : new SchemaTableName(new SchemaName(null), new 
TableName(SOURCE_TABLE_ORDER_NAME));
+                ? new SchemaTableName(new 
SchemaName(PipelineContainerComposer.SCHEMA_NAME), new 
TableName(SOURCE_TABLE_NAME))
+                : new SchemaTableName(new SchemaName(null), new 
TableName(SOURCE_TABLE_NAME));
         PipelineDataSourceWrapper targetDataSource = new 
PipelineDataSourceWrapper(StorageContainerUtil.generateDataSource(
                 
containerComposer.getActualJdbcUrlTemplate(PipelineContainerComposer.DS_4, 
false),
                 containerComposer.getUsername(), 
containerComposer.getPassword()), containerComposer.getDatabaseType());
@@ -177,7 +177,7 @@ public final class CDCE2EIT {
     
     private void initSchemaAndTable(final Connection connection, final int 
sleepSeconds) throws SQLException {
         containerComposer.createSchema(connection, sleepSeconds);
-        String sql = 
containerComposer.getExtraSQLCommand().getCreateTableOrder(SOURCE_TABLE_ORDER_NAME);
+        String sql = 
containerComposer.getExtraSQLCommand().getCreateTableOrder(SOURCE_TABLE_NAME);
         log.info("create table sql: {}", sql);
         connection.createStatement().execute(sql);
         if (sleepSeconds > 0) {
@@ -197,7 +197,7 @@ public final class CDCE2EIT {
         // TODO add full=false test case later
         parameter.setFull(true);
         String schema = 
containerComposer.getDatabaseType().isSchemaAvailable() ? "test" : "";
-        
parameter.setSchemaTables(Collections.singletonList(SchemaTable.newBuilder().setTable(SOURCE_TABLE_ORDER_NAME).setSchema(schema).build()));
+        
parameter.setSchemaTables(Collections.singletonList(SchemaTable.newBuilder().setTable(SOURCE_TABLE_NAME).setSchema(schema).build()));
         
parameter.setDatabaseType(containerComposer.getDatabaseType().getType());
         CompletableFuture.runAsync(() -> new CDCClient(parameter).start(), 
executor).whenComplete((unused, throwable) -> {
             if (null != throwable) {
@@ -216,6 +216,6 @@ public final class CDCE2EIT {
     }
     
     private String getOrderTableNameWithSchema() {
-        return containerComposer.getDatabaseType().isSchemaAvailable() ? 
String.join(".", PipelineContainerComposer.SCHEMA_NAME, 
SOURCE_TABLE_ORDER_NAME) : SOURCE_TABLE_ORDER_NAME;
+        return containerComposer.getDatabaseType().isSchemaAvailable() ? 
String.join(".", PipelineContainerComposer.SCHEMA_NAME, SOURCE_TABLE_NAME) : 
SOURCE_TABLE_NAME;
     }
 }
diff --git 
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/general/MySQLMigrationGeneralE2EIT.java
 
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/general/MySQLMigrationGeneralE2EIT.java
index 9b762db9da6..7097a152d65 100644
--- 
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/general/MySQLMigrationGeneralE2EIT.java
+++ 
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/general/MySQLMigrationGeneralE2EIT.java
@@ -52,9 +52,9 @@ import static org.junit.jupiter.api.Assertions.assertTrue;
 @Slf4j
 public final class MySQLMigrationGeneralE2EIT extends AbstractMigrationE2EIT {
     
-    private static final String SOURCE_TABLE_ORDER_NAME = "t_order_copy";
+    private static final String SOURCE_TABLE_NAME = "t_order_copy";
     
-    private static final String TARGET_TABLE_ORDER_NAME = "t_order";
+    private static final String TARGET_TABLE_NAME = "t_order";
     
     public MySQLMigrationGeneralE2EIT(final PipelineTestParameter testParam) {
         super(testParam, new MigrationJobType());
@@ -76,7 +76,7 @@ public final class MySQLMigrationGeneralE2EIT extends 
AbstractMigrationE2EIT {
     @Test
     public void assertMigrationSuccess() throws SQLException, 
InterruptedException {
         addMigrationProcessConfig();
-        getContainerComposer().createSourceOrderTable(SOURCE_TABLE_ORDER_NAME);
+        getContainerComposer().createSourceOrderTable(SOURCE_TABLE_NAME);
         getContainerComposer().createSourceOrderItemTable();
         addMigrationSourceResource();
         addMigrationTargetResource();
@@ -85,15 +85,15 @@ public final class MySQLMigrationGeneralE2EIT extends 
AbstractMigrationE2EIT {
         createTargetOrderItemTableRule();
         Pair<List<Object[]>, List<Object[]>> dataPair = 
PipelineCaseHelper.generateFullInsertData(getContainerComposer().getDatabaseType(),
 PipelineContainerComposer.TABLE_INIT_ROW_COUNT);
         log.info("init data begin: {}", LocalDateTime.now());
-        
DataSourceExecuteUtil.execute(getContainerComposer().getSourceDataSource(), 
getContainerComposer().getExtraSQLCommand().getFullInsertOrder(SOURCE_TABLE_ORDER_NAME),
 dataPair.getLeft());
+        
DataSourceExecuteUtil.execute(getContainerComposer().getSourceDataSource(), 
getContainerComposer().getExtraSQLCommand().getFullInsertOrder(SOURCE_TABLE_NAME),
 dataPair.getLeft());
         
DataSourceExecuteUtil.execute(getContainerComposer().getSourceDataSource(), 
getContainerComposer().getExtraSQLCommand().getFullInsertOrderItem(), 
dataPair.getRight());
         log.info("init data end: {}", LocalDateTime.now());
-        startMigration(SOURCE_TABLE_ORDER_NAME, TARGET_TABLE_ORDER_NAME);
+        startMigration(SOURCE_TABLE_NAME, TARGET_TABLE_NAME);
         startMigration("t_order_item", "t_order_item");
-        String orderJobId = getJobIdByTableName("ds_0." + 
SOURCE_TABLE_ORDER_NAME);
+        String orderJobId = getJobIdByTableName("ds_0." + SOURCE_TABLE_NAME);
         getContainerComposer().waitJobPrepareSuccess(String.format("SHOW 
MIGRATION STATUS '%s'", orderJobId));
         getContainerComposer().startIncrementTask(
-                new 
E2EIncrementalTask(getContainerComposer().getSourceDataSource(), 
SOURCE_TABLE_ORDER_NAME, new SnowflakeKeyGenerateAlgorithm(), 
getContainerComposer().getDatabaseType(), 30));
+                new 
E2EIncrementalTask(getContainerComposer().getSourceDataSource(), 
SOURCE_TABLE_NAME, new SnowflakeKeyGenerateAlgorithm(), 
getContainerComposer().getDatabaseType(), 30));
         assertMigrationSuccessById(orderJobId, "DATA_MATCH");
         String orderItemJobId = getJobIdByTableName("ds_0.t_order_item");
         assertMigrationSuccessById(orderItemJobId, "DATA_MATCH");
diff --git 
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/general/PostgreSQLMigrationGeneralE2EIT.java
 
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/general/PostgreSQLMigrationGeneralE2EIT.java
index 031ba2f4103..d966f9c2eaf 100644
--- 
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/general/PostgreSQLMigrationGeneralE2EIT.java
+++ 
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/general/PostgreSQLMigrationGeneralE2EIT.java
@@ -50,9 +50,9 @@ import static org.junit.jupiter.api.Assertions.assertTrue;
 @Slf4j
 public final class PostgreSQLMigrationGeneralE2EIT extends 
AbstractMigrationE2EIT {
     
-    private static final String SOURCE_TABLE_ORDER_NAME = "t_order_copy";
+    private static final String SOURCE_TABLE_NAME = "t_order_copy";
     
-    private static final String TARGET_TABLE_ORDER_NAME = "t_order";
+    private static final String TARGET_TABLE_NAME = "t_order";
     
     public PostgreSQLMigrationGeneralE2EIT(final PipelineTestParameter 
testParam) {
         super(testParam, new MigrationJobType());
@@ -77,25 +77,25 @@ public final class PostgreSQLMigrationGeneralE2EIT extends 
AbstractMigrationE2EI
     public void assertMigrationSuccess() throws SQLException, 
InterruptedException {
         addMigrationProcessConfig();
         createSourceSchema(PipelineContainerComposer.SCHEMA_NAME);
-        getContainerComposer().createSourceOrderTable(SOURCE_TABLE_ORDER_NAME);
+        getContainerComposer().createSourceOrderTable(SOURCE_TABLE_NAME);
         getContainerComposer().createSourceOrderItemTable();
-        
getContainerComposer().createSourceTableIndexList(PipelineContainerComposer.SCHEMA_NAME,
 SOURCE_TABLE_ORDER_NAME);
-        
getContainerComposer().createSourceCommentOnList(PipelineContainerComposer.SCHEMA_NAME,
 SOURCE_TABLE_ORDER_NAME);
+        
getContainerComposer().createSourceTableIndexList(PipelineContainerComposer.SCHEMA_NAME,
 SOURCE_TABLE_NAME);
+        
getContainerComposer().createSourceCommentOnList(PipelineContainerComposer.SCHEMA_NAME,
 SOURCE_TABLE_NAME);
         addMigrationSourceResource();
         addMigrationTargetResource();
         createTargetOrderTableRule();
         createTargetOrderItemTableRule();
         Pair<List<Object[]>, List<Object[]>> dataPair = 
PipelineCaseHelper.generateFullInsertData(getContainerComposer().getDatabaseType(),
 PipelineContainerComposer.TABLE_INIT_ROW_COUNT);
         log.info("init data begin: {}", LocalDateTime.now());
-        
DataSourceExecuteUtil.execute(getContainerComposer().getSourceDataSource(), 
getContainerComposer().getExtraSQLCommand().getFullInsertOrder(SOURCE_TABLE_ORDER_NAME),
 dataPair.getLeft());
+        
DataSourceExecuteUtil.execute(getContainerComposer().getSourceDataSource(), 
getContainerComposer().getExtraSQLCommand().getFullInsertOrder(SOURCE_TABLE_NAME),
 dataPair.getLeft());
         
DataSourceExecuteUtil.execute(getContainerComposer().getSourceDataSource(), 
getContainerComposer().getExtraSQLCommand().getFullInsertOrderItem(), 
dataPair.getRight());
         log.info("init data end: {}", LocalDateTime.now());
-        startMigrationWithSchema(SOURCE_TABLE_ORDER_NAME, "t_order");
+        startMigrationWithSchema(SOURCE_TABLE_NAME, "t_order");
         Awaitility.await().atMost(10, TimeUnit.SECONDS).pollInterval(1, 
TimeUnit.SECONDS).until(() -> listJobId().size() > 0);
-        String jobId = getJobIdByTableName("ds_0.test." + 
SOURCE_TABLE_ORDER_NAME);
+        String jobId = getJobIdByTableName("ds_0.test." + SOURCE_TABLE_NAME);
         getContainerComposer().waitIncrementTaskFinished(String.format("SHOW 
MIGRATION STATUS '%s'", jobId));
         getContainerComposer().startIncrementTask(new E2EIncrementalTask(
-                getContainerComposer().getSourceDataSource(), String.join(".", 
PipelineContainerComposer.SCHEMA_NAME, SOURCE_TABLE_ORDER_NAME),
+                getContainerComposer().getSourceDataSource(), String.join(".", 
PipelineContainerComposer.SCHEMA_NAME, SOURCE_TABLE_NAME),
                 new SnowflakeKeyGenerateAlgorithm(), 
getContainerComposer().getDatabaseType(), 20));
         checkOrderMigration(jobId);
         checkOrderItemMigration();
@@ -113,11 +113,11 @@ public final class PostgreSQLMigrationGeneralE2EIT 
extends AbstractMigrationE2EI
         stopMigrationByJobId(jobId);
         long recordId = new SnowflakeKeyGenerateAlgorithm().generateKey();
         getContainerComposer().sourceExecuteWithLog(String.format("INSERT INTO 
%s (order_id,user_id,status) VALUES (%s, %s, '%s')",
-                String.join(".", PipelineContainerComposer.SCHEMA_NAME, 
SOURCE_TABLE_ORDER_NAME), recordId, 1, "afterStop"));
+                String.join(".", PipelineContainerComposer.SCHEMA_NAME, 
SOURCE_TABLE_NAME), recordId, 1, "afterStop"));
         startMigrationByJobId(jobId);
         // must refresh firstly, otherwise proxy can't get schema and table 
info
         getContainerComposer().proxyExecuteWithLog("REFRESH TABLE METADATA;", 
2);
-        getContainerComposer().assertProxyOrderRecordExist(String.join(".", 
PipelineContainerComposer.SCHEMA_NAME, TARGET_TABLE_ORDER_NAME), recordId);
+        getContainerComposer().assertProxyOrderRecordExist(String.join(".", 
PipelineContainerComposer.SCHEMA_NAME, TARGET_TABLE_NAME), recordId);
         assertCheckMigrationSuccess(jobId, "DATA_MATCH");
     }
     
diff --git 
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/general/RulesMigrationE2EIT.java
 
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/general/RulesMigrationE2EIT.java
index 0ac8a5d5126..afb97fe244e 100644
--- 
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/general/RulesMigrationE2EIT.java
+++ 
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/general/RulesMigrationE2EIT.java
@@ -50,9 +50,9 @@ import static org.hamcrest.Matchers.is;
 @Slf4j
 public final class RulesMigrationE2EIT extends AbstractMigrationE2EIT {
     
-    private static final String SOURCE_TABLE_ORDER_NAME = "t_order";
+    private static final String SOURCE_TABLE_NAME = "t_order";
     
-    private static final String TARGET_TABLE_ORDER_NAME = "t_order";
+    private static final String TARGET_TABLE_NAME = "t_order";
     
     public RulesMigrationE2EIT(final PipelineTestParameter testParam) {
         super(testParam, new MigrationJobType());
@@ -86,22 +86,22 @@ public final class RulesMigrationE2EIT extends 
AbstractMigrationE2EIT {
     }
     
     private void assertMigrationSuccess(final Callable<Void> addRuleFn) throws 
Exception {
-        getContainerComposer().createSourceOrderTable(SOURCE_TABLE_ORDER_NAME);
+        getContainerComposer().createSourceOrderTable(SOURCE_TABLE_NAME);
         try (Connection connection = 
getContainerComposer().getSourceDataSource().getConnection()) {
-            
PipelineCaseHelper.batchInsertOrderRecordsWithGeneralColumns(connection, new 
UUIDKeyGenerateAlgorithm(), SOURCE_TABLE_ORDER_NAME, 
PipelineContainerComposer.TABLE_INIT_ROW_COUNT);
+            
PipelineCaseHelper.batchInsertOrderRecordsWithGeneralColumns(connection, new 
UUIDKeyGenerateAlgorithm(), SOURCE_TABLE_NAME, 
PipelineContainerComposer.TABLE_INIT_ROW_COUNT);
         }
         addMigrationSourceResource();
         addMigrationTargetResource();
         if (null != addRuleFn) {
             addRuleFn.call();
         }
-        startMigration(SOURCE_TABLE_ORDER_NAME, TARGET_TABLE_ORDER_NAME);
+        startMigration(SOURCE_TABLE_NAME, TARGET_TABLE_NAME);
         String jobId = listJobId().get(0);
         getContainerComposer().waitJobPrepareSuccess(String.format("SHOW 
MIGRATION STATUS '%s'", jobId));
         getContainerComposer().waitIncrementTaskFinished(String.format("SHOW 
MIGRATION STATUS '%s'", jobId));
         assertCheckMigrationSuccess(jobId, "DATA_MATCH");
         commitMigrationByJobId(jobId);
         getContainerComposer().proxyExecuteWithLog("REFRESH TABLE METADATA", 
1);
-        
assertThat(getContainerComposer().getTargetTableRecordsCount(SOURCE_TABLE_ORDER_NAME),
 is(PipelineContainerComposer.TABLE_INIT_ROW_COUNT));
+        
assertThat(getContainerComposer().getTargetTableRecordsCount(SOURCE_TABLE_NAME),
 is(PipelineContainerComposer.TABLE_INIT_ROW_COUNT));
     }
 }
diff --git 
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/primarykey/IndexesMigrationE2EIT.java
 
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/primarykey/IndexesMigrationE2EIT.java
index 421f91f2f9b..8accc74d08f 100644
--- 
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/primarykey/IndexesMigrationE2EIT.java
+++ 
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/primarykey/IndexesMigrationE2EIT.java
@@ -65,9 +65,9 @@ public final class IndexesMigrationE2EIT extends 
AbstractMigrationE2EIT {
             + "TYPE(NAME=\"hash_mod\",PROPERTIES(\"sharding-count\"=\"6\"))\n"
             + ");";
     
-    private static final String SOURCE_TABLE_ORDER_NAME = "t_order";
+    private static final String SOURCE_TABLE_NAME = "t_order";
     
-    private static final String TARGET_TABLE_ORDER_NAME = "t_order";
+    private static final String TARGET_TABLE_NAME = "t_order";
     
     public IndexesMigrationE2EIT(final PipelineTestParameter testParam) {
         super(testParam, new MigrationJobType());
@@ -185,15 +185,15 @@ public final class IndexesMigrationE2EIT extends 
AbstractMigrationE2EIT {
     
     private void assertMigrationSuccess(final String sqlPattern, final String 
shardingColumn, final KeyGenerateAlgorithm keyGenerateAlgorithm,
                                         final String 
consistencyCheckAlgorithmType, final Callable<Void> incrementalTaskFn) throws 
Exception {
-        getContainerComposer().sourceExecuteWithLog(String.format(sqlPattern, 
SOURCE_TABLE_ORDER_NAME));
+        getContainerComposer().sourceExecuteWithLog(String.format(sqlPattern, 
SOURCE_TABLE_NAME));
         try (Connection connection = 
getContainerComposer().getSourceDataSource().getConnection()) {
-            
PipelineCaseHelper.batchInsertOrderRecordsWithGeneralColumns(connection, 
keyGenerateAlgorithm, SOURCE_TABLE_ORDER_NAME, 
PipelineContainerComposer.TABLE_INIT_ROW_COUNT);
+            
PipelineCaseHelper.batchInsertOrderRecordsWithGeneralColumns(connection, 
keyGenerateAlgorithm, SOURCE_TABLE_NAME, 
PipelineContainerComposer.TABLE_INIT_ROW_COUNT);
         }
         addMigrationProcessConfig();
         addMigrationSourceResource();
         addMigrationTargetResource();
         
getContainerComposer().proxyExecuteWithLog(String.format(ORDER_TABLE_SHARDING_RULE_FORMAT,
 shardingColumn), 2);
-        startMigration(SOURCE_TABLE_ORDER_NAME, TARGET_TABLE_ORDER_NAME);
+        startMigration(SOURCE_TABLE_NAME, TARGET_TABLE_NAME);
         String jobId = listJobId().get(0);
         getContainerComposer().waitJobPrepareSuccess(String.format("SHOW 
MIGRATION STATUS '%s'", jobId));
         incrementalTaskFn.call();
@@ -203,7 +203,7 @@ public final class IndexesMigrationE2EIT extends 
AbstractMigrationE2EIT {
         }
         commitMigrationByJobId(jobId);
         getContainerComposer().proxyExecuteWithLog("REFRESH TABLE METADATA", 
1);
-        
assertThat(getContainerComposer().getTargetTableRecordsCount(SOURCE_TABLE_ORDER_NAME),
 is(PipelineContainerComposer.TABLE_INIT_ROW_COUNT + 1));
+        
assertThat(getContainerComposer().getTargetTableRecordsCount(SOURCE_TABLE_NAME),
 is(PipelineContainerComposer.TABLE_INIT_ROW_COUNT + 1));
         List<String> lastJobIds = listJobId();
         assertTrue(lastJobIds.isEmpty());
     }
diff --git 
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/primarykey/MariaDBMigrationE2EIT.java
 
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/primarykey/MariaDBMigrationE2EIT.java
index 3913483748a..2ed54d39654 100644
--- 
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/primarykey/MariaDBMigrationE2EIT.java
+++ 
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/primarykey/MariaDBMigrationE2EIT.java
@@ -47,9 +47,9 @@ import static org.junit.jupiter.api.Assertions.assertTrue;
 @Slf4j
 public final class MariaDBMigrationE2EIT extends AbstractMigrationE2EIT {
     
-    private static final String SOURCE_TABLE_ORDER_NAME = "t_order";
+    private static final String SOURCE_TABLE_NAME = "t_order";
     
-    private static final String TARGET_TABLE_ORDER_NAME = "t_order";
+    private static final String TARGET_TABLE_NAME = "t_order";
     
     public MariaDBMigrationE2EIT(final PipelineTestParameter testParam) {
         super(testParam, new MigrationJobType());
@@ -73,16 +73,16 @@ public final class MariaDBMigrationE2EIT extends 
AbstractMigrationE2EIT {
     @Test
     public void assertMigrationSuccess() throws SQLException, 
InterruptedException {
         String sqlPattern = "CREATE TABLE `%s` (`order_id` VARCHAR(64) NOT 
NULL, `user_id` INT NOT NULL, `status` varchar(255), PRIMARY KEY (`order_id`)) 
ENGINE=InnoDB DEFAULT CHARSET=utf8mb4";
-        getContainerComposer().sourceExecuteWithLog(String.format(sqlPattern, 
SOURCE_TABLE_ORDER_NAME));
+        getContainerComposer().sourceExecuteWithLog(String.format(sqlPattern, 
SOURCE_TABLE_NAME));
         try (Connection connection = 
getContainerComposer().getSourceDataSource().getConnection()) {
             KeyGenerateAlgorithm generateAlgorithm = new 
UUIDKeyGenerateAlgorithm();
-            
PipelineCaseHelper.batchInsertOrderRecordsWithGeneralColumns(connection, 
generateAlgorithm, SOURCE_TABLE_ORDER_NAME, 
PipelineContainerComposer.TABLE_INIT_ROW_COUNT);
+            
PipelineCaseHelper.batchInsertOrderRecordsWithGeneralColumns(connection, 
generateAlgorithm, SOURCE_TABLE_NAME, 
PipelineContainerComposer.TABLE_INIT_ROW_COUNT);
         }
         addMigrationProcessConfig();
         addMigrationSourceResource();
         addMigrationTargetResource();
         createTargetOrderTableRule();
-        startMigration(SOURCE_TABLE_ORDER_NAME, TARGET_TABLE_ORDER_NAME);
+        startMigration(SOURCE_TABLE_NAME, TARGET_TABLE_NAME);
         String jobId = listJobId().get(0);
         getContainerComposer().waitJobPrepareSuccess(String.format("SHOW 
MIGRATION STATUS '%s'", jobId));
         getContainerComposer().sourceExecuteWithLog("INSERT INTO t_order 
(order_id, user_id, status) VALUES ('a1', 1, 'OK')");
@@ -91,7 +91,7 @@ public final class MariaDBMigrationE2EIT extends 
AbstractMigrationE2EIT {
         assertCheckMigrationSuccess(jobId, "CRC32_MATCH");
         commitMigrationByJobId(jobId);
         getContainerComposer().proxyExecuteWithLog("REFRESH TABLE METADATA", 
1);
-        
assertThat(getContainerComposer().getTargetTableRecordsCount(SOURCE_TABLE_ORDER_NAME),
 is(PipelineContainerComposer.TABLE_INIT_ROW_COUNT + 1));
+        
assertThat(getContainerComposer().getTargetTableRecordsCount(SOURCE_TABLE_NAME),
 is(PipelineContainerComposer.TABLE_INIT_ROW_COUNT + 1));
         List<String> lastJobIds = listJobId();
         assertTrue(lastJobIds.isEmpty());
     }
diff --git 
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/primarykey/TextPrimaryKeyMigrationE2EIT.java
 
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/primarykey/TextPrimaryKeyMigrationE2EIT.java
index d257521895d..16561dfdf53 100644
--- 
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/primarykey/TextPrimaryKeyMigrationE2EIT.java
+++ 
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/primarykey/TextPrimaryKeyMigrationE2EIT.java
@@ -47,7 +47,7 @@ import static org.junit.jupiter.api.Assertions.assertTrue;
 @Slf4j
 public class TextPrimaryKeyMigrationE2EIT extends AbstractMigrationE2EIT {
     
-    private static final String TARGET_TABLE_ORDER_NAME = "t_order";
+    private static final String TARGET_TABLE_NAME = "t_order";
     
     public TextPrimaryKeyMigrationE2EIT(final PipelineTestParameter testParam) 
{
         super(testParam, new MigrationJobType());
@@ -73,18 +73,18 @@ public class TextPrimaryKeyMigrationE2EIT extends 
AbstractMigrationE2EIT {
     
     @Test
     public void assertTextPrimaryMigrationSuccess() throws SQLException, 
InterruptedException {
-        
getContainerComposer().createSourceOrderTable(getSourceTableOrderName());
+        getContainerComposer().createSourceOrderTable(getSourceTableName());
         try (Connection connection = 
getContainerComposer().getSourceDataSource().getConnection()) {
             UUIDKeyGenerateAlgorithm keyGenerateAlgorithm = new 
UUIDKeyGenerateAlgorithm();
-            
PipelineCaseHelper.batchInsertOrderRecordsWithGeneralColumns(connection, 
keyGenerateAlgorithm, getSourceTableOrderName(), 
PipelineContainerComposer.TABLE_INIT_ROW_COUNT);
+            
PipelineCaseHelper.batchInsertOrderRecordsWithGeneralColumns(connection, 
keyGenerateAlgorithm, getSourceTableName(), 
PipelineContainerComposer.TABLE_INIT_ROW_COUNT);
         }
         addMigrationProcessConfig();
         addMigrationSourceResource();
         addMigrationTargetResource();
         createTargetOrderTableRule();
-        startMigration(getSourceTableOrderName(), TARGET_TABLE_ORDER_NAME);
+        startMigration(getSourceTableName(), TARGET_TABLE_NAME);
         String jobId = listJobId().get(0);
-        getContainerComposer().sourceExecuteWithLog(String.format("INSERT INTO 
%s (order_id,user_id,status) VALUES (%s, %s, '%s')", getSourceTableOrderName(), 
"1000000000", 1, "afterStop"));
+        getContainerComposer().sourceExecuteWithLog(String.format("INSERT INTO 
%s (order_id,user_id,status) VALUES (%s, %s, '%s')", getSourceTableName(), 
"1000000000", 1, "afterStop"));
         getContainerComposer().waitIncrementTaskFinished(String.format("SHOW 
MIGRATION STATUS '%s'", jobId));
         assertCheckMigrationSuccess(jobId, "DATA_MATCH");
         commitMigrationByJobId(jobId);
@@ -92,7 +92,7 @@ public class TextPrimaryKeyMigrationE2EIT extends 
AbstractMigrationE2EIT {
         assertTrue(lastJobIds.isEmpty());
     }
     
-    private String getSourceTableOrderName() {
+    private String getSourceTableName() {
         return 
DatabaseTypeUtil.isMySQL(getContainerComposer().getDatabaseType()) ? "T_ORDER" 
: "t_order";
     }
 }

Reply via email to