This is an automated email from the ASF dual-hosted git repository.
sunnianjun pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/shardingsphere.git
The following commit(s) were added to refs/heads/master by this push:
new 01b5f54e81c Rename SOURCE_TABLE_NAME and TARGET_TABLE_NAME (#24602)
01b5f54e81c is described below
commit 01b5f54e81c218da82d8de3176ea12feb4aaea51
Author: Liang Zhang <[email protected]>
AuthorDate: Wed Mar 15 06:15:59 2023 +0800
Rename SOURCE_TABLE_NAME and TARGET_TABLE_NAME (#24602)
---
.../test/e2e/data/pipeline/cases/cdc/CDCE2EIT.java | 16 ++++++++--------
.../general/MySQLMigrationGeneralE2EIT.java | 14 +++++++-------
.../general/PostgreSQLMigrationGeneralE2EIT.java | 22 +++++++++++-----------
.../migration/general/RulesMigrationE2EIT.java | 12 ++++++------
.../primarykey/IndexesMigrationE2EIT.java | 12 ++++++------
.../primarykey/MariaDBMigrationE2EIT.java | 12 ++++++------
.../primarykey/TextPrimaryKeyMigrationE2EIT.java | 12 ++++++------
7 files changed, 50 insertions(+), 50 deletions(-)
diff --git
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/cdc/CDCE2EIT.java
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/cdc/CDCE2EIT.java
index 6326cfc5abf..3cb539f26b3 100644
---
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/cdc/CDCE2EIT.java
+++
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/cdc/CDCE2EIT.java
@@ -89,7 +89,7 @@ public final class CDCE2EIT {
+ "KEY_GENERATE_STRATEGY(COLUMN=order_id,TYPE(NAME='snowflake'))"
+ ")", PipelineContainerComposer.DS_0,
PipelineContainerComposer.DS_1);
- private static final String SOURCE_TABLE_ORDER_NAME = "t_order";
+ private static final String SOURCE_TABLE_NAME = "t_order";
private final PipelineContainerComposer containerComposer;
@@ -135,7 +135,7 @@ public final class CDCE2EIT {
DataSource jdbcDataSource =
containerComposer.generateShardingSphereDataSourceFromProxy();
Pair<List<Object[]>, List<Object[]>> dataPair =
PipelineCaseHelper.generateFullInsertData(containerComposer.getDatabaseType(),
20);
log.info("init data begin: {}", LocalDateTime.now());
- DataSourceExecuteUtil.execute(jdbcDataSource,
containerComposer.getExtraSQLCommand().getFullInsertOrder(SOURCE_TABLE_ORDER_NAME),
dataPair.getLeft());
+ DataSourceExecuteUtil.execute(jdbcDataSource,
containerComposer.getExtraSQLCommand().getFullInsertOrder(SOURCE_TABLE_NAME),
dataPair.getLeft());
log.info("init data end: {}", LocalDateTime.now());
try (
Connection connection =
DriverManager.getConnection(containerComposer.getActualJdbcUrlTemplate(PipelineContainerComposer.DS_4,
false),
@@ -146,7 +146,7 @@ public final class CDCE2EIT {
Awaitility.await().atMost(10, TimeUnit.SECONDS).pollInterval(1,
TimeUnit.SECONDS).until(() -> !containerComposer.queryForListWithLog("SHOW
STREAMING LIST").isEmpty());
String jobId = containerComposer.queryForListWithLog("SHOW STREAMING
LIST").get(0).get("id").toString();
containerComposer.waitIncrementTaskFinished(String.format("SHOW
STREAMING STATUS '%s'", jobId));
- containerComposer.startIncrementTask(new
E2EIncrementalTask(jdbcDataSource, SOURCE_TABLE_ORDER_NAME, new
SnowflakeKeyGenerateAlgorithm(), containerComposer.getDatabaseType(), 20));
+ containerComposer.startIncrementTask(new
E2EIncrementalTask(jdbcDataSource, SOURCE_TABLE_NAME, new
SnowflakeKeyGenerateAlgorithm(), containerComposer.getDatabaseType(), 20));
containerComposer.getIncreaseTaskThread().join(10000L);
List<Map<String, Object>> actualProxyList;
try (Connection connection = jdbcDataSource.getConnection()) {
@@ -155,8 +155,8 @@ public final class CDCE2EIT {
}
Awaitility.await().atMost(20, TimeUnit.SECONDS).pollInterval(2,
TimeUnit.SECONDS).until(() ->
listOrderRecords(getOrderTableNameWithSchema()).size() ==
actualProxyList.size());
SchemaTableName schemaTableName =
containerComposer.getDatabaseType().isSchemaAvailable()
- ? new SchemaTableName(new
SchemaName(PipelineContainerComposer.SCHEMA_NAME), new
TableName(SOURCE_TABLE_ORDER_NAME))
- : new SchemaTableName(new SchemaName(null), new
TableName(SOURCE_TABLE_ORDER_NAME));
+ ? new SchemaTableName(new
SchemaName(PipelineContainerComposer.SCHEMA_NAME), new
TableName(SOURCE_TABLE_NAME))
+ : new SchemaTableName(new SchemaName(null), new
TableName(SOURCE_TABLE_NAME));
PipelineDataSourceWrapper targetDataSource = new
PipelineDataSourceWrapper(StorageContainerUtil.generateDataSource(
containerComposer.getActualJdbcUrlTemplate(PipelineContainerComposer.DS_4,
false),
containerComposer.getUsername(),
containerComposer.getPassword()), containerComposer.getDatabaseType());
@@ -177,7 +177,7 @@ public final class CDCE2EIT {
private void initSchemaAndTable(final Connection connection, final int
sleepSeconds) throws SQLException {
containerComposer.createSchema(connection, sleepSeconds);
- String sql =
containerComposer.getExtraSQLCommand().getCreateTableOrder(SOURCE_TABLE_ORDER_NAME);
+ String sql =
containerComposer.getExtraSQLCommand().getCreateTableOrder(SOURCE_TABLE_NAME);
log.info("create table sql: {}", sql);
connection.createStatement().execute(sql);
if (sleepSeconds > 0) {
@@ -197,7 +197,7 @@ public final class CDCE2EIT {
// TODO add full=false test case later
parameter.setFull(true);
String schema =
containerComposer.getDatabaseType().isSchemaAvailable() ? "test" : "";
-
parameter.setSchemaTables(Collections.singletonList(SchemaTable.newBuilder().setTable(SOURCE_TABLE_ORDER_NAME).setSchema(schema).build()));
+
parameter.setSchemaTables(Collections.singletonList(SchemaTable.newBuilder().setTable(SOURCE_TABLE_NAME).setSchema(schema).build()));
parameter.setDatabaseType(containerComposer.getDatabaseType().getType());
CompletableFuture.runAsync(() -> new CDCClient(parameter).start(),
executor).whenComplete((unused, throwable) -> {
if (null != throwable) {
@@ -216,6 +216,6 @@ public final class CDCE2EIT {
}
private String getOrderTableNameWithSchema() {
- return containerComposer.getDatabaseType().isSchemaAvailable() ?
String.join(".", PipelineContainerComposer.SCHEMA_NAME,
SOURCE_TABLE_ORDER_NAME) : SOURCE_TABLE_ORDER_NAME;
+ return containerComposer.getDatabaseType().isSchemaAvailable() ?
String.join(".", PipelineContainerComposer.SCHEMA_NAME, SOURCE_TABLE_NAME) :
SOURCE_TABLE_NAME;
}
}
diff --git
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/general/MySQLMigrationGeneralE2EIT.java
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/general/MySQLMigrationGeneralE2EIT.java
index 9b762db9da6..7097a152d65 100644
---
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/general/MySQLMigrationGeneralE2EIT.java
+++
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/general/MySQLMigrationGeneralE2EIT.java
@@ -52,9 +52,9 @@ import static org.junit.jupiter.api.Assertions.assertTrue;
@Slf4j
public final class MySQLMigrationGeneralE2EIT extends AbstractMigrationE2EIT {
- private static final String SOURCE_TABLE_ORDER_NAME = "t_order_copy";
+ private static final String SOURCE_TABLE_NAME = "t_order_copy";
- private static final String TARGET_TABLE_ORDER_NAME = "t_order";
+ private static final String TARGET_TABLE_NAME = "t_order";
public MySQLMigrationGeneralE2EIT(final PipelineTestParameter testParam) {
super(testParam, new MigrationJobType());
@@ -76,7 +76,7 @@ public final class MySQLMigrationGeneralE2EIT extends
AbstractMigrationE2EIT {
@Test
public void assertMigrationSuccess() throws SQLException,
InterruptedException {
addMigrationProcessConfig();
- getContainerComposer().createSourceOrderTable(SOURCE_TABLE_ORDER_NAME);
+ getContainerComposer().createSourceOrderTable(SOURCE_TABLE_NAME);
getContainerComposer().createSourceOrderItemTable();
addMigrationSourceResource();
addMigrationTargetResource();
@@ -85,15 +85,15 @@ public final class MySQLMigrationGeneralE2EIT extends
AbstractMigrationE2EIT {
createTargetOrderItemTableRule();
Pair<List<Object[]>, List<Object[]>> dataPair =
PipelineCaseHelper.generateFullInsertData(getContainerComposer().getDatabaseType(),
PipelineContainerComposer.TABLE_INIT_ROW_COUNT);
log.info("init data begin: {}", LocalDateTime.now());
-
DataSourceExecuteUtil.execute(getContainerComposer().getSourceDataSource(),
getContainerComposer().getExtraSQLCommand().getFullInsertOrder(SOURCE_TABLE_ORDER_NAME),
dataPair.getLeft());
+
DataSourceExecuteUtil.execute(getContainerComposer().getSourceDataSource(),
getContainerComposer().getExtraSQLCommand().getFullInsertOrder(SOURCE_TABLE_NAME),
dataPair.getLeft());
DataSourceExecuteUtil.execute(getContainerComposer().getSourceDataSource(),
getContainerComposer().getExtraSQLCommand().getFullInsertOrderItem(),
dataPair.getRight());
log.info("init data end: {}", LocalDateTime.now());
- startMigration(SOURCE_TABLE_ORDER_NAME, TARGET_TABLE_ORDER_NAME);
+ startMigration(SOURCE_TABLE_NAME, TARGET_TABLE_NAME);
startMigration("t_order_item", "t_order_item");
- String orderJobId = getJobIdByTableName("ds_0." +
SOURCE_TABLE_ORDER_NAME);
+ String orderJobId = getJobIdByTableName("ds_0." + SOURCE_TABLE_NAME);
getContainerComposer().waitJobPrepareSuccess(String.format("SHOW
MIGRATION STATUS '%s'", orderJobId));
getContainerComposer().startIncrementTask(
- new
E2EIncrementalTask(getContainerComposer().getSourceDataSource(),
SOURCE_TABLE_ORDER_NAME, new SnowflakeKeyGenerateAlgorithm(),
getContainerComposer().getDatabaseType(), 30));
+ new
E2EIncrementalTask(getContainerComposer().getSourceDataSource(),
SOURCE_TABLE_NAME, new SnowflakeKeyGenerateAlgorithm(),
getContainerComposer().getDatabaseType(), 30));
assertMigrationSuccessById(orderJobId, "DATA_MATCH");
String orderItemJobId = getJobIdByTableName("ds_0.t_order_item");
assertMigrationSuccessById(orderItemJobId, "DATA_MATCH");
diff --git
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/general/PostgreSQLMigrationGeneralE2EIT.java
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/general/PostgreSQLMigrationGeneralE2EIT.java
index 031ba2f4103..d966f9c2eaf 100644
---
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/general/PostgreSQLMigrationGeneralE2EIT.java
+++
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/general/PostgreSQLMigrationGeneralE2EIT.java
@@ -50,9 +50,9 @@ import static org.junit.jupiter.api.Assertions.assertTrue;
@Slf4j
public final class PostgreSQLMigrationGeneralE2EIT extends
AbstractMigrationE2EIT {
- private static final String SOURCE_TABLE_ORDER_NAME = "t_order_copy";
+ private static final String SOURCE_TABLE_NAME = "t_order_copy";
- private static final String TARGET_TABLE_ORDER_NAME = "t_order";
+ private static final String TARGET_TABLE_NAME = "t_order";
public PostgreSQLMigrationGeneralE2EIT(final PipelineTestParameter
testParam) {
super(testParam, new MigrationJobType());
@@ -77,25 +77,25 @@ public final class PostgreSQLMigrationGeneralE2EIT extends
AbstractMigrationE2EI
public void assertMigrationSuccess() throws SQLException,
InterruptedException {
addMigrationProcessConfig();
createSourceSchema(PipelineContainerComposer.SCHEMA_NAME);
- getContainerComposer().createSourceOrderTable(SOURCE_TABLE_ORDER_NAME);
+ getContainerComposer().createSourceOrderTable(SOURCE_TABLE_NAME);
getContainerComposer().createSourceOrderItemTable();
-
getContainerComposer().createSourceTableIndexList(PipelineContainerComposer.SCHEMA_NAME,
SOURCE_TABLE_ORDER_NAME);
-
getContainerComposer().createSourceCommentOnList(PipelineContainerComposer.SCHEMA_NAME,
SOURCE_TABLE_ORDER_NAME);
+
getContainerComposer().createSourceTableIndexList(PipelineContainerComposer.SCHEMA_NAME,
SOURCE_TABLE_NAME);
+
getContainerComposer().createSourceCommentOnList(PipelineContainerComposer.SCHEMA_NAME,
SOURCE_TABLE_NAME);
addMigrationSourceResource();
addMigrationTargetResource();
createTargetOrderTableRule();
createTargetOrderItemTableRule();
Pair<List<Object[]>, List<Object[]>> dataPair =
PipelineCaseHelper.generateFullInsertData(getContainerComposer().getDatabaseType(),
PipelineContainerComposer.TABLE_INIT_ROW_COUNT);
log.info("init data begin: {}", LocalDateTime.now());
-
DataSourceExecuteUtil.execute(getContainerComposer().getSourceDataSource(),
getContainerComposer().getExtraSQLCommand().getFullInsertOrder(SOURCE_TABLE_ORDER_NAME),
dataPair.getLeft());
+
DataSourceExecuteUtil.execute(getContainerComposer().getSourceDataSource(),
getContainerComposer().getExtraSQLCommand().getFullInsertOrder(SOURCE_TABLE_NAME),
dataPair.getLeft());
DataSourceExecuteUtil.execute(getContainerComposer().getSourceDataSource(),
getContainerComposer().getExtraSQLCommand().getFullInsertOrderItem(),
dataPair.getRight());
log.info("init data end: {}", LocalDateTime.now());
- startMigrationWithSchema(SOURCE_TABLE_ORDER_NAME, "t_order");
+ startMigrationWithSchema(SOURCE_TABLE_NAME, "t_order");
Awaitility.await().atMost(10, TimeUnit.SECONDS).pollInterval(1,
TimeUnit.SECONDS).until(() -> listJobId().size() > 0);
- String jobId = getJobIdByTableName("ds_0.test." +
SOURCE_TABLE_ORDER_NAME);
+ String jobId = getJobIdByTableName("ds_0.test." + SOURCE_TABLE_NAME);
getContainerComposer().waitIncrementTaskFinished(String.format("SHOW
MIGRATION STATUS '%s'", jobId));
getContainerComposer().startIncrementTask(new E2EIncrementalTask(
- getContainerComposer().getSourceDataSource(), String.join(".",
PipelineContainerComposer.SCHEMA_NAME, SOURCE_TABLE_ORDER_NAME),
+ getContainerComposer().getSourceDataSource(), String.join(".",
PipelineContainerComposer.SCHEMA_NAME, SOURCE_TABLE_NAME),
new SnowflakeKeyGenerateAlgorithm(),
getContainerComposer().getDatabaseType(), 20));
checkOrderMigration(jobId);
checkOrderItemMigration();
@@ -113,11 +113,11 @@ public final class PostgreSQLMigrationGeneralE2EIT
extends AbstractMigrationE2EI
stopMigrationByJobId(jobId);
long recordId = new SnowflakeKeyGenerateAlgorithm().generateKey();
getContainerComposer().sourceExecuteWithLog(String.format("INSERT INTO
%s (order_id,user_id,status) VALUES (%s, %s, '%s')",
- String.join(".", PipelineContainerComposer.SCHEMA_NAME,
SOURCE_TABLE_ORDER_NAME), recordId, 1, "afterStop"));
+ String.join(".", PipelineContainerComposer.SCHEMA_NAME,
SOURCE_TABLE_NAME), recordId, 1, "afterStop"));
startMigrationByJobId(jobId);
// must refresh firstly, otherwise proxy can't get schema and table
info
getContainerComposer().proxyExecuteWithLog("REFRESH TABLE METADATA;",
2);
- getContainerComposer().assertProxyOrderRecordExist(String.join(".",
PipelineContainerComposer.SCHEMA_NAME, TARGET_TABLE_ORDER_NAME), recordId);
+ getContainerComposer().assertProxyOrderRecordExist(String.join(".",
PipelineContainerComposer.SCHEMA_NAME, TARGET_TABLE_NAME), recordId);
assertCheckMigrationSuccess(jobId, "DATA_MATCH");
}
diff --git
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/general/RulesMigrationE2EIT.java
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/general/RulesMigrationE2EIT.java
index 0ac8a5d5126..afb97fe244e 100644
---
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/general/RulesMigrationE2EIT.java
+++
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/general/RulesMigrationE2EIT.java
@@ -50,9 +50,9 @@ import static org.hamcrest.Matchers.is;
@Slf4j
public final class RulesMigrationE2EIT extends AbstractMigrationE2EIT {
- private static final String SOURCE_TABLE_ORDER_NAME = "t_order";
+ private static final String SOURCE_TABLE_NAME = "t_order";
- private static final String TARGET_TABLE_ORDER_NAME = "t_order";
+ private static final String TARGET_TABLE_NAME = "t_order";
public RulesMigrationE2EIT(final PipelineTestParameter testParam) {
super(testParam, new MigrationJobType());
@@ -86,22 +86,22 @@ public final class RulesMigrationE2EIT extends
AbstractMigrationE2EIT {
}
private void assertMigrationSuccess(final Callable<Void> addRuleFn) throws
Exception {
- getContainerComposer().createSourceOrderTable(SOURCE_TABLE_ORDER_NAME);
+ getContainerComposer().createSourceOrderTable(SOURCE_TABLE_NAME);
try (Connection connection =
getContainerComposer().getSourceDataSource().getConnection()) {
-
PipelineCaseHelper.batchInsertOrderRecordsWithGeneralColumns(connection, new
UUIDKeyGenerateAlgorithm(), SOURCE_TABLE_ORDER_NAME,
PipelineContainerComposer.TABLE_INIT_ROW_COUNT);
+
PipelineCaseHelper.batchInsertOrderRecordsWithGeneralColumns(connection, new
UUIDKeyGenerateAlgorithm(), SOURCE_TABLE_NAME,
PipelineContainerComposer.TABLE_INIT_ROW_COUNT);
}
addMigrationSourceResource();
addMigrationTargetResource();
if (null != addRuleFn) {
addRuleFn.call();
}
- startMigration(SOURCE_TABLE_ORDER_NAME, TARGET_TABLE_ORDER_NAME);
+ startMigration(SOURCE_TABLE_NAME, TARGET_TABLE_NAME);
String jobId = listJobId().get(0);
getContainerComposer().waitJobPrepareSuccess(String.format("SHOW
MIGRATION STATUS '%s'", jobId));
getContainerComposer().waitIncrementTaskFinished(String.format("SHOW
MIGRATION STATUS '%s'", jobId));
assertCheckMigrationSuccess(jobId, "DATA_MATCH");
commitMigrationByJobId(jobId);
getContainerComposer().proxyExecuteWithLog("REFRESH TABLE METADATA",
1);
-
assertThat(getContainerComposer().getTargetTableRecordsCount(SOURCE_TABLE_ORDER_NAME),
is(PipelineContainerComposer.TABLE_INIT_ROW_COUNT));
+
assertThat(getContainerComposer().getTargetTableRecordsCount(SOURCE_TABLE_NAME),
is(PipelineContainerComposer.TABLE_INIT_ROW_COUNT));
}
}
diff --git
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/primarykey/IndexesMigrationE2EIT.java
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/primarykey/IndexesMigrationE2EIT.java
index 421f91f2f9b..8accc74d08f 100644
---
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/primarykey/IndexesMigrationE2EIT.java
+++
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/primarykey/IndexesMigrationE2EIT.java
@@ -65,9 +65,9 @@ public final class IndexesMigrationE2EIT extends
AbstractMigrationE2EIT {
+ "TYPE(NAME=\"hash_mod\",PROPERTIES(\"sharding-count\"=\"6\"))\n"
+ ");";
- private static final String SOURCE_TABLE_ORDER_NAME = "t_order";
+ private static final String SOURCE_TABLE_NAME = "t_order";
- private static final String TARGET_TABLE_ORDER_NAME = "t_order";
+ private static final String TARGET_TABLE_NAME = "t_order";
public IndexesMigrationE2EIT(final PipelineTestParameter testParam) {
super(testParam, new MigrationJobType());
@@ -185,15 +185,15 @@ public final class IndexesMigrationE2EIT extends
AbstractMigrationE2EIT {
private void assertMigrationSuccess(final String sqlPattern, final String
shardingColumn, final KeyGenerateAlgorithm keyGenerateAlgorithm,
final String
consistencyCheckAlgorithmType, final Callable<Void> incrementalTaskFn) throws
Exception {
- getContainerComposer().sourceExecuteWithLog(String.format(sqlPattern,
SOURCE_TABLE_ORDER_NAME));
+ getContainerComposer().sourceExecuteWithLog(String.format(sqlPattern,
SOURCE_TABLE_NAME));
try (Connection connection =
getContainerComposer().getSourceDataSource().getConnection()) {
-
PipelineCaseHelper.batchInsertOrderRecordsWithGeneralColumns(connection,
keyGenerateAlgorithm, SOURCE_TABLE_ORDER_NAME,
PipelineContainerComposer.TABLE_INIT_ROW_COUNT);
+
PipelineCaseHelper.batchInsertOrderRecordsWithGeneralColumns(connection,
keyGenerateAlgorithm, SOURCE_TABLE_NAME,
PipelineContainerComposer.TABLE_INIT_ROW_COUNT);
}
addMigrationProcessConfig();
addMigrationSourceResource();
addMigrationTargetResource();
getContainerComposer().proxyExecuteWithLog(String.format(ORDER_TABLE_SHARDING_RULE_FORMAT,
shardingColumn), 2);
- startMigration(SOURCE_TABLE_ORDER_NAME, TARGET_TABLE_ORDER_NAME);
+ startMigration(SOURCE_TABLE_NAME, TARGET_TABLE_NAME);
String jobId = listJobId().get(0);
getContainerComposer().waitJobPrepareSuccess(String.format("SHOW
MIGRATION STATUS '%s'", jobId));
incrementalTaskFn.call();
@@ -203,7 +203,7 @@ public final class IndexesMigrationE2EIT extends
AbstractMigrationE2EIT {
}
commitMigrationByJobId(jobId);
getContainerComposer().proxyExecuteWithLog("REFRESH TABLE METADATA",
1);
-
assertThat(getContainerComposer().getTargetTableRecordsCount(SOURCE_TABLE_ORDER_NAME),
is(PipelineContainerComposer.TABLE_INIT_ROW_COUNT + 1));
+
assertThat(getContainerComposer().getTargetTableRecordsCount(SOURCE_TABLE_NAME),
is(PipelineContainerComposer.TABLE_INIT_ROW_COUNT + 1));
List<String> lastJobIds = listJobId();
assertTrue(lastJobIds.isEmpty());
}
diff --git
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/primarykey/MariaDBMigrationE2EIT.java
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/primarykey/MariaDBMigrationE2EIT.java
index 3913483748a..2ed54d39654 100644
---
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/primarykey/MariaDBMigrationE2EIT.java
+++
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/primarykey/MariaDBMigrationE2EIT.java
@@ -47,9 +47,9 @@ import static org.junit.jupiter.api.Assertions.assertTrue;
@Slf4j
public final class MariaDBMigrationE2EIT extends AbstractMigrationE2EIT {
- private static final String SOURCE_TABLE_ORDER_NAME = "t_order";
+ private static final String SOURCE_TABLE_NAME = "t_order";
- private static final String TARGET_TABLE_ORDER_NAME = "t_order";
+ private static final String TARGET_TABLE_NAME = "t_order";
public MariaDBMigrationE2EIT(final PipelineTestParameter testParam) {
super(testParam, new MigrationJobType());
@@ -73,16 +73,16 @@ public final class MariaDBMigrationE2EIT extends
AbstractMigrationE2EIT {
@Test
public void assertMigrationSuccess() throws SQLException,
InterruptedException {
String sqlPattern = "CREATE TABLE `%s` (`order_id` VARCHAR(64) NOT
NULL, `user_id` INT NOT NULL, `status` varchar(255), PRIMARY KEY (`order_id`))
ENGINE=InnoDB DEFAULT CHARSET=utf8mb4";
- getContainerComposer().sourceExecuteWithLog(String.format(sqlPattern,
SOURCE_TABLE_ORDER_NAME));
+ getContainerComposer().sourceExecuteWithLog(String.format(sqlPattern,
SOURCE_TABLE_NAME));
try (Connection connection =
getContainerComposer().getSourceDataSource().getConnection()) {
KeyGenerateAlgorithm generateAlgorithm = new
UUIDKeyGenerateAlgorithm();
-
PipelineCaseHelper.batchInsertOrderRecordsWithGeneralColumns(connection,
generateAlgorithm, SOURCE_TABLE_ORDER_NAME,
PipelineContainerComposer.TABLE_INIT_ROW_COUNT);
+
PipelineCaseHelper.batchInsertOrderRecordsWithGeneralColumns(connection,
generateAlgorithm, SOURCE_TABLE_NAME,
PipelineContainerComposer.TABLE_INIT_ROW_COUNT);
}
addMigrationProcessConfig();
addMigrationSourceResource();
addMigrationTargetResource();
createTargetOrderTableRule();
- startMigration(SOURCE_TABLE_ORDER_NAME, TARGET_TABLE_ORDER_NAME);
+ startMigration(SOURCE_TABLE_NAME, TARGET_TABLE_NAME);
String jobId = listJobId().get(0);
getContainerComposer().waitJobPrepareSuccess(String.format("SHOW
MIGRATION STATUS '%s'", jobId));
getContainerComposer().sourceExecuteWithLog("INSERT INTO t_order
(order_id, user_id, status) VALUES ('a1', 1, 'OK')");
@@ -91,7 +91,7 @@ public final class MariaDBMigrationE2EIT extends
AbstractMigrationE2EIT {
assertCheckMigrationSuccess(jobId, "CRC32_MATCH");
commitMigrationByJobId(jobId);
getContainerComposer().proxyExecuteWithLog("REFRESH TABLE METADATA",
1);
-
assertThat(getContainerComposer().getTargetTableRecordsCount(SOURCE_TABLE_ORDER_NAME),
is(PipelineContainerComposer.TABLE_INIT_ROW_COUNT + 1));
+
assertThat(getContainerComposer().getTargetTableRecordsCount(SOURCE_TABLE_NAME),
is(PipelineContainerComposer.TABLE_INIT_ROW_COUNT + 1));
List<String> lastJobIds = listJobId();
assertTrue(lastJobIds.isEmpty());
}
diff --git
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/primarykey/TextPrimaryKeyMigrationE2EIT.java
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/primarykey/TextPrimaryKeyMigrationE2EIT.java
index d257521895d..16561dfdf53 100644
---
a/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/primarykey/TextPrimaryKeyMigrationE2EIT.java
+++
b/test/e2e/pipeline/src/test/java/org/apache/shardingsphere/test/e2e/data/pipeline/cases/migration/primarykey/TextPrimaryKeyMigrationE2EIT.java
@@ -47,7 +47,7 @@ import static org.junit.jupiter.api.Assertions.assertTrue;
@Slf4j
public class TextPrimaryKeyMigrationE2EIT extends AbstractMigrationE2EIT {
- private static final String TARGET_TABLE_ORDER_NAME = "t_order";
+ private static final String TARGET_TABLE_NAME = "t_order";
public TextPrimaryKeyMigrationE2EIT(final PipelineTestParameter testParam)
{
super(testParam, new MigrationJobType());
@@ -73,18 +73,18 @@ public class TextPrimaryKeyMigrationE2EIT extends
AbstractMigrationE2EIT {
@Test
public void assertTextPrimaryMigrationSuccess() throws SQLException,
InterruptedException {
-
getContainerComposer().createSourceOrderTable(getSourceTableOrderName());
+ getContainerComposer().createSourceOrderTable(getSourceTableName());
try (Connection connection =
getContainerComposer().getSourceDataSource().getConnection()) {
UUIDKeyGenerateAlgorithm keyGenerateAlgorithm = new
UUIDKeyGenerateAlgorithm();
-
PipelineCaseHelper.batchInsertOrderRecordsWithGeneralColumns(connection,
keyGenerateAlgorithm, getSourceTableOrderName(),
PipelineContainerComposer.TABLE_INIT_ROW_COUNT);
+
PipelineCaseHelper.batchInsertOrderRecordsWithGeneralColumns(connection,
keyGenerateAlgorithm, getSourceTableName(),
PipelineContainerComposer.TABLE_INIT_ROW_COUNT);
}
addMigrationProcessConfig();
addMigrationSourceResource();
addMigrationTargetResource();
createTargetOrderTableRule();
- startMigration(getSourceTableOrderName(), TARGET_TABLE_ORDER_NAME);
+ startMigration(getSourceTableName(), TARGET_TABLE_NAME);
String jobId = listJobId().get(0);
- getContainerComposer().sourceExecuteWithLog(String.format("INSERT INTO
%s (order_id,user_id,status) VALUES (%s, %s, '%s')", getSourceTableOrderName(),
"1000000000", 1, "afterStop"));
+ getContainerComposer().sourceExecuteWithLog(String.format("INSERT INTO
%s (order_id,user_id,status) VALUES (%s, %s, '%s')", getSourceTableName(),
"1000000000", 1, "afterStop"));
getContainerComposer().waitIncrementTaskFinished(String.format("SHOW
MIGRATION STATUS '%s'", jobId));
assertCheckMigrationSuccess(jobId, "DATA_MATCH");
commitMigrationByJobId(jobId);
@@ -92,7 +92,7 @@ public class TextPrimaryKeyMigrationE2EIT extends
AbstractMigrationE2EIT {
assertTrue(lastJobIds.isEmpty());
}
- private String getSourceTableOrderName() {
+ private String getSourceTableName() {
return
DatabaseTypeUtil.isMySQL(getContainerComposer().getDatabaseType()) ? "T_ORDER"
: "t_order";
}
}