This is an automated email from the ASF dual-hosted git repository.

adoroszlai pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/ozone.git


The following commit(s) were added to refs/heads/master by this push:
     new bdb67a18af Revert "HDDS-9240. Test container schema V2 and V3 replica 
mutual import (#5250)"
bdb67a18af is described below

commit bdb67a18af6b58682e57b22e9f85eb43cfa9b193
Author: Doroszlai, Attila <[email protected]>
AuthorDate: Wed Sep 20 09:59:08 2023 +0200

    Revert "HDDS-9240. Test container schema V2 and V3 replica mutual import 
(#5250)"
    
    This reverts commit 0e323bd678bd3059fb3c8ce1474f30611f0e1cad.
---
 .../container/keyvalue/TestKeyValueContainer.java  | 108 ---------------------
 1 file changed, 108 deletions(-)

diff --git 
a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueContainer.java
 
b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueContainer.java
index d48d7eb53a..3c1a57c9d2 100644
--- 
a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueContainer.java
+++ 
b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/keyvalue/TestKeyValueContainer.java
@@ -33,7 +33,6 @@ import org.apache.hadoop.hdds.utils.db.RDBStore;
 import org.apache.hadoop.hdds.utils.db.RocksDatabase.ColumnFamily;
 import org.apache.hadoop.hdds.utils.db.Table;
 import org.apache.hadoop.ozone.OzoneConsts;
-import org.apache.hadoop.ozone.container.ContainerTestHelper;
 import org.apache.hadoop.ozone.container.common.helpers.BlockData;
 import org.apache.hadoop.ozone.container.common.impl.ContainerLayoutVersion;
 import org.apache.hadoop.ozone.container.common.impl.ContainerDataYaml;
@@ -63,7 +62,6 @@ import org.junit.Assume;
 import org.junit.Before;
 import org.junit.Rule;
 import org.junit.Test;
-import org.junit.jupiter.api.Assertions;
 import org.junit.rules.TemporaryFolder;
 
 import org.junit.runner.RunWith;
@@ -92,11 +90,7 @@ import java.util.stream.Collectors;
 import java.util.stream.IntStream;
 import java.util.stream.Stream;
 
-import static java.util.Collections.singletonList;
 import static org.apache.hadoop.hdds.HddsConfigKeys.HDDS_DB_PROFILE;
-import static org.apache.hadoop.ozone.OzoneConsts.SCHEMA_V2;
-import static org.apache.hadoop.ozone.OzoneConsts.SCHEMA_V3;
-import static 
org.apache.hadoop.ozone.container.common.statemachine.DatanodeConfiguration.CONTAINER_SCHEMA_V3_ENABLED;
 import static 
org.apache.hadoop.ozone.container.keyvalue.helpers.KeyValueContainerUtil.isSameSchemaVersion;
 import static 
org.apache.hadoop.ozone.container.replication.CopyContainerCompression.NO_COMPRESSION;
 import static org.apache.ratis.util.Preconditions.assertTrue;
@@ -942,106 +936,4 @@ public class TestKeyValueContainer {
       Assert.assertTrue(container.getContainerData().isEmpty());
     }
   }
-
-  /**
-   * Test import schema V2 replica to V3 enabled HddsVolume.
-   */
-  @Test
-  public void testImportV2ReplicaToV3HddsVolume() throws Exception {
-    final String testDir = GenericTestUtils.getTempPath(
-        TestKeyValueContainer.class.getSimpleName() + "-"
-            + UUID.randomUUID());
-    try {
-      testMixedSchemaImport(testDir, false);
-    } finally {
-      FileUtils.deleteDirectory(new File(testDir));
-    }
-  }
-
-  /**
-   * Test import schema V3 replica to V3 disabled HddsVolume.
-   */
-  @Test
-  public void testImportV3ReplicaToV2HddsVolume() throws Exception {
-    final String testDir = GenericTestUtils.getTempPath(
-        TestKeyValueContainer.class.getSimpleName() + "-"
-            + UUID.randomUUID());
-    try {
-      testMixedSchemaImport(testDir, true);
-    } finally {
-      FileUtils.deleteDirectory(new File(testDir));
-    }
-  }
-
-  private void testMixedSchemaImport(String dir,
-      boolean schemaV3Enabled) throws IOException {
-    OzoneConfiguration conf = new OzoneConfiguration();
-    final String dir1 = dir + (schemaV3Enabled ? "/v3" : "/v2");
-
-    // create HddsVolume
-    HddsVolume hddsVolume1 = new HddsVolume.Builder(dir1)
-        .conf(conf).datanodeUuid(datanodeId.toString()).build();
-    conf.setBoolean(CONTAINER_SCHEMA_V3_ENABLED, schemaV3Enabled);
-    StorageVolumeUtil.checkVolume(hddsVolume1, scmId, scmId, conf, null, null);
-    Mockito.when(volumeChoosingPolicy.chooseVolume(anyList(), anyLong()))
-        .thenReturn(hddsVolume1);
-
-    // create container
-    long containerId = 1;
-    KeyValueContainerData data = new KeyValueContainerData(containerId,
-        ContainerLayoutVersion.FILE_PER_BLOCK,
-        ContainerTestHelper.CONTAINER_MAX_SIZE, UUID.randomUUID().toString(),
-        UUID.randomUUID().toString());
-    KeyValueContainer container = new KeyValueContainer(data, conf);
-    container.create(volumeSet, volumeChoosingPolicy, scmId);
-    long pendingDeleteBlockCount = 20;
-    try (DBHandle meta = BlockUtils.getDB(data, conf)) {
-      Table<String, Long> metadataTable = meta.getStore().getMetadataTable();
-      metadataTable.put(data.getPendingDeleteBlockCountKey(),
-          pendingDeleteBlockCount);
-    }
-    container.close();
-
-    // verify container schema
-    if (schemaV3Enabled) {
-      Assert.assertEquals(SCHEMA_V3,
-          container.getContainerData().getSchemaVersion());
-    } else {
-      Assert.assertEquals(SCHEMA_V2,
-          container.getContainerData().getSchemaVersion());
-    }
-
-    //export container
-    TarContainerPacker packer = new TarContainerPacker(NO_COMPRESSION);
-    File file1 = new File(dir1 + "/" + containerId);
-    if (!file1.createNewFile()) {
-      Assertions.fail("Failed to create file " + file1.getAbsolutePath());
-    }
-    try (FileOutputStream fos = new FileOutputStream(file1)) {
-      container.exportContainerData(fos, packer);
-    }
-
-    // create new HddsVolume
-    conf.setBoolean(CONTAINER_SCHEMA_V3_ENABLED, !schemaV3Enabled);
-    final String dir2 = dir + (schemaV3Enabled ? "/v2" : "/v3");
-    HddsVolume hddsVolume2 = new HddsVolume.Builder(dir2)
-        .conf(conf).datanodeUuid(datanodeId.toString()).build();
-    StorageVolumeUtil.checkVolume(hddsVolume2, scmId, scmId, conf, null, null);
-    Mockito.when(volumeChoosingPolicy.chooseVolume(anyList(), anyLong()))
-        .thenReturn(hddsVolume2);
-    Mockito.when(volumeSet.getVolumesList()).thenReturn(
-        singletonList(hddsVolume2));
-
-    // import container to new HddsVolume
-    KeyValueContainer importedContainer = new KeyValueContainer(data, conf);
-    importedContainer.populatePathFields(scmId, hddsVolume2);
-    try (FileInputStream fio = new FileInputStream(file1)) {
-      importedContainer.importContainerData(fio, packer);
-    }
-
-    Assert.assertEquals(schemaV3Enabled ? SCHEMA_V3 : SCHEMA_V2,
-        importedContainer.getContainerData().getSchemaVersion());
-    Assert.assertEquals(pendingDeleteBlockCount,
-        importedContainer.getContainerData().getNumPendingDeletionBlocks());
-  }
 }


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to