This is an automated email from the ASF dual-hosted git repository.

ckj pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/incubator-uniffle.git


The following commit(s) were added to refs/heads/master by this push:
     new 76559ce3 [#722] test: cleanup residue files in tmp directory after 
tests (#1134)
76559ce3 is described below

commit 76559ce3b16c1595818947eb8a5d208d02659a39
Author: Neo Chien <[email protected]>
AuthorDate: Mon Aug 21 19:34:06 2023 +0800

    [#722] test: cleanup residue files in tmp directory after tests (#1134)
    
    ### What changes were proposed in this pull request?
    
    Cleanup residue files in tmp directory after tests
    
    ### Why are the changes needed?
    
    Fix: #722
    
    ### Does this PR introduce _any_ user-facing change?
    
    No.
    
    ### How was this patch tested?
    
    current UT
---
 .../java/org/apache/uniffle/common/util/ChecksumUtilsTest.java |  5 +++--
 .../test/java/org/apache/uniffle/test/IntegrationTestBase.java |  7 ++++---
 .../uniffle/test/ShuffleServerWithKerberizedHadoopTest.java    |  8 ++++----
 .../src/test/java/org/apache/uniffle/test/RepartitionTest.java |  5 +++--
 .../src/test/java/org/apache/uniffle/test/SparkSQLTest.java    |  5 +++--
 .../uniffle/test/ContinuousSelectPartitionStrategyTest.java    |  8 ++++----
 .../apache/uniffle/test/GetShuffleReportForMultiPartTest.java  |  8 ++++----
 .../java/org/apache/uniffle/server/ShuffleTaskManagerTest.java | 10 ++++++----
 8 files changed, 31 insertions(+), 25 deletions(-)

diff --git 
a/common/src/test/java/org/apache/uniffle/common/util/ChecksumUtilsTest.java 
b/common/src/test/java/org/apache/uniffle/common/util/ChecksumUtilsTest.java
index 37c31052..100be0f7 100644
--- a/common/src/test/java/org/apache/uniffle/common/util/ChecksumUtilsTest.java
+++ b/common/src/test/java/org/apache/uniffle/common/util/ChecksumUtilsTest.java
@@ -21,18 +21,20 @@ import java.io.File;
 import java.io.FileOutputStream;
 import java.nio.ByteBuffer;
 import java.nio.channels.FileChannel;
-import java.nio.file.Files;
 import java.nio.file.Path;
 import java.nio.file.Paths;
 import java.util.Random;
 import java.util.zip.CRC32;
 
 import org.junit.jupiter.api.Test;
+import org.junit.jupiter.api.io.TempDir;
 
 import static org.junit.jupiter.api.Assertions.assertEquals;
 
 public class ChecksumUtilsTest {
 
+  @TempDir File tempDir;
+
   @Test
   public void crc32TestWithByte() {
     byte[] data = new byte[32 * 1024 * 1024];
@@ -56,7 +58,6 @@ public class ChecksumUtilsTest {
     byte[] data = new byte[length];
     new Random().nextBytes(data);
 
-    String tempDir = Files.createTempDirectory("rss").toString();
     File file = new File(tempDir, "crc_test.txt");
     file.createNewFile();
     file.deleteOnExit();
diff --git 
a/integration-test/common/src/test/java/org/apache/uniffle/test/IntegrationTestBase.java
 
b/integration-test/common/src/test/java/org/apache/uniffle/test/IntegrationTestBase.java
index e73ccfa7..269a7d12 100644
--- 
a/integration-test/common/src/test/java/org/apache/uniffle/test/IntegrationTestBase.java
+++ 
b/integration-test/common/src/test/java/org/apache/uniffle/test/IntegrationTestBase.java
@@ -27,6 +27,7 @@ import java.util.concurrent.atomic.AtomicInteger;
 
 import com.google.common.collect.Lists;
 import org.junit.jupiter.api.AfterAll;
+import org.junit.jupiter.api.io.TempDir;
 
 import org.apache.uniffle.common.util.RssUtils;
 import org.apache.uniffle.coordinator.CoordinatorConf;
@@ -64,6 +65,8 @@ public abstract class IntegrationTestBase extends 
HadoopTestBase {
   protected static final int NETTY_PORT = 21000;
   protected static AtomicInteger nettyPortCounter = new AtomicInteger();
 
+  static @TempDir File tempDir;
+
   public static void startServers() throws Exception {
     for (CoordinatorServer coordinator : coordinators) {
       coordinator.start();
@@ -106,12 +109,10 @@ public abstract class IntegrationTestBase extends 
HadoopTestBase {
   }
 
   protected static ShuffleServerConf getShuffleServerConf() throws Exception {
-    File dataFolder = Files.createTempDirectory("rssdata").toFile();
     ShuffleServerConf serverConf = new ShuffleServerConf();
-    dataFolder.deleteOnExit();
     serverConf.setInteger("rss.rpc.server.port", SHUFFLE_SERVER_PORT);
     serverConf.setString("rss.storage.type", 
StorageType.MEMORY_LOCALFILE_HDFS.name());
-    serverConf.setString("rss.storage.basePath", dataFolder.getAbsolutePath());
+    serverConf.setString("rss.storage.basePath", tempDir.getAbsolutePath());
     serverConf.setString("rss.server.buffer.capacity", "671088640");
     serverConf.setString("rss.server.memory.shuffle.highWaterMark", "50.0");
     serverConf.setString("rss.server.memory.shuffle.lowWaterMark", "0.0");
diff --git 
a/integration-test/common/src/test/java/org/apache/uniffle/test/ShuffleServerWithKerberizedHadoopTest.java
 
b/integration-test/common/src/test/java/org/apache/uniffle/test/ShuffleServerWithKerberizedHadoopTest.java
index 8bced501..60ad0535 100644
--- 
a/integration-test/common/src/test/java/org/apache/uniffle/test/ShuffleServerWithKerberizedHadoopTest.java
+++ 
b/integration-test/common/src/test/java/org/apache/uniffle/test/ShuffleServerWithKerberizedHadoopTest.java
@@ -18,7 +18,6 @@
 package org.apache.uniffle.test;
 
 import java.io.File;
-import java.nio.file.Files;
 import java.util.HashMap;
 import java.util.List;
 import java.util.Map;
@@ -31,6 +30,7 @@ import org.junit.jupiter.api.AfterEach;
 import org.junit.jupiter.api.BeforeAll;
 import org.junit.jupiter.api.BeforeEach;
 import org.junit.jupiter.api.Test;
+import org.junit.jupiter.api.io.TempDir;
 import org.roaringbitmap.longlong.Roaring64NavigableMap;
 
 import org.apache.uniffle.client.TestUtils;
@@ -80,13 +80,13 @@ public class ShuffleServerWithKerberizedHadoopTest extends 
KerberizedHadoopBase
   private static CoordinatorServer coordinatorServer;
   private static ShuffleServer shuffleServer;
 
+  static @TempDir File tempDir;
+
   private static ShuffleServerConf getShuffleServerConf() throws Exception {
-    File dataFolder = Files.createTempDirectory("rssdata").toFile();
     ShuffleServerConf serverConf = new ShuffleServerConf();
-    dataFolder.deleteOnExit();
     serverConf.setInteger("rss.rpc.server.port", SHUFFLE_SERVER_PORT);
     serverConf.setString("rss.storage.type", 
StorageType.MEMORY_LOCALFILE_HDFS.name());
-    serverConf.setString("rss.storage.basePath", dataFolder.getAbsolutePath());
+    serverConf.setString("rss.storage.basePath", tempDir.getAbsolutePath());
     serverConf.setString("rss.server.buffer.capacity", "671088640");
     serverConf.setString("rss.server.memory.shuffle.highWaterMark", "50.0");
     serverConf.setString("rss.server.memory.shuffle.lowWaterMark", "0.0");
diff --git 
a/integration-test/spark-common/src/test/java/org/apache/uniffle/test/RepartitionTest.java
 
b/integration-test/spark-common/src/test/java/org/apache/uniffle/test/RepartitionTest.java
index 277649fc..a5d72a42 100644
--- 
a/integration-test/spark-common/src/test/java/org/apache/uniffle/test/RepartitionTest.java
+++ 
b/integration-test/spark-common/src/test/java/org/apache/uniffle/test/RepartitionTest.java
@@ -20,7 +20,6 @@ package org.apache.uniffle.test;
 import java.io.File;
 import java.io.FileWriter;
 import java.io.PrintWriter;
-import java.nio.file.Files;
 import java.util.Arrays;
 import java.util.Map;
 import java.util.Random;
@@ -33,6 +32,7 @@ import org.apache.spark.api.java.JavaPairRDD;
 import org.apache.spark.api.java.JavaRDD;
 import org.apache.spark.sql.SparkSession;
 import org.junit.jupiter.api.Test;
+import org.junit.jupiter.api.io.TempDir;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
@@ -40,6 +40,8 @@ public abstract class RepartitionTest extends 
SparkIntegrationTestBase {
 
   private static final Logger LOG = 
LoggerFactory.getLogger(RepartitionTest.class);
 
+  static @TempDir File tempDir;
+
   @Test
   public void resultCompareTest() throws Exception {
     run();
@@ -63,7 +65,6 @@ public abstract class RepartitionTest extends 
SparkIntegrationTestBase {
   public abstract void updateRssStorage(SparkConf sparkConf);
 
   protected String generateTextFile(int wordsPerRow, int rows) throws 
Exception {
-    String tempDir = Files.createTempDirectory("rss").toString();
     File file = new File(tempDir, "wordcount.txt");
     file.createNewFile();
     LOG.info("Create file:" + file.getAbsolutePath());
diff --git 
a/integration-test/spark-common/src/test/java/org/apache/uniffle/test/SparkSQLTest.java
 
b/integration-test/spark-common/src/test/java/org/apache/uniffle/test/SparkSQLTest.java
index effce773..1231a6cd 100644
--- 
a/integration-test/spark-common/src/test/java/org/apache/uniffle/test/SparkSQLTest.java
+++ 
b/integration-test/spark-common/src/test/java/org/apache/uniffle/test/SparkSQLTest.java
@@ -20,7 +20,6 @@ package org.apache.uniffle.test;
 import java.io.File;
 import java.io.FileWriter;
 import java.io.PrintWriter;
-import java.nio.file.Files;
 import java.util.Map;
 import java.util.Random;
 
@@ -31,6 +30,7 @@ import org.apache.spark.sql.Dataset;
 import org.apache.spark.sql.Row;
 import org.apache.spark.sql.SparkSession;
 import org.junit.jupiter.api.Test;
+import org.junit.jupiter.api.io.TempDir;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
@@ -38,6 +38,8 @@ public abstract class SparkSQLTest extends 
SparkIntegrationTestBase {
 
   private static final Logger LOG = 
LoggerFactory.getLogger(SparkSQLTest.class);
 
+  static @TempDir File tempDir;
+
   @Test
   public void resultCompareTest() throws Exception {
     run();
@@ -76,7 +78,6 @@ public abstract class SparkSQLTest extends 
SparkIntegrationTestBase {
 
   protected String generateCsvFile() throws Exception {
     int rows = 1000;
-    String tempDir = Files.createTempDirectory("rss").toString();
     File file = new File(tempDir, "test.csv");
     file.createNewFile();
     LOG.info("Create file:" + file.getAbsolutePath());
diff --git 
a/integration-test/spark3/src/test/java/org/apache/uniffle/test/ContinuousSelectPartitionStrategyTest.java
 
b/integration-test/spark3/src/test/java/org/apache/uniffle/test/ContinuousSelectPartitionStrategyTest.java
index f15d26bf..e1dd2ac3 100644
--- 
a/integration-test/spark3/src/test/java/org/apache/uniffle/test/ContinuousSelectPartitionStrategyTest.java
+++ 
b/integration-test/spark3/src/test/java/org/apache/uniffle/test/ContinuousSelectPartitionStrategyTest.java
@@ -18,7 +18,6 @@
 package org.apache.uniffle.test;
 
 import java.io.File;
-import java.nio.file.Files;
 import java.util.Comparator;
 import java.util.List;
 import java.util.Map;
@@ -37,6 +36,7 @@ import org.apache.spark.sql.functions;
 import org.apache.spark.sql.internal.SQLConf;
 import org.junit.jupiter.api.BeforeAll;
 import org.junit.jupiter.api.Test;
+import org.junit.jupiter.api.io.TempDir;
 
 import org.apache.uniffle.coordinator.CoordinatorConf;
 import 
org.apache.uniffle.coordinator.strategy.assignment.AbstractAssignmentStrategy;
@@ -54,6 +54,8 @@ public class ContinuousSelectPartitionStrategyTest extends 
SparkIntegrationTestB
   private static final int replicateWrite = 3;
   private static final int replicateRead = 2;
 
+  static @TempDir File tempDir;
+
   @BeforeAll
   public static void setupServers() throws Exception {
     CoordinatorConf coordinatorConf = getCoordinatorConf();
@@ -75,13 +77,11 @@ public class ContinuousSelectPartitionStrategyTest extends 
SparkIntegrationTestB
   private static void createShuffleServers() throws Exception {
     for (int i = 0; i < 3; i++) {
       // Copy from IntegrationTestBase#getShuffleServerConf
-      File dataFolder = Files.createTempDirectory("rssdata" + i).toFile();
       ShuffleServerConf serverConf = new ShuffleServerConf();
-      dataFolder.deleteOnExit();
       serverConf.setInteger("rss.rpc.server.port", SHUFFLE_SERVER_PORT + i);
       serverConf.setInteger("rss.server.netty.port", NETTY_PORT + i);
       serverConf.setString("rss.storage.type", 
StorageType.MEMORY_LOCALFILE_HDFS.name());
-      serverConf.setString("rss.storage.basePath", 
dataFolder.getAbsolutePath());
+      serverConf.setString("rss.storage.basePath", tempDir.getAbsolutePath());
       serverConf.setString("rss.server.buffer.capacity", 
String.valueOf(671088640 - i));
       serverConf.setString("rss.server.memory.shuffle.highWaterMark", "50.0");
       serverConf.setString("rss.server.memory.shuffle.lowWaterMark", "0.0");
diff --git 
a/integration-test/spark3/src/test/java/org/apache/uniffle/test/GetShuffleReportForMultiPartTest.java
 
b/integration-test/spark3/src/test/java/org/apache/uniffle/test/GetShuffleReportForMultiPartTest.java
index f7944ceb..d9791702 100644
--- 
a/integration-test/spark3/src/test/java/org/apache/uniffle/test/GetShuffleReportForMultiPartTest.java
+++ 
b/integration-test/spark3/src/test/java/org/apache/uniffle/test/GetShuffleReportForMultiPartTest.java
@@ -18,7 +18,6 @@
 package org.apache.uniffle.test;
 
 import java.io.File;
-import java.nio.file.Files;
 import java.util.Comparator;
 import java.util.List;
 import java.util.Map;
@@ -43,6 +42,7 @@ import org.apache.spark.sql.functions;
 import org.apache.spark.sql.internal.SQLConf;
 import org.junit.jupiter.api.BeforeAll;
 import org.junit.jupiter.api.Test;
+import org.junit.jupiter.api.io.TempDir;
 import org.roaringbitmap.longlong.Roaring64NavigableMap;
 
 import org.apache.uniffle.common.ShuffleServerInfo;
@@ -62,6 +62,8 @@ public class GetShuffleReportForMultiPartTest extends 
SparkIntegrationTestBase {
   private static final int replicateWrite = 3;
   private static final int replicateRead = 2;
 
+  static @TempDir File tempDir;
+
   @BeforeAll
   public static void setupServers() throws Exception {
     CoordinatorConf coordinatorConf = getCoordinatorConf();
@@ -79,13 +81,11 @@ public class GetShuffleReportForMultiPartTest extends 
SparkIntegrationTestBase {
   private static void createShuffleServers() throws Exception {
     for (int i = 0; i < 4; i++) {
       // Copy from IntegrationTestBase#getShuffleServerConf
-      File dataFolder = Files.createTempDirectory("rssdata" + i).toFile();
       ShuffleServerConf serverConf = new ShuffleServerConf();
-      dataFolder.deleteOnExit();
       serverConf.setInteger("rss.rpc.server.port", SHUFFLE_SERVER_PORT + i);
       serverConf.setInteger("rss.server.netty.port", NETTY_PORT + i);
       serverConf.setString("rss.storage.type", 
StorageType.MEMORY_LOCALFILE_HDFS.name());
-      serverConf.setString("rss.storage.basePath", 
dataFolder.getAbsolutePath());
+      serverConf.setString("rss.storage.basePath", tempDir.getAbsolutePath());
       serverConf.setString("rss.server.buffer.capacity", "671088640");
       serverConf.setString("rss.server.memory.shuffle.highWaterMark", "50.0");
       serverConf.setString("rss.server.memory.shuffle.lowWaterMark", "0.0");
diff --git 
a/server/src/test/java/org/apache/uniffle/server/ShuffleTaskManagerTest.java 
b/server/src/test/java/org/apache/uniffle/server/ShuffleTaskManagerTest.java
index 3458655d..ea32911e 100644
--- a/server/src/test/java/org/apache/uniffle/server/ShuffleTaskManagerTest.java
+++ b/server/src/test/java/org/apache/uniffle/server/ShuffleTaskManagerTest.java
@@ -18,7 +18,6 @@
 package org.apache.uniffle.server;
 
 import java.io.File;
-import java.nio.file.Files;
 import java.util.Arrays;
 import java.util.HashSet;
 import java.util.List;
@@ -80,10 +79,15 @@ public class ShuffleTaskManagerTest extends HadoopTestBase {
 
   private ShuffleServer shuffleServer;
 
+  @TempDir File tempDir1;
+  @TempDir File tempDir2;
+
   @BeforeEach
   public void beforeEach() {
     ShuffleServerMetrics.clear();
     ShuffleServerMetrics.register();
+    assertTrue(this.tempDir1.isDirectory());
+    assertTrue(this.tempDir2.isDirectory());
   }
 
   @AfterEach
@@ -463,11 +467,9 @@ public class ShuffleTaskManagerTest extends HadoopTestBase 
{
 
     conf.setString(ShuffleServerConf.RSS_STORAGE_TYPE.key(), "LOCALFILE");
     conf.set(ShuffleServerConf.RSS_TEST_MODE_ENABLE, true);
-    java.nio.file.Path path1 = 
Files.createTempDirectory("removeShuffleDataWithLocalfileTest");
-    java.nio.file.Path path2 = 
Files.createTempDirectory("removeShuffleDataWithLocalfileTest");
     conf.setString(
         ShuffleServerConf.RSS_STORAGE_BASE_PATH.key(),
-        path1.toAbsolutePath().toString() + "," + 
path2.toAbsolutePath().toString());
+        tempDir1.getAbsolutePath() + "," + tempDir2.getAbsolutePath());
 
     shuffleServer = new ShuffleServer(conf);
     ShuffleTaskManager shuffleTaskManager = 
shuffleServer.getShuffleTaskManager();

Reply via email to