This is an automated email from the ASF dual-hosted git repository.
huaxingao pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/iceberg.git
The following commit(s) were added to refs/heads/main by this push:
new bd54026d73 Spark: Delegate temp file deletion to JUnit in
TestParquetVectorizedReads (#15557)
bd54026d73 is described below
commit bd54026d732a1eac49d0a6852036d1cd113ea65d
Author: Noritaka Sekiyama <[email protected]>
AuthorDate: Fri Mar 13 04:54:33 2026 +0800
Spark: Delegate temp file deletion to JUnit in TestParquetVectorizedReads
(#15557)
* Delegate temp file deletion to JUnit in TestParquetVectorizedReads
(#13506)
Replace the manual File.createTempFile + delete pattern with
temp.resolve(...).toFile() so JUnit's @TempDir handles cleanup
automatically. Applied to Spark v3.5, v4.0, and v4.1 modules.
Co-Authored-By: Claude Sonnet 4.6 <[email protected]>
* Rename junit.parquet to data.parquet per review feedback
Co-authored-by: Isaac
---------
Co-authored-by: Claude Sonnet 4.6 <[email protected]>
---
.../vectorized/parquet/TestParquetVectorizedReads.java | 15 +++++----------
.../vectorized/parquet/TestParquetVectorizedReads.java | 15 +++++----------
.../vectorized/parquet/TestParquetVectorizedReads.java | 15 +++++----------
3 files changed, 15 insertions(+), 30 deletions(-)
diff --git
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/data/vectorized/parquet/TestParquetVectorizedReads.java
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/data/vectorized/parquet/TestParquetVectorizedReads.java
index 2885563863..24a3603dc7 100644
---
a/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/data/vectorized/parquet/TestParquetVectorizedReads.java
+++
b/spark/v3.5/spark/src/test/java/org/apache/iceberg/spark/data/vectorized/parquet/TestParquetVectorizedReads.java
@@ -213,8 +213,7 @@ public class TestParquetVectorizedReads extends
AvroDataTestBase {
Map<Integer, Object> idToConstant)
throws IOException {
// write a test parquet file using iceberg writer
- File testFile = File.createTempFile("junit", null, temp.toFile());
- assertThat(testFile.delete()).as("Delete should succeed").isTrue();
+ File testFile = temp.resolve("data.parquet").toFile();
try (FileAppender<Record> writer = getParquetWriter(writeSchema,
testFile)) {
writer.addAll(expected);
@@ -390,8 +389,7 @@ public class TestParquetVectorizedReads extends
AvroDataTestBase {
optional(102, "float_data", Types.FloatType.get()),
optional(103, "decimal_data", Types.DecimalType.of(10, 5)));
- File dataFile = File.createTempFile("junit", null, temp.toFile());
- assertThat(dataFile.delete()).as("Delete should succeed").isTrue();
+ File dataFile = temp.resolve("data.parquet").toFile();
Iterable<Record> data =
generateData(writeSchema, 30000, 0L,
RandomData.DEFAULT_NULL_PERCENTAGE, IDENTITY);
try (FileAppender<Record> writer = getParquetWriter(writeSchema,
dataFile)) {
@@ -423,8 +421,7 @@ public class TestParquetVectorizedReads extends
AvroDataTestBase {
optional(107, "string_data", Types.StringType.get()),
optional(108, "binary_data", Types.BinaryType.get()));
- File dataFile = File.createTempFile("junit", null, temp.toFile());
- assertThat(dataFile.delete()).as("Delete should succeed").isTrue();
+ File dataFile = temp.resolve("data.parquet").toFile();
Iterable<Record> data =
generateData(schema, 30000, 0L, RandomData.DEFAULT_NULL_PERCENTAGE,
IDENTITY);
try (FileAppender<Record> writer = getParquetV2Writer(schema, dataFile)) {
@@ -437,8 +434,7 @@ public class TestParquetVectorizedReads extends
AvroDataTestBase {
public void testUnsupportedReadsForParquetV2() throws Exception {
// Some types use delta encoding and which are not supported for
vectorized reads
Schema schema = new Schema(SUPPORTED_PRIMITIVES.fields());
- File dataFile = File.createTempFile("junit", null, temp.toFile());
- assertThat(dataFile.delete()).as("Delete should succeed").isTrue();
+ File dataFile = temp.resolve("data.parquet").toFile();
Iterable<Record> data =
generateData(schema, 30000, 0L, RandomData.DEFAULT_NULL_PERCENTAGE,
IDENTITY);
try (FileAppender<Record> writer = getParquetV2Writer(schema, dataFile)) {
@@ -456,8 +452,7 @@ public class TestParquetVectorizedReads extends
AvroDataTestBase {
int numRows = 1;
Schema schema = new Schema(optional(100, "uuid", Types.UUIDType.get()));
- File dataFile = File.createTempFile("junit", null, temp.toFile());
- assertThat(dataFile.delete()).as("Delete should succeed").isTrue();
+ File dataFile = temp.resolve("data.parquet").toFile();
Iterable<Record> data = generateData(schema, numRows, 0L, 0, IDENTITY);
try (FileAppender<Record> writer = getParquetV2Writer(schema, dataFile)) {
writer.addAll(data);
diff --git
a/spark/v4.0/spark/src/test/java/org/apache/iceberg/spark/data/vectorized/parquet/TestParquetVectorizedReads.java
b/spark/v4.0/spark/src/test/java/org/apache/iceberg/spark/data/vectorized/parquet/TestParquetVectorizedReads.java
index 4eabac7ab8..6011c6dad7 100644
---
a/spark/v4.0/spark/src/test/java/org/apache/iceberg/spark/data/vectorized/parquet/TestParquetVectorizedReads.java
+++
b/spark/v4.0/spark/src/test/java/org/apache/iceberg/spark/data/vectorized/parquet/TestParquetVectorizedReads.java
@@ -213,8 +213,7 @@ public class TestParquetVectorizedReads extends
AvroDataTestBase {
Map<Integer, Object> idToConstant)
throws IOException {
// write a test parquet file using iceberg writer
- File testFile = File.createTempFile("junit", null, temp.toFile());
- assertThat(testFile.delete()).as("Delete should succeed").isTrue();
+ File testFile = temp.resolve("data.parquet").toFile();
try (FileAppender<Record> writer = getParquetWriter(writeSchema,
testFile)) {
writer.addAll(expected);
@@ -390,8 +389,7 @@ public class TestParquetVectorizedReads extends
AvroDataTestBase {
optional(102, "float_data", Types.FloatType.get()),
optional(103, "decimal_data", Types.DecimalType.of(10, 5)));
- File dataFile = File.createTempFile("junit", null, temp.toFile());
- assertThat(dataFile.delete()).as("Delete should succeed").isTrue();
+ File dataFile = temp.resolve("data.parquet").toFile();
Iterable<Record> data =
generateData(writeSchema, 30000, 0L,
RandomData.DEFAULT_NULL_PERCENTAGE, IDENTITY);
try (FileAppender<Record> writer = getParquetWriter(writeSchema,
dataFile)) {
@@ -423,8 +421,7 @@ public class TestParquetVectorizedReads extends
AvroDataTestBase {
optional(107, "string_data", Types.StringType.get()),
optional(108, "binary_data", Types.BinaryType.get()));
- File dataFile = File.createTempFile("junit", null, temp.toFile());
- assertThat(dataFile.delete()).as("Delete should succeed").isTrue();
+ File dataFile = temp.resolve("data.parquet").toFile();
Iterable<Record> data =
generateData(schema, 30000, 0L, RandomData.DEFAULT_NULL_PERCENTAGE,
IDENTITY);
try (FileAppender<Record> writer = getParquetV2Writer(schema, dataFile)) {
@@ -437,8 +434,7 @@ public class TestParquetVectorizedReads extends
AvroDataTestBase {
public void testUnsupportedReadsForParquetV2() throws Exception {
// Some types use delta encoding and which are not supported for
vectorized reads
Schema schema = new Schema(SUPPORTED_PRIMITIVES.fields());
- File dataFile = File.createTempFile("junit", null, temp.toFile());
- assertThat(dataFile.delete()).as("Delete should succeed").isTrue();
+ File dataFile = temp.resolve("data.parquet").toFile();
Iterable<Record> data =
generateData(schema, 30000, 0L, RandomData.DEFAULT_NULL_PERCENTAGE,
IDENTITY);
try (FileAppender<Record> writer = getParquetV2Writer(schema, dataFile)) {
@@ -456,8 +452,7 @@ public class TestParquetVectorizedReads extends
AvroDataTestBase {
int numRows = 1;
Schema schema = new Schema(optional(100, "uuid", Types.UUIDType.get()));
- File dataFile = File.createTempFile("junit", null, temp.toFile());
- assertThat(dataFile.delete()).as("Delete should succeed").isTrue();
+ File dataFile = temp.resolve("data.parquet").toFile();
Iterable<Record> data = generateData(schema, numRows, 0L, 0, IDENTITY);
try (FileAppender<Record> writer = getParquetV2Writer(schema, dataFile)) {
writer.addAll(data);
diff --git
a/spark/v4.1/spark/src/test/java/org/apache/iceberg/spark/data/vectorized/parquet/TestParquetVectorizedReads.java
b/spark/v4.1/spark/src/test/java/org/apache/iceberg/spark/data/vectorized/parquet/TestParquetVectorizedReads.java
index 4eabac7ab8..6011c6dad7 100644
---
a/spark/v4.1/spark/src/test/java/org/apache/iceberg/spark/data/vectorized/parquet/TestParquetVectorizedReads.java
+++
b/spark/v4.1/spark/src/test/java/org/apache/iceberg/spark/data/vectorized/parquet/TestParquetVectorizedReads.java
@@ -213,8 +213,7 @@ public class TestParquetVectorizedReads extends
AvroDataTestBase {
Map<Integer, Object> idToConstant)
throws IOException {
// write a test parquet file using iceberg writer
- File testFile = File.createTempFile("junit", null, temp.toFile());
- assertThat(testFile.delete()).as("Delete should succeed").isTrue();
+ File testFile = temp.resolve("data.parquet").toFile();
try (FileAppender<Record> writer = getParquetWriter(writeSchema,
testFile)) {
writer.addAll(expected);
@@ -390,8 +389,7 @@ public class TestParquetVectorizedReads extends
AvroDataTestBase {
optional(102, "float_data", Types.FloatType.get()),
optional(103, "decimal_data", Types.DecimalType.of(10, 5)));
- File dataFile = File.createTempFile("junit", null, temp.toFile());
- assertThat(dataFile.delete()).as("Delete should succeed").isTrue();
+ File dataFile = temp.resolve("data.parquet").toFile();
Iterable<Record> data =
generateData(writeSchema, 30000, 0L,
RandomData.DEFAULT_NULL_PERCENTAGE, IDENTITY);
try (FileAppender<Record> writer = getParquetWriter(writeSchema,
dataFile)) {
@@ -423,8 +421,7 @@ public class TestParquetVectorizedReads extends
AvroDataTestBase {
optional(107, "string_data", Types.StringType.get()),
optional(108, "binary_data", Types.BinaryType.get()));
- File dataFile = File.createTempFile("junit", null, temp.toFile());
- assertThat(dataFile.delete()).as("Delete should succeed").isTrue();
+ File dataFile = temp.resolve("data.parquet").toFile();
Iterable<Record> data =
generateData(schema, 30000, 0L, RandomData.DEFAULT_NULL_PERCENTAGE,
IDENTITY);
try (FileAppender<Record> writer = getParquetV2Writer(schema, dataFile)) {
@@ -437,8 +434,7 @@ public class TestParquetVectorizedReads extends
AvroDataTestBase {
public void testUnsupportedReadsForParquetV2() throws Exception {
// Some types use delta encoding and which are not supported for
vectorized reads
Schema schema = new Schema(SUPPORTED_PRIMITIVES.fields());
- File dataFile = File.createTempFile("junit", null, temp.toFile());
- assertThat(dataFile.delete()).as("Delete should succeed").isTrue();
+ File dataFile = temp.resolve("data.parquet").toFile();
Iterable<Record> data =
generateData(schema, 30000, 0L, RandomData.DEFAULT_NULL_PERCENTAGE,
IDENTITY);
try (FileAppender<Record> writer = getParquetV2Writer(schema, dataFile)) {
@@ -456,8 +452,7 @@ public class TestParquetVectorizedReads extends
AvroDataTestBase {
int numRows = 1;
Schema schema = new Schema(optional(100, "uuid", Types.UUIDType.get()));
- File dataFile = File.createTempFile("junit", null, temp.toFile());
- assertThat(dataFile.delete()).as("Delete should succeed").isTrue();
+ File dataFile = temp.resolve("data.parquet").toFile();
Iterable<Record> data = generateData(schema, numRows, 0L, 0, IDENTITY);
try (FileAppender<Record> writer = getParquetV2Writer(schema, dataFile)) {
writer.addAll(data);