This is an automated email from the ASF dual-hosted git repository.

alamb pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/arrow-datafusion.git


The following commit(s) were added to refs/heads/main by this push:
     new 06e9f53637 Minor: consolidate test data (#6217)
06e9f53637 is described below

commit 06e9f53637f20dd91bef43b74942ec36c38c22d5
Author: Andrew Lamb <[email protected]>
AuthorDate: Fri May 5 09:19:19 2023 -0400

    Minor: consolidate test data (#6217)
    
    * Minor: consolidate test data directories
    
    * move json data
    
    * move parquet data
---
 datafusion/core/src/datasource/file_format/json.rs          |   4 ++--
 datafusion/core/src/physical_plan/file_format/csv.rs        |   2 +-
 datafusion/core/src/physical_plan/file_format/json.rs       |   4 ++--
 datafusion/core/src/physical_plan/file_format/parquet.rs    |   2 +-
 datafusion/core/tests/{jsons => data}/1.json                |   0
 datafusion/core/tests/{jsons => data}/2.json                |   0
 datafusion/core/tests/{jsons => data}/3.json                |   0
 datafusion/core/tests/{csv => data}/corrupt.csv             |   0
 .../tests/{parquet => }/data/repeat_much.snappy.parquet     | Bin
 .../core/tests/{jsons => data}/schema_infer_limit.json      |   0
 .../core/tests/{parquet => }/data/test_binary.parquet       | Bin
 .../core/tests/{parquet => }/data/timestamp_with_tz.parquet | Bin
 datafusion/core/tests/sql/json.rs                           |   4 ++--
 datafusion/core/tests/sql/order.rs                          |   2 +-
 datafusion/core/tests/sql/parquet.rs                        |   4 ++--
 15 files changed, 11 insertions(+), 11 deletions(-)

diff --git a/datafusion/core/src/datasource/file_format/json.rs 
b/datafusion/core/src/datasource/file_format/json.rs
index a66edab888..21cd22f070 100644
--- a/datafusion/core/src/datasource/file_format/json.rs
+++ b/datafusion/core/src/datasource/file_format/json.rs
@@ -256,7 +256,7 @@ mod tests {
         projection: Option<Vec<usize>>,
         limit: Option<usize>,
     ) -> Result<Arc<dyn ExecutionPlan>> {
-        let filename = "tests/jsons/2.json";
+        let filename = "tests/data/2.json";
         let format = JsonFormat::default();
         scan_format(state, &format, ".", filename, projection, limit).await
     }
@@ -266,7 +266,7 @@ mod tests {
         let session = SessionContext::new();
         let ctx = session.state();
         let store = Arc::new(LocalFileSystem::new()) as _;
-        let filename = "tests/jsons/schema_infer_limit.json";
+        let filename = "tests/data/schema_infer_limit.json";
         let format = JsonFormat::default().with_schema_infer_max_rec(Some(3));
 
         let file_schema = format
diff --git a/datafusion/core/src/physical_plan/file_format/csv.rs 
b/datafusion/core/src/physical_plan/file_format/csv.rs
index e2d2bb8ef7..e7633807e0 100644
--- a/datafusion/core/src/physical_plan/file_format/csv.rs
+++ b/datafusion/core/src/physical_plan/file_format/csv.rs
@@ -785,7 +785,7 @@ mod tests {
         let options = CsvReadOptions::default()
             .schema_infer_max_records(2)
             .has_header(true);
-        let df = ctx.read_csv("tests/csv/corrupt.csv", options).await?;
+        let df = ctx.read_csv("tests/data/corrupt.csv", options).await?;
         let tmp_dir = TempDir::new()?;
         let out_dir = tmp_dir.as_ref().to_str().unwrap().to_string() + "/out";
         let e = df
diff --git a/datafusion/core/src/physical_plan/file_format/json.rs 
b/datafusion/core/src/physical_plan/file_format/json.rs
index d122fd78b4..bb686b55a0 100644
--- a/datafusion/core/src/physical_plan/file_format/json.rs
+++ b/datafusion/core/src/physical_plan/file_format/json.rs
@@ -309,7 +309,7 @@ mod tests {
 
     use super::*;
 
-    const TEST_DATA_BASE: &str = "tests/jsons";
+    const TEST_DATA_BASE: &str = "tests/data";
 
     async fn prepare_store(
         state: &SessionState,
@@ -707,7 +707,7 @@ mod tests {
         let options = CsvReadOptions::default()
             .schema_infer_max_records(2)
             .has_header(true);
-        let df = ctx.read_csv("tests/csv/corrupt.csv", options).await?;
+        let df = ctx.read_csv("tests/data/corrupt.csv", options).await?;
         let tmp_dir = TempDir::new()?;
         let out_dir = tmp_dir.as_ref().to_str().unwrap().to_string() + "/out";
         let e = df
diff --git a/datafusion/core/src/physical_plan/file_format/parquet.rs 
b/datafusion/core/src/physical_plan/file_format/parquet.rs
index 0afb819d43..641be002b6 100644
--- a/datafusion/core/src/physical_plan/file_format/parquet.rs
+++ b/datafusion/core/src/physical_plan/file_format/parquet.rs
@@ -940,7 +940,7 @@ mod tests {
         let options = CsvReadOptions::default()
             .schema_infer_max_records(2)
             .has_header(true);
-        let df = ctx.read_csv("tests/csv/corrupt.csv", options).await?;
+        let df = ctx.read_csv("tests/data/corrupt.csv", options).await?;
         let tmp_dir = TempDir::new()?;
         let out_dir = tmp_dir.as_ref().to_str().unwrap().to_string() + "/out";
         let e = df
diff --git a/datafusion/core/tests/jsons/1.json 
b/datafusion/core/tests/data/1.json
similarity index 100%
rename from datafusion/core/tests/jsons/1.json
rename to datafusion/core/tests/data/1.json
diff --git a/datafusion/core/tests/jsons/2.json 
b/datafusion/core/tests/data/2.json
similarity index 100%
rename from datafusion/core/tests/jsons/2.json
rename to datafusion/core/tests/data/2.json
diff --git a/datafusion/core/tests/jsons/3.json 
b/datafusion/core/tests/data/3.json
similarity index 100%
rename from datafusion/core/tests/jsons/3.json
rename to datafusion/core/tests/data/3.json
diff --git a/datafusion/core/tests/csv/corrupt.csv 
b/datafusion/core/tests/data/corrupt.csv
similarity index 100%
rename from datafusion/core/tests/csv/corrupt.csv
rename to datafusion/core/tests/data/corrupt.csv
diff --git a/datafusion/core/tests/parquet/data/repeat_much.snappy.parquet 
b/datafusion/core/tests/data/repeat_much.snappy.parquet
similarity index 100%
rename from datafusion/core/tests/parquet/data/repeat_much.snappy.parquet
rename to datafusion/core/tests/data/repeat_much.snappy.parquet
diff --git a/datafusion/core/tests/jsons/schema_infer_limit.json 
b/datafusion/core/tests/data/schema_infer_limit.json
similarity index 100%
rename from datafusion/core/tests/jsons/schema_infer_limit.json
rename to datafusion/core/tests/data/schema_infer_limit.json
diff --git a/datafusion/core/tests/parquet/data/test_binary.parquet 
b/datafusion/core/tests/data/test_binary.parquet
similarity index 100%
rename from datafusion/core/tests/parquet/data/test_binary.parquet
rename to datafusion/core/tests/data/test_binary.parquet
diff --git a/datafusion/core/tests/parquet/data/timestamp_with_tz.parquet 
b/datafusion/core/tests/data/timestamp_with_tz.parquet
similarity index 100%
rename from datafusion/core/tests/parquet/data/timestamp_with_tz.parquet
rename to datafusion/core/tests/data/timestamp_with_tz.parquet
diff --git a/datafusion/core/tests/sql/json.rs 
b/datafusion/core/tests/sql/json.rs
index 10fcdfda20..8608305f15 100644
--- a/datafusion/core/tests/sql/json.rs
+++ b/datafusion/core/tests/sql/json.rs
@@ -17,7 +17,7 @@
 
 use super::*;
 
-const TEST_DATA_BASE: &str = "tests/jsons";
+const TEST_DATA_BASE: &str = "tests/data";
 
 #[tokio::test]
 async fn json_query() {
@@ -92,7 +92,7 @@ async fn json_explain() {
             \n  CoalescePartitionsExec\
             \n    AggregateExec: mode=Partial, gby=[], aggr=[COUNT(UInt8(1))]\
             \n      RepartitionExec: partitioning=RoundRobinBatch(NUM_CORES), 
input_partitions=1\
-            \n        JsonExec: file_groups={1 group: 
[[WORKING_DIR/tests/jsons/2.json]]}, projection=[a]\n",
+            \n        JsonExec: file_groups={1 group: 
[[WORKING_DIR/tests/data/2.json]]}, projection=[a]\n",
         ],
     ];
     assert_eq!(expected, actual);
diff --git a/datafusion/core/tests/sql/order.rs 
b/datafusion/core/tests/sql/order.rs
index a85fee776d..fa6ac61244 100644
--- a/datafusion/core/tests/sql/order.rs
+++ b/datafusion/core/tests/sql/order.rs
@@ -25,7 +25,7 @@ use test_utils::{batches_to_vec, partitions_to_sorted_vec};
 #[tokio::test]
 async fn sort_with_lots_of_repetition_values() -> Result<()> {
     let ctx = SessionContext::new();
-    let filename = "tests/parquet/data/repeat_much.snappy.parquet";
+    let filename = "tests/data/repeat_much.snappy.parquet";
 
     ctx.register_parquet("rep", filename, ParquetReadOptions::default())
         .await?;
diff --git a/datafusion/core/tests/sql/parquet.rs 
b/datafusion/core/tests/sql/parquet.rs
index d2998209a2..532a0414f9 100644
--- a/datafusion/core/tests/sql/parquet.rs
+++ b/datafusion/core/tests/sql/parquet.rs
@@ -151,7 +151,7 @@ async fn fixed_size_binary_columns() {
     let ctx = SessionContext::new();
     ctx.register_parquet(
         "t0",
-        "tests/parquet/data/test_binary.parquet",
+        "tests/data/test_binary.parquet",
         ParquetReadOptions::default(),
     )
     .await
@@ -170,7 +170,7 @@ async fn window_fn_timestamp_tz() {
     let ctx = SessionContext::new();
     ctx.register_parquet(
         "t0",
-        "tests/parquet/data/timestamp_with_tz.parquet",
+        "tests/data/timestamp_with_tz.parquet",
         ParquetReadOptions::default(),
     )
     .await

Reply via email to