This is an automated email from the ASF dual-hosted git repository.

agrove pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/datafusion-ballista.git


The following commit(s) were added to refs/heads/main by this push:
     new 343a0999 remove support for HDFS ... (#1073)
343a0999 is described below

commit 343a0999d2c8713e9a38948c1c07d87250e0e283
Author: Marko Milenković <[email protected]>
AuthorDate: Sat Oct 12 22:50:27 2024 +0100

    remove support for HDFS ... (#1073)
---
 README.md                                      | 2 +-
 ballista/client/Cargo.toml                     | 2 --
 ballista/core/Cargo.toml                       | 4 ----
 ballista/core/src/object_store_registry/mod.rs | 9 ---------
 4 files changed, 1 insertion(+), 16 deletions(-)

diff --git a/README.md b/README.md
index 80c10817..f614a16e 100644
--- a/README.md
+++ b/README.md
@@ -66,7 +66,7 @@ See the [architecture 
guide](docs/source/contributors-guide/architecture.md) for
 
 ## Features
 
-- Supports HDFS as well as cloud object stores. S3 is supported today and GCS 
and Azure support is planned.
+- Supports cloud object stores. S3 is supported today and GCS and Azure 
support is planned.
 - DataFrame and SQL APIs available from Python and Rust.
 - Clients can connect to a Ballista cluster using [Flight SQL][flight-sql].
 - JDBC support via Arrow Flight SQL JDBC Driver
diff --git a/ballista/client/Cargo.toml b/ballista/client/Cargo.toml
index 75bb7b48..2fb0c3d1 100644
--- a/ballista/client/Cargo.toml
+++ b/ballista/client/Cargo.toml
@@ -43,7 +43,5 @@ tokio = "1.0"
 [features]
 azure = ["ballista-core/azure"]
 default = []
-hdfs = ["ballista-core/hdfs"]
-hdfs3 = ["ballista-core/hdfs3"]
 s3 = ["ballista-core/s3"]
 standalone = ["ballista-executor", "ballista-scheduler"]
diff --git a/ballista/core/Cargo.toml b/ballista/core/Cargo.toml
index 884b3f60..519b3d9f 100644
--- a/ballista/core/Cargo.toml
+++ b/ballista/core/Cargo.toml
@@ -39,9 +39,6 @@ docsrs = []
 # Used for testing ONLY: causes all values to hash to the same value (test for 
collisions)
 force_hash_collisions = ["datafusion/force_hash_collisions"]
 gcs = ["object_store/gcp"]
-# Used to enable hdfs to be registered in the ObjectStoreRegistry by default
-hdfs = ["datafusion-objectstore-hdfs/hdfs"]
-hdfs3 = ["datafusion-objectstore-hdfs/hdfs3"]
 s3 = ["object_store/aws"]
 
 [dependencies]
@@ -52,7 +49,6 @@ bytes = "1.0"
 chrono = { version = "0.4", default-features = false }
 clap = { workspace = true }
 datafusion = { workspace = true }
-datafusion-objectstore-hdfs = { version = "0.1.4", default-features = false, 
optional = true }
 datafusion-proto = { workspace = true }
 datafusion-proto-common = { workspace = true }
 futures = "0.3"
diff --git a/ballista/core/src/object_store_registry/mod.rs 
b/ballista/core/src/object_store_registry/mod.rs
index 2e394a7c..aedccc5e 100644
--- a/ballista/core/src/object_store_registry/mod.rs
+++ b/ballista/core/src/object_store_registry/mod.rs
@@ -20,8 +20,6 @@ use datafusion::datasource::object_store::{
     DefaultObjectStoreRegistry, ObjectStoreRegistry,
 };
 use datafusion::execution::runtime_env::RuntimeConfig;
-#[cfg(any(feature = "hdfs", feature = "hdfs3"))]
-use datafusion_objectstore_hdfs::object_store::hdfs::HadoopFileSystem;
 #[cfg(feature = "s3")]
 use object_store::aws::AmazonS3Builder;
 #[cfg(feature = "azure")]
@@ -54,13 +52,6 @@ impl BallistaObjectStoreRegistry {
         &self,
         url: &Url,
     ) -> datafusion::error::Result<Arc<dyn ObjectStore>> {
-        #[cfg(any(feature = "hdfs", feature = "hdfs3"))]
-        {
-            if let Some(store) = HadoopFileSystem::new(url.as_str()) {
-                return Ok(Arc::new(store));
-            }
-        }
-
         #[cfg(feature = "s3")]
         {
             if url.as_str().starts_with("s3://") {


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to