yjshen commented on a change in pull request #1905: URL: https://github.com/apache/arrow-datafusion/pull/1905#discussion_r818717630
########## File path: datafusion/src/datasource/object_store/mod.rs ########## @@ -39,27 +39,34 @@ use crate::error::{DataFusionError, Result}; /// Note that the dynamic dispatch on the reader might /// have some performance impacts. #[async_trait] -pub trait ObjectReader: Send + Sync { +pub trait ObjectReader: Read + Seek + Send { /// Get reader for a part [start, start + length] in the file asynchronously async fn chunk_reader(&self, start: u64, length: usize) -> Result<Box<dyn AsyncRead>>; - /// Get reader for a part [start, start + length] in the file - fn sync_chunk_reader( - &self, - start: u64, - length: usize, - ) -> Result<Box<dyn Read + Send + Sync>>; - - /// Get reader for the entire file - fn sync_reader(&self) -> Result<Box<dyn Read + Send + Sync>> { - self.sync_chunk_reader(0, self.length() as usize) - } Review comment: Hi @rdettai long time no see! I think the chunk is only a term from parquet reader implementation. And we are always using chunks from one same file sequentially in DataFusion. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: github-unsubscr...@arrow.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org