nevi-me commented on pull request #7309:
URL: https://github.com/apache/arrow/pull/7309#issuecomment-636794722


   > It also opens the way to have other formats support
   
   Do you mean other compression formats?
   
   > However, this needs for the user to do manual seek (if the BufRead is on a 
File) and does not rely on the Builder. Unless a new builder with buffer is 
added.
   
   I suppose the broader question to ask would be whether we'd like to support 
reading (and writing) compressed data. There have also been some changes on the 
`arrow::csv` side, such as allowing inference of multiple files, which might 
also be convenient to have in `arrow::json`.
   
   What's your opinion on the 2 above? We need not implement them as part of 
this PR, but I suppose some user feedback is helpful :)
   
   I'm still pro returning the reader back to the start, or is there a 
performance impact in doing so? I wouldn't want to place the burden of seeking 
on the user, because I'd expect the common inference case to be getting the 
schema then reading the file.
   
   ```rust
       // return the reader seek back to the start      
       reader.into_inner().seek(SeekFrom::Start(0))?;
   ```


----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
[email protected]


Reply via email to