nevi-me commented on pull request #7309:
URL: https://github.com/apache/arrow/pull/7309#issuecomment-636794722
> It also opens the way to have other formats support
Do you mean other compression formats?
> However, this needs for the user to do manual seek (if the BufRead is on a
File) and does not rely on the Builder. Unless a new builder with buffer is
added.
I suppose the broader question to ask would be whether we'd like to support
reading (and writing) compressed data. There have also been some changes on the
`arrow::csv` side, such as allowing inference of multiple files, which might
also be convenient to have in `arrow::json`.
What's your opinion on the 2 above? We need not implement them as part of
this PR, but I suppose some user feedback is helpful :)
I'm still pro returning the reader back to the start, or is there a
performance impact in doing so? I wouldn't want to place the burden of seeking
on the user, because I'd expect the common inference case to be getting the
schema then reading the file.
```rust
// return the reader seek back to the start
reader.into_inner().seek(SeekFrom::Start(0))?;
```
----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
For queries about this service, please contact Infrastructure at:
[email protected]