thisisnic commented on a change in pull request #10765:
URL: https://github.com/apache/arrow/pull/10765#discussion_r679808079
##########
File path: r/vignettes/dataset.Rmd
##########
@@ -313,27 +330,29 @@ instead of a file path, or simply concatenate them like
`big_dataset <- c(ds1, d
As you can see, querying a large dataset can be made quite fast by storage in
an
efficient binary columnar format like Parquet or Feather and partitioning
based on
-columns commonly used for filtering. However, we don't always get our data
delivered
-to us that way. Sometimes we start with one giant CSV. Our first step in
analyzing data
+columns commonly used for filtering. However, data isn't always stored that
way.
+Sometimes you might start with one giant CSV. The first step in analyzing data
is cleaning is up and reshaping it into a more usable form.
-The `write_dataset()` function allows you to take a Dataset or other tabular
data object---an Arrow `Table` or `RecordBatch`, or an R `data.frame`---and
write it to a different file format, partitioned into multiple files.
+The `write_dataset()` function allows you to take a Dataset or another tabular
+data object - an Arrow Table or RecordBatch, or an R data frame - and write
Review comment:
I'm leaning towards this interpretation too, so have added guidance in
`STYLE.md`.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]