alamb commented on code in PR #60:
URL: https://github.com/apache/datafusion-site/pull/60#discussion_r2001947189


##########
content/blog/2025-03-18-parquet-pruning.md:
##########
@@ -0,0 +1,111 @@
+---
+layout: post
+title: Parquet pruning in DataFusion: Read Only What Matters
+date: 2025-03-18
+author: Xiangpeng Hao
+categories: [performance]
+---
+
+<!--
+{% comment %}
+Licensed to the Apache Software Foundation (ASF) under one or more
+contributor license agreements.  See the NOTICE file distributed with
+this work for additional information regarding copyright ownership.
+The ASF licenses this file to you under the Apache License, Version 2.0
+(the "License"); you may not use this file except in compliance with
+the License.  You may obtain a copy of the License at
+
+http://www.apache.org/licenses/LICENSE-2.0
+
+Unless required by applicable law or agreed to in writing, software
+distributed under the License is distributed on an "AS IS" BASIS,
+WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+See the License for the specific language governing permissions and
+limitations under the License.
+{% endcomment %}
+-->
+
+
+_Editor's Note: This blog was first published on [Xiangpeng Hao's blog]. 
Thanks to [InfluxData] for sponsoring this work as part of his PhD funding._
+
+[Xiangpeng Hao's blog]: https://blog.xiangpeng.systems/posts/parquet-to-arrow/
+[InfluxData]: https://www.influxdata.com/
+<hr/>
+
+Parquet has become the industry standard for storing columnar data, and 
reading Parquet efficiently is crucial for query performance.
+
+To optimize this, DataFusion implements advanced Parquet support for effective 
data pruning and decoding.

Review Comment:
   ```suggestion
   To optimize this, [Apache DataFusion] implements advanced Parquet support 
for effective data pruning and decoding.
   
   [Apache DataFusion]: https://datafusion.apache.org/
   ```



##########
content/blog/2025-03-18-parquet-pruning.md:
##########
@@ -0,0 +1,111 @@
+---
+layout: post
+title: Parquet pruning in DataFusion: Read Only What Matters
+date: 2025-03-18
+author: Xiangpeng Hao
+categories: [performance]
+---
+
+<!--
+{% comment %}
+Licensed to the Apache Software Foundation (ASF) under one or more
+contributor license agreements.  See the NOTICE file distributed with
+this work for additional information regarding copyright ownership.
+The ASF licenses this file to you under the Apache License, Version 2.0
+(the "License"); you may not use this file except in compliance with
+the License.  You may obtain a copy of the License at
+
+http://www.apache.org/licenses/LICENSE-2.0
+
+Unless required by applicable law or agreed to in writing, software
+distributed under the License is distributed on an "AS IS" BASIS,
+WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+See the License for the specific language governing permissions and
+limitations under the License.
+{% endcomment %}
+-->
+
+
+_Editor's Note: This blog was first published on [Xiangpeng Hao's blog]. 
Thanks to [InfluxData] for sponsoring this work as part of his PhD funding._
+
+[Xiangpeng Hao's blog]: https://blog.xiangpeng.systems/posts/parquet-to-arrow/
+[InfluxData]: https://www.influxdata.com/
+<hr/>
+
+Parquet has become the industry standard for storing columnar data, and 
reading Parquet efficiently is crucial for query performance.

Review Comment:
   I recommend adding links and referring to this as Apache Parquet and Apache 
DataFusion on first mention
   
   ```suggestion
   [Apache Parquet] has become the industry standard for storing columnar data, 
and reading Parquet efficiently is crucial for query performance.
   
   [Apache Parquet]: https://parquet.apache.org/
   ```



##########
content/blog/2025-03-18-parquet-pruning.md:
##########
@@ -0,0 +1,111 @@
+---
+layout: post
+title: Parquet pruning in DataFusion: Read Only What Matters
+date: 2025-03-18
+author: Xiangpeng Hao
+categories: [performance]
+---
+
+<!--
+{% comment %}
+Licensed to the Apache Software Foundation (ASF) under one or more
+contributor license agreements.  See the NOTICE file distributed with
+this work for additional information regarding copyright ownership.
+The ASF licenses this file to you under the Apache License, Version 2.0
+(the "License"); you may not use this file except in compliance with
+the License.  You may obtain a copy of the License at
+
+http://www.apache.org/licenses/LICENSE-2.0
+
+Unless required by applicable law or agreed to in writing, software
+distributed under the License is distributed on an "AS IS" BASIS,
+WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+See the License for the specific language governing permissions and
+limitations under the License.
+{% endcomment %}
+-->
+
+
+_Editor's Note: This blog was first published on [Xiangpeng Hao's blog]. 
Thanks to [InfluxData] for sponsoring this work as part of his PhD funding._
+
+[Xiangpeng Hao's blog]: https://blog.xiangpeng.systems/posts/parquet-to-arrow/
+[InfluxData]: https://www.influxdata.com/
+<hr/>
+
+Parquet has become the industry standard for storing columnar data, and 
reading Parquet efficiently is crucial for query performance.
+
+To optimize this, DataFusion implements advanced Parquet support for effective 
data pruning and decoding.
+
+However, achieving high performance adds complexity, and this is no exception. 
This post provides an overview of the techniques used in DataFusion to 
selectively read Parquet files.
+
+### The pipeline
+The diagram below illustrates the Parquet reading pipeline in DataFusion, 
highlighting how data flows through various pruning stages before being 
converted to Arrow format:

Review Comment:
   ```suggestion
   The diagram below illustrates the [Parquet reading pipeline] in DataFusion, 
highlighting how data flows through various pruning stages before being 
converted to Arrow format:
   
   [Parquet reading pipeline]: 
https://docs.rs/datafusion/46.0.0/datafusion/datasource/physical_plan/parquet/source/struct.ParquetSource.html```



##########
content/blog/2025-03-18-parquet-pruning.md:
##########
@@ -0,0 +1,111 @@
+---
+layout: post
+title: Parquet pruning in DataFusion: Read Only What Matters
+date: 2025-03-18
+author: Xiangpeng Hao
+categories: [performance]
+---
+
+<!--
+{% comment %}
+Licensed to the Apache Software Foundation (ASF) under one or more
+contributor license agreements.  See the NOTICE file distributed with
+this work for additional information regarding copyright ownership.
+The ASF licenses this file to you under the Apache License, Version 2.0
+(the "License"); you may not use this file except in compliance with
+the License.  You may obtain a copy of the License at
+
+http://www.apache.org/licenses/LICENSE-2.0
+
+Unless required by applicable law or agreed to in writing, software
+distributed under the License is distributed on an "AS IS" BASIS,
+WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+See the License for the specific language governing permissions and
+limitations under the License.
+{% endcomment %}
+-->
+
+
+_Editor's Note: This blog was first published on [Xiangpeng Hao's blog]. 
Thanks to [InfluxData] for sponsoring this work as part of his PhD funding._
+
+[Xiangpeng Hao's blog]: https://blog.xiangpeng.systems/posts/parquet-to-arrow/
+[InfluxData]: https://www.influxdata.com/
+<hr/>
+
+Parquet has become the industry standard for storing columnar data, and 
reading Parquet efficiently is crucial for query performance.
+
+To optimize this, DataFusion implements advanced Parquet support for effective 
data pruning and decoding.
+
+However, achieving high performance adds complexity, and this is no exception. 
This post provides an overview of the techniques used in DataFusion to 
selectively read Parquet files.
+
+### The pipeline
+The diagram below illustrates the Parquet reading pipeline in DataFusion, 
highlighting how data flows through various pruning stages before being 
converted to Arrow format:
+
+<img src="/blog/images/parquet-pruning/read-parquet.jpg" alt="Parquet pruning 
pipeline in DataFusion" width="100%" class="img-responsive">
+
+
+#### Background: Parquet file structure
+As shown in the figure above, each Parquet file has multiple row groups. Each 
row group contains a set of columns, and each column contains a set of pages.
+
+Pages are the smallest units of data in Parquet files and typically contain 
compressed and encoded values for a specific column. This hierarchical 
structure enables efficient columnar access and forms the foundation for the 
pruning techniques we'll discuss.
+
+Check out [Querying Parquet with Millisecond 
Latency](https://www.influxdata.com/blog/querying-parquet-millisecond-latency/) 
for more details on the Parquet file structure.
+
+#### 1. Read metadata
+DataFusion first reads the Parquet metadata to understand the data in the 
file. 
+Metadata often includes data schema, the exact location of each row group and 
column chunk, and their corresponding statistics (e.g., min/max values).
+It also optionally includes [page-level 
stats](https://parquet.apache.org/docs/file-format/pageindex/) and [Bloom 
filters](https://www.influxdata.com/blog/using-parquets-bloom-filters/).
+This information is used to prune the file before reading the actual data.
+
+[Fetching 
metadata](https://github.com/apache/datafusion/blob/31701b8dc9c6486856c06a29a32107d9f4549cec/datafusion/core/src/datasource/physical_plan/parquet/reader.rs#L118)
 requires up to two network requests: one to read the footer size from the end 
of the file, and another to read the footer itself. 
+
+[Decoding 
metadata](https://www.influxdata.com/blog/how-good-parquet-wide-tables/) is 
generally fast since it only requires parsing a small amount of data. However, 
for tables with hundreds or thousands of columns, the metadata can become quite 
large and decoding it can become a bottleneck. This is particularly noticeable 
when scanning many small files.
+
+Reading metadata is latency-critical, so DataFusion allows users to cache 
metadata through the 
[ParquetFileReaderFactory](https://github.com/apache/datafusion/blob/31701b8dc9c6486856c06a29a32107d9f4549cec/datafusion/core/src/datasource/physical_plan/parquet/reader.rs#L39)
 trait.
+
+
+#### 2. Prune by projection
+The simplest yet perhaps most effective pruning is to read only the columns 
that are needed.
+This is because queries usually don't select all columns, e.g., `SELECT a FROM 
table` only reads column `a`.
+As a **columnar** format, Parquet allows DataFusion to [only 
read](https://github.com/apache/datafusion/blob/31701b8dc9c6486856c06a29a32107d9f4549cec/datafusion/core/src/datasource/physical_plan/parquet/mod.rs#L778)
 the **columns** that are needed.
+
+This projection pruning happens at the column level and can dramatically 
reduce I/O when working with wide tables where queries typically access only a 
small subset of columns.
+
+
+#### 3. Prune by row group stats and Bloom filters
+Each row group has [basic 
stats](https://github.com/apache/datafusion/blob/31701b8dc9c6486856c06a29a32107d9f4549cec/datafusion/core/src/physical_optimizer/pruning.rs#L81)
 like min/max values for each column.
+DataFusion applies the query predicates to these stats to prune row groups, 
e.g., `SELECT * FROM table WHERE a > 10` will only read row groups where `a` 
has a max value greater than 10.
+
+Sometimes min/max stats are too simple to prune effectively, so Parquet also 
supports [Bloom 
filters](https://www.influxdata.com/blog/using-parquets-bloom-filters/). 
DataFusion [uses Bloom filters when 
available](https://github.com/apache/datafusion/blob/31701b8dc9c6486856c06a29a32107d9f4549cec/datafusion/core/src/datasource/physical_plan/parquet/opener.rs#L202).
+
+Bloom filters are particularly effective for equality predicates (`WHERE a = 
10`) and can significantly reduce the number of row groups that need to be read 
for point queries or queries with highly selective predicates.
+
+
+#### 4. Prune by page stats
+Parquet optionally supports [page-level 
stats](https://github.com/apache/parquet-format/blob/master/PageIndex.md) -- 
similar to row group stats but more fine-grained.
+DataFusion implements [page 
pruning](https://github.com/apache/datafusion/blob/31701b8dc9c6486856c06a29a32107d9f4549cec/datafusion/core/src/datasource/physical_plan/parquet/opener.rs#L219)
 when the stats are present.
+
+Page-level pruning provides an additional layer of filtering after row group 
pruning. It allows DataFusion to skip individual pages within a row group, 
further reducing the amount of data that needs to be read and decoded.
+
+
+#### 5. Read from storage 
+Now we (hopefully) have pruned the Parquet file into small ranges of bytes, 
i.e., the [Access 
Plan](https://github.com/apache/datafusion/blob/76a7789ace33ced54c973fa0d5fc9d1866e1bf19/datafusion/datasource-parquet/src/access_plan.rs#L86).
+The last step is to [make 
requests](https://github.com/apache/datafusion/blob/31701b8dc9c6486856c06a29a32107d9f4549cec/datafusion/core/src/datasource/physical_plan/parquet/reader.rs#L103)
 to fetch those bytes and decode them into Arrow RecordBatch. 
+
+
+### Bonus: filter pushdown
+So far we have discussed techniques that prune the Parquet file using only the 
metadata, i.e., before reading the actual data.
+
+Filter pushdown, also known as predicate pushdown, is a technique that prunes 
data during scanning, with filters being generated and applied in the Parquet 
reader.

Review Comment:
   ```suggestion
   Filter pushdown, also known as predicate pushdown or late materialization, 
is a technique that prunes data during scanning, with filters being generated 
and applied in the Parquet reader.
   ```



##########
content/blog/2025-03-18-parquet-pruning.md:
##########
@@ -0,0 +1,111 @@
+---
+layout: post
+title: Parquet pruning in DataFusion: Read Only What Matters
+date: 2025-03-18
+author: Xiangpeng Hao
+categories: [performance]
+---
+
+<!--
+{% comment %}
+Licensed to the Apache Software Foundation (ASF) under one or more
+contributor license agreements.  See the NOTICE file distributed with
+this work for additional information regarding copyright ownership.
+The ASF licenses this file to you under the Apache License, Version 2.0
+(the "License"); you may not use this file except in compliance with
+the License.  You may obtain a copy of the License at
+
+http://www.apache.org/licenses/LICENSE-2.0
+
+Unless required by applicable law or agreed to in writing, software
+distributed under the License is distributed on an "AS IS" BASIS,
+WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+See the License for the specific language governing permissions and
+limitations under the License.
+{% endcomment %}
+-->
+
+
+_Editor's Note: This blog was first published on [Xiangpeng Hao's blog]. 
Thanks to [InfluxData] for sponsoring this work as part of his PhD funding._
+
+[Xiangpeng Hao's blog]: https://blog.xiangpeng.systems/posts/parquet-to-arrow/
+[InfluxData]: https://www.influxdata.com/
+<hr/>
+
+Parquet has become the industry standard for storing columnar data, and 
reading Parquet efficiently is crucial for query performance.
+
+To optimize this, DataFusion implements advanced Parquet support for effective 
data pruning and decoding.
+
+However, achieving high performance adds complexity, and this is no exception. 
This post provides an overview of the techniques used in DataFusion to 
selectively read Parquet files.
+
+### The pipeline
+The diagram below illustrates the Parquet reading pipeline in DataFusion, 
highlighting how data flows through various pruning stages before being 
converted to Arrow format:
+
+<img src="/blog/images/parquet-pruning/read-parquet.jpg" alt="Parquet pruning 
pipeline in DataFusion" width="100%" class="img-responsive">
+
+
+#### Background: Parquet file structure
+As shown in the figure above, each Parquet file has multiple row groups. Each 
row group contains a set of columns, and each column contains a set of pages.
+
+Pages are the smallest units of data in Parquet files and typically contain 
compressed and encoded values for a specific column. This hierarchical 
structure enables efficient columnar access and forms the foundation for the 
pruning techniques we'll discuss.
+
+Check out [Querying Parquet with Millisecond 
Latency](https://www.influxdata.com/blog/querying-parquet-millisecond-latency/) 
for more details on the Parquet file structure.
+
+#### 1. Read metadata
+DataFusion first reads the Parquet metadata to understand the data in the 
file. 
+Metadata often includes data schema, the exact location of each row group and 
column chunk, and their corresponding statistics (e.g., min/max values).
+It also optionally includes [page-level 
stats](https://parquet.apache.org/docs/file-format/pageindex/) and [Bloom 
filters](https://www.influxdata.com/blog/using-parquets-bloom-filters/).
+This information is used to prune the file before reading the actual data.
+
+[Fetching 
metadata](https://github.com/apache/datafusion/blob/31701b8dc9c6486856c06a29a32107d9f4549cec/datafusion/core/src/datasource/physical_plan/parquet/reader.rs#L118)
 requires up to two network requests: one to read the footer size from the end 
of the file, and another to read the footer itself. 
+
+[Decoding 
metadata](https://www.influxdata.com/blog/how-good-parquet-wide-tables/) is 
generally fast since it only requires parsing a small amount of data. However, 
for tables with hundreds or thousands of columns, the metadata can become quite 
large and decoding it can become a bottleneck. This is particularly noticeable 
when scanning many small files.
+
+Reading metadata is latency-critical, so DataFusion allows users to cache 
metadata through the 
[ParquetFileReaderFactory](https://github.com/apache/datafusion/blob/31701b8dc9c6486856c06a29a32107d9f4549cec/datafusion/core/src/datasource/physical_plan/parquet/reader.rs#L39)
 trait.
+
+
+#### 2. Prune by projection
+The simplest yet perhaps most effective pruning is to read only the columns 
that are needed.
+This is because queries usually don't select all columns, e.g., `SELECT a FROM 
table` only reads column `a`.
+As a **columnar** format, Parquet allows DataFusion to [only 
read](https://github.com/apache/datafusion/blob/31701b8dc9c6486856c06a29a32107d9f4549cec/datafusion/core/src/datasource/physical_plan/parquet/mod.rs#L778)
 the **columns** that are needed.
+
+This projection pruning happens at the column level and can dramatically 
reduce I/O when working with wide tables where queries typically access only a 
small subset of columns.
+
+
+#### 3. Prune by row group stats and Bloom filters
+Each row group has [basic 
stats](https://github.com/apache/datafusion/blob/31701b8dc9c6486856c06a29a32107d9f4549cec/datafusion/core/src/physical_optimizer/pruning.rs#L81)
 like min/max values for each column.
+DataFusion applies the query predicates to these stats to prune row groups, 
e.g., `SELECT * FROM table WHERE a > 10` will only read row groups where `a` 
has a max value greater than 10.
+
+Sometimes min/max stats are too simple to prune effectively, so Parquet also 
supports [Bloom 
filters](https://www.influxdata.com/blog/using-parquets-bloom-filters/). 
DataFusion [uses Bloom filters when 
available](https://github.com/apache/datafusion/blob/31701b8dc9c6486856c06a29a32107d9f4549cec/datafusion/core/src/datasource/physical_plan/parquet/opener.rs#L202).
+
+Bloom filters are particularly effective for equality predicates (`WHERE a = 
10`) and can significantly reduce the number of row groups that need to be read 
for point queries or queries with highly selective predicates.
+
+
+#### 4. Prune by page stats
+Parquet optionally supports [page-level 
stats](https://github.com/apache/parquet-format/blob/master/PageIndex.md) -- 
similar to row group stats but more fine-grained.
+DataFusion implements [page 
pruning](https://github.com/apache/datafusion/blob/31701b8dc9c6486856c06a29a32107d9f4549cec/datafusion/core/src/datasource/physical_plan/parquet/opener.rs#L219)
 when the stats are present.
+
+Page-level pruning provides an additional layer of filtering after row group 
pruning. It allows DataFusion to skip individual pages within a row group, 
further reducing the amount of data that needs to be read and decoded.
+
+
+#### 5. Read from storage 
+Now we (hopefully) have pruned the Parquet file into small ranges of bytes, 
i.e., the [Access 
Plan](https://github.com/apache/datafusion/blob/76a7789ace33ced54c973fa0d5fc9d1866e1bf19/datafusion/datasource-parquet/src/access_plan.rs#L86).
+The last step is to [make 
requests](https://github.com/apache/datafusion/blob/31701b8dc9c6486856c06a29a32107d9f4549cec/datafusion/core/src/datasource/physical_plan/parquet/reader.rs#L103)
 to fetch those bytes and decode them into Arrow RecordBatch. 
+
+
+### Bonus: filter pushdown

Review Comment:
   Maybe instead of `bonus` this would be clearer like
   
   ```suggestion
   ### Preview of Coming Attractions: filter pushdown
   ```



##########
content/blog/2025-03-18-parquet-pruning.md:
##########
@@ -0,0 +1,111 @@
+---
+layout: post
+title: Parquet pruning in DataFusion: Read Only What Matters
+date: 2025-03-18
+author: Xiangpeng Hao
+categories: [performance]
+---
+
+<!--
+{% comment %}
+Licensed to the Apache Software Foundation (ASF) under one or more
+contributor license agreements.  See the NOTICE file distributed with
+this work for additional information regarding copyright ownership.
+The ASF licenses this file to you under the Apache License, Version 2.0
+(the "License"); you may not use this file except in compliance with
+the License.  You may obtain a copy of the License at
+
+http://www.apache.org/licenses/LICENSE-2.0
+
+Unless required by applicable law or agreed to in writing, software
+distributed under the License is distributed on an "AS IS" BASIS,
+WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+See the License for the specific language governing permissions and
+limitations under the License.
+{% endcomment %}
+-->
+
+
+_Editor's Note: This blog was first published on [Xiangpeng Hao's blog]. 
Thanks to [InfluxData] for sponsoring this work as part of his PhD funding._
+
+[Xiangpeng Hao's blog]: https://blog.xiangpeng.systems/posts/parquet-to-arrow/
+[InfluxData]: https://www.influxdata.com/
+<hr/>
+
+Parquet has become the industry standard for storing columnar data, and 
reading Parquet efficiently is crucial for query performance.
+
+To optimize this, DataFusion implements advanced Parquet support for effective 
data pruning and decoding.
+
+However, achieving high performance adds complexity, and this is no exception. 
This post provides an overview of the techniques used in DataFusion to 
selectively read Parquet files.
+
+### The pipeline
+The diagram below illustrates the Parquet reading pipeline in DataFusion, 
highlighting how data flows through various pruning stages before being 
converted to Arrow format:
+
+<img src="/blog/images/parquet-pruning/read-parquet.jpg" alt="Parquet pruning 
pipeline in DataFusion" width="100%" class="img-responsive">
+
+
+#### Background: Parquet file structure
+As shown in the figure above, each Parquet file has multiple row groups. Each 
row group contains a set of columns, and each column contains a set of pages.
+
+Pages are the smallest units of data in Parquet files and typically contain 
compressed and encoded values for a specific column. This hierarchical 
structure enables efficient columnar access and forms the foundation for the 
pruning techniques we'll discuss.
+
+Check out [Querying Parquet with Millisecond 
Latency](https://www.influxdata.com/blog/querying-parquet-millisecond-latency/) 
for more details on the Parquet file structure.
+
+#### 1. Read metadata
+DataFusion first reads the Parquet metadata to understand the data in the 
file. 
+Metadata often includes data schema, the exact location of each row group and 
column chunk, and their corresponding statistics (e.g., min/max values).
+It also optionally includes [page-level 
stats](https://parquet.apache.org/docs/file-format/pageindex/) and [Bloom 
filters](https://www.influxdata.com/blog/using-parquets-bloom-filters/).
+This information is used to prune the file before reading the actual data.
+
+[Fetching 
metadata](https://github.com/apache/datafusion/blob/31701b8dc9c6486856c06a29a32107d9f4549cec/datafusion/core/src/datasource/physical_plan/parquet/reader.rs#L118)
 requires up to two network requests: one to read the footer size from the end 
of the file, and another to read the footer itself. 

Review Comment:
   While it is true it requires up to 2 network requests, I thought it was 
almost always one
   
   However, I dug around the code and it seems like by default Datafusion does 
in fact do 2 network requests 🤔 
   
   https://datafusion.apache.org/user-guide/configs.html
   
   Maybe we should consider changing that default (like read 512K or something)



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: github-unsubscr...@datafusion.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


---------------------------------------------------------------------
To unsubscribe, e-mail: github-unsubscr...@datafusion.apache.org
For additional commands, e-mail: github-h...@datafusion.apache.org

Reply via email to