This is an automated email from the ASF dual-hosted git repository.

github-bot pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/datafusion-comet.git


The following commit(s) were added to refs/heads/asf-site by this push:
     new 290e67e9b Publish built docs triggered by 
7022983198f00469c6bb830b38f13c9e83513528
290e67e9b is described below

commit 290e67e9be6faf700ead302f86967035f25e1162
Author: github-actions[bot] <github-actions[bot]@users.noreply.github.com>
AuthorDate: Tue Feb 17 19:23:44 2026 +0000

    Publish built docs triggered by 7022983198f00469c6bb830b38f13c9e83513528
---
 _sources/contributor-guide/roadmap.md.txt | 19 +++++++++++++++----
 _sources/user-guide/latest/configs.md.txt |  1 +
 contributor-guide/roadmap.html            | 14 +++++++++++---
 searchindex.js                            |  2 +-
 user-guide/latest/configs.html            | 16 ++++++++++------
 5 files changed, 38 insertions(+), 14 deletions(-)

diff --git a/_sources/contributor-guide/roadmap.md.txt 
b/_sources/contributor-guide/roadmap.md.txt
index 3abe6f3f1..ce9c41416 100644
--- a/_sources/contributor-guide/roadmap.md.txt
+++ b/_sources/contributor-guide/roadmap.md.txt
@@ -26,11 +26,11 @@ helpful to have a roadmap for some of the major items that 
require coordination
 
 ### Iceberg Integration
 
-Iceberg integration is still a work-in-progress ([#2060]), with major 
improvements expected in the next few
-releases. The default `auto` scan mode now uses `native_iceberg_compat` 
instead of `native_comet`, enabling
-support for complex types.
+Iceberg tables reads are now fully native, powered by a scan operator backed 
by Iceberg-rust ([#2528]). We anticipate
+major improvements expected in the next few releases, including bringing 
Iceberg table format V3 features (_e.g._,
+encryption) to the reader.
 
-[#2060]: https://github.com/apache/datafusion-comet/issues/2060
+[#2528]: https://github.com/apache/datafusion-comet/pull/2528
 
 ### Spark 4.0 Support
 
@@ -40,6 +40,17 @@ more Spark SQL tests and fully implementing ANSI support 
([#313]) for all suppor
 [#313]: https://github.com/apache/datafusion-comet/issues/313
 [#1637]: https://github.com/apache/datafusion-comet/issues/1637
 
+### Dynamic Partition Pruning
+
+Iceberg table scans support Dynamic Partition Pruning (DPP) filters generated 
by Spark's `PlanDynamicPruningFilters`
+optimizer rule ([#3349]). However, we still need to bring this functionality 
to our Parquet reader. Furthermore,
+Spark's `PlanAdaptiveDynamicPruningFilters` optimizer rule runs after Comet's 
rules, so DPP with Adaptive Query
+Execution requires a redesign of Comet's plan translation. We are focused on 
implementing DPP to keep Comet competitive
+with benchmarks that benefit from this feature like TPC-DS. This effort can be 
tracked at [#3510].
+
+[#3349]: https://github.com/apache/datafusion-comet/pull/3349
+[#3510]: https://github.com/apache/datafusion-comet/issues/3510
+
 ### Removing the native_comet scan implementation
 
 The `native_comet` scan implementation is now deprecated and will be removed 
in a future release ([#2186], [#2177]).
diff --git a/_sources/user-guide/latest/configs.md.txt 
b/_sources/user-guide/latest/configs.md.txt
index a4e389b43..6ae0ec6fb 100644
--- a/_sources/user-guide/latest/configs.md.txt
+++ b/_sources/user-guide/latest/configs.md.txt
@@ -139,6 +139,7 @@ These settings can be used to determine which parts of the 
plan are accelerated
 | `spark.comet.debug.memory` | When enabled, log all native memory pool 
interactions. For more information, refer to the Comet Debugging Guide 
(https://datafusion.apache.org/comet/contributor-guide/debugging.html). | false 
|
 | `spark.comet.exec.onHeap.enabled` | Whether to allow Comet to run in on-heap 
mode. Required for running Spark SQL tests. It can be overridden by the 
environment variable `ENABLE_COMET_ONHEAP`. | false |
 | `spark.comet.exec.onHeap.memoryPool` | The type of memory pool to be used 
for Comet native execution when running Spark in on-heap mode. Available pool 
types are `greedy`, `fair_spill`, `greedy_task_shared`, 
`fair_spill_task_shared`, `greedy_global`, `fair_spill_global`, and 
`unbounded`. | greedy_task_shared |
+| `spark.comet.exec.respectDataFusionConfigs` | Development and testing 
configuration option to allow DataFusion configs set in Spark configuration 
settings starting with `spark.comet.datafusion.` to be passed into native 
execution. | false |
 | `spark.comet.memoryOverhead` | The amount of additional memory to be 
allocated per executor process for Comet, in MiB, when running Spark in on-heap 
mode. | 1024 MiB |
 | `spark.comet.parquet.write.enabled` | Whether to enable native Parquet write 
through Comet. When enabled, Comet will intercept Parquet write operations and 
execute them natively. This feature is highly experimental and only partially 
implemented. It should not be used in production. | false |
 | `spark.comet.scan.csv.v2.enabled` | Whether to use the native Comet V2 CSV 
reader for improved performance. Default: false (uses standard Spark CSV 
reader) Experimental: Performance benefits are workload-dependent. | false |
diff --git a/contributor-guide/roadmap.html b/contributor-guide/roadmap.html
index 91709dd5f..be85cf9ef 100644
--- a/contributor-guide/roadmap.html
+++ b/contributor-guide/roadmap.html
@@ -463,15 +463,23 @@ helpful to have a roadmap for some of the major items 
that require coordination
 <h2>Major Initiatives<a class="headerlink" href="#major-initiatives" 
title="Link to this heading">#</a></h2>
 <section id="iceberg-integration">
 <h3>Iceberg Integration<a class="headerlink" href="#iceberg-integration" 
title="Link to this heading">#</a></h3>
-<p>Iceberg integration is still a work-in-progress (<a class="reference 
external" 
href="https://github.com/apache/datafusion-comet/issues/2060";>#2060</a>), with 
major improvements expected in the next few
-releases. The default <code class="docutils literal notranslate"><span 
class="pre">auto</span></code> scan mode now uses <code class="docutils literal 
notranslate"><span class="pre">native_iceberg_compat</span></code> instead of 
<code class="docutils literal notranslate"><span 
class="pre">native_comet</span></code>, enabling
-support for complex types.</p>
+<p>Iceberg tables reads are now fully native, powered by a scan operator 
backed by Iceberg-rust (<a class="reference external" 
href="https://github.com/apache/datafusion-comet/pull/2528";>#2528</a>). We 
anticipate
+major improvements expected in the next few releases, including bringing 
Iceberg table format V3 features (<em>e.g.</em>,
+encryption) to the reader.</p>
 </section>
 <section id="spark-4-0-support">
 <h3>Spark 4.0 Support<a class="headerlink" href="#spark-4-0-support" 
title="Link to this heading">#</a></h3>
 <p>Comet has experimental support for Spark 4.0, but there is more work to do 
(<a class="reference external" 
href="https://github.com/apache/datafusion-comet/issues/1637";>#1637</a>), such 
as enabling
 more Spark SQL tests and fully implementing ANSI support (<a class="reference 
external" 
href="https://github.com/apache/datafusion-comet/issues/313";>#313</a>) for all 
supported expressions.</p>
 </section>
+<section id="dynamic-partition-pruning">
+<h3>Dynamic Partition Pruning<a class="headerlink" 
href="#dynamic-partition-pruning" title="Link to this heading">#</a></h3>
+<p>Iceberg table scans support Dynamic Partition Pruning (DPP) filters 
generated by Spark’s <code class="docutils literal notranslate"><span 
class="pre">PlanDynamicPruningFilters</span></code>
+optimizer rule (<a class="reference external" 
href="https://github.com/apache/datafusion-comet/pull/3349";>#3349</a>). 
However, we still need to bring this functionality to our Parquet reader. 
Furthermore,
+Spark’s <code class="docutils literal notranslate"><span 
class="pre">PlanAdaptiveDynamicPruningFilters</span></code> optimizer rule runs 
after Comet’s rules, so DPP with Adaptive Query
+Execution requires a redesign of Comet’s plan translation. We are focused on 
implementing DPP to keep Comet competitive
+with benchmarks that benefit from this feature like TPC-DS. This effort can be 
tracked at <a class="reference external" 
href="https://github.com/apache/datafusion-comet/issues/3510";>#3510</a>.</p>
+</section>
 <section id="removing-the-native-comet-scan-implementation">
 <h3>Removing the native_comet scan implementation<a class="headerlink" 
href="#removing-the-native-comet-scan-implementation" title="Link to this 
heading">#</a></h3>
 <p>The <code class="docutils literal notranslate"><span 
class="pre">native_comet</span></code> scan implementation is now deprecated 
and will be removed in a future release (<a class="reference external" 
href="https://github.com/apache/datafusion-comet/issues/2186";>#2186</a>, <a 
class="reference external" 
href="https://github.com/apache/datafusion-comet/issues/2177";>#2177</a>).
diff --git a/searchindex.js b/searchindex.js
index 9904bc795..1d278c792 100644
--- a/searchindex.js
+++ b/searchindex.js
@@ -1 +1 @@
-Search.setIndex({"alltitles": {"1. Format Your Code": [[12, 
"format-your-code"]], "1. Install Comet": [[22, "install-comet"]], "1. Native 
Operators (nativeExecs map)": [[4, "native-operators-nativeexecs-map"]], "2. 
Build and Verify": [[12, "build-and-verify"]], "2. Clone Spark and Apply Diff": 
[[22, "clone-spark-and-apply-diff"]], "2. Sink Operators (sinks map)": [[4, 
"sink-operators-sinks-map"]], "3. Comet JVM Operators": [[4, 
"comet-jvm-operators"]], "3. Run Clippy (Recommended)": [[12 [...]
\ No newline at end of file
+Search.setIndex({"alltitles": {"1. Format Your Code": [[12, 
"format-your-code"]], "1. Install Comet": [[22, "install-comet"]], "1. Native 
Operators (nativeExecs map)": [[4, "native-operators-nativeexecs-map"]], "2. 
Build and Verify": [[12, "build-and-verify"]], "2. Clone Spark and Apply Diff": 
[[22, "clone-spark-and-apply-diff"]], "2. Sink Operators (sinks map)": [[4, 
"sink-operators-sinks-map"]], "3. Comet JVM Operators": [[4, 
"comet-jvm-operators"]], "3. Run Clippy (Recommended)": [[12 [...]
\ No newline at end of file
diff --git a/user-guide/latest/configs.html b/user-guide/latest/configs.html
index c640704da..465e3e6ba 100644
--- a/user-guide/latest/configs.html
+++ b/user-guide/latest/configs.html
@@ -806,27 +806,31 @@ under the License.
 <td><p>The type of memory pool to be used for Comet native execution when 
running Spark in on-heap mode. Available pool types are <code class="docutils 
literal notranslate"><span class="pre">greedy</span></code>, <code 
class="docutils literal notranslate"><span 
class="pre">fair_spill</span></code>, <code class="docutils literal 
notranslate"><span class="pre">greedy_task_shared</span></code>, <code 
class="docutils literal notranslate"><span 
class="pre">fair_spill_task_shared</span></code> [...]
 <td><p>greedy_task_shared</p></td>
 </tr>
-<tr class="row-odd"><td><p><code class="docutils literal notranslate"><span 
class="pre">spark.comet.memoryOverhead</span></code></p></td>
+<tr class="row-odd"><td><p><code class="docutils literal notranslate"><span 
class="pre">spark.comet.exec.respectDataFusionConfigs</span></code></p></td>
+<td><p>Development and testing configuration option to allow DataFusion 
configs set in Spark configuration settings starting with <code class="docutils 
literal notranslate"><span class="pre">spark.comet.datafusion.</span></code> to 
be passed into native execution.</p></td>
+<td><p>false</p></td>
+</tr>
+<tr class="row-even"><td><p><code class="docutils literal notranslate"><span 
class="pre">spark.comet.memoryOverhead</span></code></p></td>
 <td><p>The amount of additional memory to be allocated per executor process 
for Comet, in MiB, when running Spark in on-heap mode.</p></td>
 <td><p>1024 MiB</p></td>
 </tr>
-<tr class="row-even"><td><p><code class="docutils literal notranslate"><span 
class="pre">spark.comet.parquet.write.enabled</span></code></p></td>
+<tr class="row-odd"><td><p><code class="docutils literal notranslate"><span 
class="pre">spark.comet.parquet.write.enabled</span></code></p></td>
 <td><p>Whether to enable native Parquet write through Comet. When enabled, 
Comet will intercept Parquet write operations and execute them natively. This 
feature is highly experimental and only partially implemented. It should not be 
used in production.</p></td>
 <td><p>false</p></td>
 </tr>
-<tr class="row-odd"><td><p><code class="docutils literal notranslate"><span 
class="pre">spark.comet.scan.csv.v2.enabled</span></code></p></td>
+<tr class="row-even"><td><p><code class="docutils literal notranslate"><span 
class="pre">spark.comet.scan.csv.v2.enabled</span></code></p></td>
 <td><p>Whether to use the native Comet V2 CSV reader for improved performance. 
Default: false (uses standard Spark CSV reader) Experimental: Performance 
benefits are workload-dependent.</p></td>
 <td><p>false</p></td>
 </tr>
-<tr class="row-even"><td><p><code class="docutils literal notranslate"><span 
class="pre">spark.comet.sparkToColumnar.enabled</span></code></p></td>
+<tr class="row-odd"><td><p><code class="docutils literal notranslate"><span 
class="pre">spark.comet.sparkToColumnar.enabled</span></code></p></td>
 <td><p>Whether to enable Spark to Arrow columnar conversion. When this is 
turned on, Comet will convert operators in <code class="docutils literal 
notranslate"><span 
class="pre">spark.comet.sparkToColumnar.supportedOperatorList</span></code> 
into Arrow columnar format before processing. This is an experimental feature 
and has known issues with non-UTC timezones.</p></td>
 <td><p>false</p></td>
 </tr>
-<tr class="row-odd"><td><p><code class="docutils literal notranslate"><span 
class="pre">spark.comet.sparkToColumnar.supportedOperatorList</span></code></p></td>
+<tr class="row-even"><td><p><code class="docutils literal notranslate"><span 
class="pre">spark.comet.sparkToColumnar.supportedOperatorList</span></code></p></td>
 <td><p>A comma-separated list of operators that will be converted to Arrow 
columnar format when <code class="docutils literal notranslate"><span 
class="pre">spark.comet.sparkToColumnar.enabled</span></code> is true.</p></td>
 <td><p>Range,InMemoryTableScan,RDDScan</p></td>
 </tr>
-<tr class="row-even"><td><p><code class="docutils literal notranslate"><span 
class="pre">spark.comet.testing.strict</span></code></p></td>
+<tr class="row-odd"><td><p><code class="docutils literal notranslate"><span 
class="pre">spark.comet.testing.strict</span></code></p></td>
 <td><p>Experimental option to enable strict testing, which will fail tests 
that could be more comprehensive, such as checking for a specific fallback 
reason. It can be overridden by the environment variable <code class="docutils 
literal notranslate"><span 
class="pre">ENABLE_COMET_STRICT_TESTING</span></code>.</p></td>
 <td><p>false</p></td>
 </tr>


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to