mbutrovich commented on code in PR #3294: URL: https://github.com/apache/datafusion-comet/pull/3294#discussion_r2733814326
########## dev/benchmarks/README.md: ########## @@ -73,3 +73,76 @@ Generating charts: ```shell python3 generate-comparison.py --benchmark tpch --labels "Spark 3.5.3" "Comet 0.9.0" "Gluten 1.4.0" --title "TPC-H @ 100 GB (single executor, 8 cores, local Parquet files)" spark-tpch-1752338506381.json comet-tpch-1752337818039.json gluten-tpch-1752337474344.json ``` + +## Iceberg Benchmarking + +Comet includes native Iceberg support via iceberg-rust integration. This enables benchmarking TPC-H queries +against Iceberg tables with native scan acceleration. + +### Prerequisites + +Download the Iceberg Spark runtime JAR: + +```shell +wget https://repo1.maven.org/maven2/org/apache/iceberg/iceberg-spark-runtime-3.5_2.12/1.8.1/iceberg-spark-runtime-3.5_2.12-1.8.1.jar +export ICEBERG_JAR=/path/to/iceberg-spark-runtime-3.5_2.12-1.8.1.jar +``` + +### Create Iceberg TPC-H tables + +Convert existing Parquet TPC-H data to Iceberg format: + +```shell +export ICEBERG_WAREHOUSE=/mnt/bigdata/iceberg-warehouse + +$SPARK_HOME/bin/spark-submit \ + --master $SPARK_MASTER \ + --jars $ICEBERG_JAR \ + --conf spark.driver.memory=8G \ + --conf spark.executor.instances=1 \ + --conf spark.executor.cores=8 \ + --conf spark.cores.max=8 \ + --conf spark.executor.memory=16g \ + --conf spark.sql.catalog.local=org.apache.iceberg.spark.SparkCatalog \ + --conf spark.sql.catalog.local.type=hadoop \ Review Comment: This hardcodes the catalog, which IMO for benchmarking is fine. Above you have `ICEBERG_CATALOG=${ICEBERG_CATALOG:-local}`. I'd be consistent, and frankly I'd just get rid of `ICEBERG_CATALOG` above. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: [email protected] For queries about this service, please contact Infrastructure at: [email protected] --------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]
