This is an automated email from the ASF dual-hosted git repository.
aokolnychyi pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/iceberg.git
The following commit(s) were added to refs/heads/main by this push:
new de3125afe6 Spark 4.1: Fix IcebergSource doc (#15359)
de3125afe6 is described below
commit de3125afe64fc2b171a52b6e884c72f901e3cba1
Author: Anton Okolnychyi <[email protected]>
AuthorDate: Wed Feb 18 23:19:57 2026 -0800
Spark 4.1: Fix IcebergSource doc (#15359)
---
.../org/apache/iceberg/spark/source/IcebergSource.java | 18 +++++++-----------
1 file changed, 7 insertions(+), 11 deletions(-)
diff --git
a/spark/v4.1/spark/src/main/java/org/apache/iceberg/spark/source/IcebergSource.java
b/spark/v4.1/spark/src/main/java/org/apache/iceberg/spark/source/IcebergSource.java
index a0462e8f89..28282eb1cc 100644
---
a/spark/v4.1/spark/src/main/java/org/apache/iceberg/spark/source/IcebergSource.java
+++
b/spark/v4.1/spark/src/main/java/org/apache/iceberg/spark/source/IcebergSource.java
@@ -48,19 +48,15 @@ import org.apache.spark.sql.types.StructType;
import org.apache.spark.sql.util.CaseInsensitiveStringMap;
/**
- * The IcebergSource loads/writes tables with format "iceberg". It can load
paths and tables.
+ * Data source for reading and writing Iceberg tables using the "iceberg"
format.
*
- * <p>How paths/tables are loaded when using
spark.read().format("iceberg").load(table)
+ * <p>The `path` parameter provided by Spark is resolved in the following
priority order:
*
- * <p>table = "file:///path/to/table" -> loads a HadoopTable at given path
table = "tablename"
- * -> loads currentCatalog.currentNamespace.tablename table =
"catalog.tablename" -> load
- * "tablename" from the specified catalog. table = "namespace.tablename" ->
load
- * "namespace.tablename" from current catalog table =
"catalog.namespace.tablename" ->
- * "namespace.tablename" from the specified catalog. table =
"namespace1.namespace2.tablename" ->
- * load "namespace1.namespace2.tablename" from current catalog
- *
- * <p>The above list is in order of priority. For example: a matching catalog
will take priority
- * over any namespace resolution.
+ * <ol>
+ * <li>Rewrite key - If `path` is a rewrite key, load a table from the
rewrite catalog
+ * <li>Table location - If `path` contains "/", load a table at the
specified location
+ * <li>Catalog identifier - Otherwise resolve `path` as an identifier per
Spark rules
+ * </ol>
*/
public class IcebergSource
implements DataSourceRegister, SupportsCatalogOptions,
SessionConfigSupport {