This is an automated email from the ASF dual-hosted git repository.

kevinjqliu pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/iceberg.git


The following commit(s) were added to refs/heads/main by this push:
     new ad116d604d Docs: Fix several nit issues in docs (#15419)
ad116d604d is described below

commit ad116d604dbeb2a4b4115ecba639d22c476ddcee
Author: Anshul Baliga <[email protected]>
AuthorDate: Mon Feb 23 23:58:33 2026 +0530

    Docs: Fix several nit issues in docs (#15419)
---
 docs/docs/aws.md                 | 4 ++--
 docs/docs/flink-configuration.md | 2 +-
 docs/docs/flink-maintenance.md   | 2 +-
 docs/docs/spark-procedures.md    | 2 +-
 format/spec.md                   | 2 +-
 5 files changed, 6 insertions(+), 6 deletions(-)

diff --git a/docs/docs/aws.md b/docs/docs/aws.md
index 9c979b9da8..1fe8674012 100644
--- a/docs/docs/aws.md
+++ b/docs/docs/aws.md
@@ -270,9 +270,9 @@ Read [this AWS 
documentation](https://docs.aws.amazon.com/AmazonRDS/latest/UserG
 With all the available options, we offer the following guidelines when 
choosing the right catalog to use for your application:
 
 1. if your organization has an existing Glue metastore or plans to use the AWS 
analytics ecosystem including Glue, [Athena](https://aws.amazon.com/athena), 
[EMR](https://aws.amazon.com/emr), [Redshift](https://aws.amazon.com/redshift) 
and [LakeFormation](https://aws.amazon.com/lake-formation), Glue catalog 
provides the easiest integration.
-2. if your application requires frequent updates to table or high read and 
write throughput (e.g. streaming write), Glue and DynamoDB catalog provides the 
best performance through optimistic locking.
+2. if your application requires frequent updates to table or high read and 
write throughput (e.g. streaming write), Glue and DynamoDB catalogs provide the 
best performance through optimistic locking.
 3. if you would like to enforce access control for tables in a catalog, Glue 
tables can be managed as an [IAM 
resource](https://docs.aws.amazon.com/service-authorization/latest/reference/list_awsglue.html),
 whereas DynamoDB catalog tables can only be managed through [item-level 
permission](https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/specifying-conditions.html)
 which is much more complicated.
-4. if you would like to query tables based on table property information 
without the need to scan the entire catalog, DynamoDB catalog allows you to 
build secondary indexes for any arbitrary property field and provide efficient 
query performance.
+4. if you would like to query tables based on table property information 
without the need to scan the entire catalog, DynamoDB catalog allows you to 
build secondary indexes for any arbitrary property field and provides efficient 
query performance.
 5. if you would like to have the benefit of DynamoDB catalog while also 
connect to Glue, you can enable [DynamoDB stream with Lambda 
trigger](https://docs.aws.amazon.com/amazondynamodb/latest/developerguide/Streams.Lambda.Tutorial.html)
 to asynchronously update your Glue metastore with table information in the 
DynamoDB catalog.
 6. if your organization already maintains an existing relational database in 
RDS or uses [serverless Aurora](https://aws.amazon.com/rds/aurora/serverless/) 
to manage tables, the JDBC catalog provides the easiest integration.
 
diff --git a/docs/docs/flink-configuration.md b/docs/docs/flink-configuration.md
index 41d3df0745..f30b422888 100644
--- a/docs/docs/flink-configuration.md
+++ b/docs/docs/flink-configuration.md
@@ -123,7 +123,7 @@ env.getConfig()
 | include-column-stats          | connector.iceberg.include-column-stats       
   | N/A                          | false                            | Create a 
new scan from this that loads the column stats with each data file. Column 
stats include: value count, null value count, lower bounds, and upper bounds.   
                                                                                
                                                                                
                   [...]
 | max-planning-snapshot-count   | 
connector.iceberg.max-planning-snapshot-count   | N/A                          
| Integer.MAX_VALUE                | Max number of snapshots limited per split 
enumeration. Applicable only to streaming read.                                 
                                                                                
                                                                                
                                                             [...]
 | limit                         | connector.iceberg.limit                      
   | N/A                          | -1                               | Limited 
output number of rows.                                                          
                                                                                
                                                                                
                                                                                
               [...]
-| max-allowed-planning-failures | 
connector.iceberg.max-allowed-planning-failures | N/A                          
| 3                                | Max allowed consecutive failures for scan 
planning before failing the job. Set to -1 for never failing the job for scan 
planing failure.                                                                
                                                                                
                                                               [...]
+| max-allowed-planning-failures | 
connector.iceberg.max-allowed-planning-failures | N/A                          
| 3                                | Max allowed consecutive failures for scan 
planning before failing the job. Set to -1 for never failing the job for scan 
planning failure.                                                               
                                                                                
                                                               [...]
 | watermark-column              | connector.iceberg.watermark-column           
   | N/A                          | null                             | 
Specifies the watermark column to use for watermark generation. If this option 
is present, the `splitAssignerFactory` will be overridden with 
`OrderedSplitAssignerFactory`.                                                  
                                                                                
                                         [...]
 | watermark-column-time-unit    | connector.iceberg.watermark-column-time-unit 
   | N/A                          | TimeUnit.MICROSECONDS            | 
Specifies the watermark time unit to use for watermark generation. The possible 
values are  DAYS, HOURS, MINUTES, SECONDS, MILLISECONDS, MICROSECONDS, 
NANOSECONDS.                                                                    
                                                                                
                                [...]
 
diff --git a/docs/docs/flink-maintenance.md b/docs/docs/flink-maintenance.md
index 37fb9a7a84..1afa35034f 100644
--- a/docs/docs/flink-maintenance.md
+++ b/docs/docs/flink-maintenance.md
@@ -224,7 +224,7 @@ env.execute("Table Maintenance Job");
 
 | Method                                   | Description                       
                                                                                
                                                                                
                                                                                
                                                                                
      | Default Value           | Type               |
 
|------------------------------------------|-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|-------------------------|--------------------|
-| `location(string)`                       | The location to start the 
recursive listing the candidate files for removal.                              
                                                                                
                                                                                
                                                                                
              | Table's location        | String             |
+| `location(string)`                       | The location to start the 
recursive listing of the candidate files for removal.                           
                                                                                
                                                                                
                                                                                
                 | Table's location        | String             |
 | `usePrefixListing(boolean)`              | When true, use prefix-based file 
listing via the SupportsPrefixOperations interface. The Table FileIO 
implementation must support SupportsPrefixOperations when this flag is 
enabled.(Note: Setting it to False will use a recursive method to obtain file 
information. If the underlying storage is object storage, it will repeatedly 
call the API to get the path.)  | False                   | boolean            |
 | `prefixMismatchMode(PrefixMismatchMode)` | Action behavior when location 
prefixes (schemes/authorities) mismatch: <ul><li>ERROR - throw an exception. 
</li><li>IGNORE - no action.</li><li>DELETE - delete files.</li></ul>           
                                                                                
                                                                                
             | ERROR                   | PrefixMismatchMode |
 | `equalSchemes(Map<String, String>)`      | Mapping of file system schemes to 
be considered equal. Key is a comma-separated list of schemes and value is a 
scheme                                                                          
                                                                                
                                                                                
         | "s3n"=>"s3","s3a"=>"s3" | Map<String,String> |  
diff --git a/docs/docs/spark-procedures.md b/docs/docs/spark-procedures.md
index c040f49ef8..4544f6533d 100644
--- a/docs/docs/spark-procedures.md
+++ b/docs/docs/spark-procedures.md
@@ -401,7 +401,7 @@ Iceberg can compact data files in parallel using Spark with 
the `rewriteDataFile
 | `max-concurrent-file-group-rewrites` | 5 | Maximum number of file groups to 
be simultaneously rewritten |
 | `partial-progress.enabled` | false | Enable committing groups of files prior 
to the entire rewrite completing |
 | `partial-progress.max-commits` | 10 | Maximum amount of commits that this 
rewrite is allowed to produce if partial progress is enabled |
-| `partial-progress.max-failed-commits` | value of 
`partital-progress.max-commits` | Maximum amount of failed commits allowed 
before job failure, if partial progress is enabled |
+| `partial-progress.max-failed-commits` | value of 
`partial-progress.max-commits` | Maximum amount of failed commits allowed 
before job failure, if partial progress is enabled |
 | `use-starting-sequence-number` | true | Use the sequence number of the 
snapshot at compaction start time instead of that of the newly produced 
snapshot |
 | `rewrite-job-order` | none | Force the rewrite job order based on the value. 
<ul><li>If rewrite-job-order=bytes-asc, then rewrite the smallest job groups 
first.</li><li>If rewrite-job-order=bytes-desc, then rewrite the largest job 
groups first.</li><li>If rewrite-job-order=files-asc, then rewrite the job 
groups with the least files first.</li><li>If rewrite-job-order=files-desc, 
then rewrite the job groups with the most files first.</li><li>If 
rewrite-job-order=none, then rewrite job g [...]
 | `target-file-size-bytes` | 536870912 (512 MB, default value of 
`write.target-file-size-bytes` from [table 
properties](configuration.md#write-properties)) | Target output file size |
diff --git a/format/spec.md b/format/spec.md
index 43348d856c..0d3c79762c 100644
--- a/format/spec.md
+++ b/format/spec.md
@@ -1485,7 +1485,7 @@ Older versions of the reference implementation can read 
tables with transforms u
 
 ### Sort Orders
 
-Sort orders are serialized as a list of JSON object, each of which contains 
the following fields:
+Sort orders are serialized as a list of JSON objects, each of which contains 
the following fields:
 
 |Field|JSON representation|Example|
 |--- |--- |--- |

Reply via email to