This is an automated email from the ASF dual-hosted git repository.

lzljs3620320 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/paimon.git


The following commit(s) were added to refs/heads/master by this push:
     new da3ffd0215 [doc] Fix typos in docs and code comments (#7341)
da3ffd0215 is described below

commit da3ffd0215f5347cb859fd8077c7fa65286c84d6
Author: Jiajia Li <[email protected]>
AuthorDate: Thu Mar 5 15:47:45 2026 +0800

    [doc] Fix typos in docs and code comments (#7341)
---
 docs/content/_index.md                                                | 2 +-
 docs/content/append-table/blob.md                                     | 2 +-
 docs/content/append-table/incremental-clustering.md                   | 2 +-
 docs/content/concepts/overview.md                                     | 2 +-
 docs/content/ecosystem/starrocks.md                                   | 2 +-
 docs/content/flink/procedures.md                                      | 2 +-
 docs/content/learn-paimon/understand-files.md                         | 2 +-
 docs/content/maintenance/filesystems.md                               | 2 +-
 docs/content/primary-key-table/chain-table.md                         | 2 +-
 docs/content/primary-key-table/changelog-producer.md                  | 2 +-
 docs/content/primary-key-table/merge-engine/aggregation.md            | 2 +-
 docs/content/primary-key-table/merge-engine/partial-update.md         | 2 +-
 docs/content/program-api/cpp-api.md                                   | 2 +-
 docs/content/project/contributing.md                                  | 2 +-
 docs/content/spark/dataframe.md                                       | 2 +-
 docs/content/spark/procedures.md                                      | 2 +-
 docs/content/spark/quick-start.md                                     | 4 ++--
 docs/content/spark/sql-ddl.md                                         | 2 +-
 docs/layouts/shortcodes/generated/core_configuration.html             | 2 +-
 docs/layouts/shortcodes/generated/orc_configuration.html              | 2 +-
 paimon-api/src/main/java/org/apache/paimon/CoreOptions.java           | 2 +-
 paimon-format/src/main/java/org/apache/paimon/format/OrcOptions.java  | 2 +-
 .../java/org/apache/paimon/tools/ci/licensecheck/JarFileChecker.java  | 2 +-
 23 files changed, 24 insertions(+), 24 deletions(-)

diff --git a/docs/content/_index.md b/docs/content/_index.md
index 2803f6be95..216df04e4f 100644
--- a/docs/content/_index.md
+++ b/docs/content/_index.md
@@ -45,7 +45,7 @@ Paimon offers the following core capabilities:
 
 ## Try Paimon
 
-If you’re interested in playing around with Paimon, check out our
+If you're interested in playing around with Paimon, check out our
 quick start guide with [Flink]({{< ref "flink/quick-start" >}}) or [Spark]({{< 
ref "spark/quick-start" >}}). It provides a step by
 step introduction to the APIs and guides you through real applications.
 
diff --git a/docs/content/append-table/blob.md 
b/docs/content/append-table/blob.md
index 9d3135d8b2..ad6e1d3f50 100644
--- a/docs/content/append-table/blob.md
+++ b/docs/content/append-table/blob.md
@@ -576,7 +576,7 @@ public class BlobDescriptorExample {
             long fileSize = 2L * 1024 * 1024 * 1024;  // 2GB
 
             BlobDescriptor descriptor = new BlobDescriptor(externalUri, 0, 
fileSize);
-            // file io should be accessable to externalUri
+            // file io should be accessible to externalUri
             FileIO fileIO = Table.fileIO();
             UriReader uriReader = UriReader.fromFile(fileIO);
             Blob blob = Blob.fromDescriptor(uriReader, descriptor);
diff --git a/docs/content/append-table/incremental-clustering.md 
b/docs/content/append-table/incremental-clustering.md
index 0ca3462e16..5358c040d1 100644
--- a/docs/content/append-table/incremental-clustering.md
+++ b/docs/content/append-table/incremental-clustering.md
@@ -101,7 +101,7 @@ only support running Incremental Clustering in batch mode.
 
 To run a Incremental Clustering job, follow these instructions. 
 
-You don’t need to specify any clustering-related parameters when running 
Incremental Clustering,
+You don't need to specify any clustering-related parameters when running 
Incremental Clustering,
 these options are already defined as table options. If you need to change 
clustering settings, please update the corresponding table options.
 
 {{< tabs "incremental-clustering" >}}
diff --git a/docs/content/concepts/overview.md 
b/docs/content/concepts/overview.md
index 659e1f04d3..28fab2017e 100644
--- a/docs/content/concepts/overview.md
+++ b/docs/content/concepts/overview.md
@@ -56,7 +56,7 @@ For streaming engines like Apache Flink, there are typically 
three types of conn
   intermediate stages in this pipeline, to guarantee the latency stay
   within seconds.
 - OLAP system, such as ClickHouse, it receives processed data in
-  streaming fashion and serving user’s ad-hoc queries. 
+  streaming fashion and serving user's ad-hoc queries.
 - Batch storage, such as Apache Hive, it supports various operations
   of the traditional batch processing, including `INSERT OVERWRITE`.
 
diff --git a/docs/content/ecosystem/starrocks.md 
b/docs/content/ecosystem/starrocks.md
index f04fba437e..49554d54aa 100644
--- a/docs/content/ecosystem/starrocks.md
+++ b/docs/content/ecosystem/starrocks.md
@@ -79,7 +79,7 @@ SELECT * FROM paimon_catalog.test_db.partition_tbl$partitions;
 ## StarRocks to Paimon type mapping
 
 This section lists all supported type conversion between StarRocks and Paimon. 
-All StarRocks’s data types can be found in this doc [StarRocks Data type 
overview](https://docs.starrocks.io/docs/sql-reference/data-types/).
+All StarRocks's data types can be found in this doc [StarRocks Data type 
overview](https://docs.starrocks.io/docs/sql-reference/data-types/).
 
 <table class="table table-bordered">
     <thead>
diff --git a/docs/content/flink/procedures.md b/docs/content/flink/procedures.md
index d83a965005..38bef4ab63 100644
--- a/docs/content/flink/procedures.md
+++ b/docs/content/flink/procedures.md
@@ -703,7 +703,7 @@ All available procedures are listed below.
       <td>
          To expire partitions. Argument:
             <li>table: the target table identifier. Cannot be empty.</li>
-            <li>expiration_time: the expiration interval of a partition. A 
partition will be expired if it‘s lifetime is over this value. Partition time 
is extracted from the partition value.</li>
+            <li>expiration_time: the expiration interval of a partition. A 
partition will be expired if it's lifetime is over this value. Partition time 
is extracted from the partition value.</li>
             <li>timestamp_formatter: the formatter to format timestamp from 
string.</li>
             <li>timestamp_pattern: the pattern to get a timestamp from 
partitions.</li>
             <li>expire_strategy: specifies the expiration strategy for 
partition expiration, possible values: 'values-time' or 'update-time' , 
'values-time' as default.</li>
diff --git a/docs/content/learn-paimon/understand-files.md 
b/docs/content/learn-paimon/understand-files.md
index 1b85293472..a8a9ac52ce 100644
--- a/docs/content/learn-paimon/understand-files.md
+++ b/docs/content/learn-paimon/understand-files.md
@@ -496,5 +496,5 @@ Maybe you think the 5 files for the primary key table are 
actually okay, but the
 may have 50 small files in a single bucket, which is very difficult to accept. 
Worse still, partitions that
 are no longer active also keep so many small files.
 
-Configure ‘full-compaction.delta-commits’ perform full-compaction periodically 
in Flink writing. And it can ensure
+Configure 'full-compaction.delta-commits' perform full-compaction periodically 
in Flink writing. And it can ensure
 that partitions are full compacted before writing ends.
diff --git a/docs/content/maintenance/filesystems.md 
b/docs/content/maintenance/filesystems.md
index b5d41c89bf..38bcc090fe 100644
--- a/docs/content/maintenance/filesystems.md
+++ b/docs/content/maintenance/filesystems.md
@@ -391,7 +391,7 @@ Please refer to [Trino 
S3](https://trino.io/docs/current/object-storage/file-sys
 
 ### S3 Compliant Object Stores
 
-The S3 Filesystem also support using S3 compliant object stores such as MinIO, 
Tencent's COS and IBM’s Cloud Object
+The S3 Filesystem also support using S3 compliant object stores such as MinIO, 
Tencent's COS and IBM's Cloud Object
 Storage. Just configure your endpoint to the provider of the object store 
service.
 
 ```yaml
diff --git a/docs/content/primary-key-table/chain-table.md 
b/docs/content/primary-key-table/chain-table.md
index 197c79efa4..c8fc2f4f73 100644
--- a/docs/content/primary-key-table/chain-table.md
+++ b/docs/content/primary-key-table/chain-table.md
@@ -30,7 +30,7 @@ Chain table is a new capability for primary key tables that 
transforms how you p
 Imagine a scenario where you periodically store a full snapshot of data (for 
example, once a day), even 
 though only a small portion changes between snapshots. ODS binlog dump is a 
typical example of this pattern.
 
-Taking a daily binlog dump job as an example. A batch job merges yesterday’s 
full dataset with today’s 
+Taking a daily binlog dump job as an example. A batch job merges yesterday's 
full dataset with today's
 incremental changes to produce a new full dataset. This approach has two clear 
drawbacks:
 * Full computation: Merge operation includes all data, and it will involve 
shuffle, which results in poor performance.
 * Full storage: Store a full set of data every day, and the changed data 
usually accounts for a very small proportion.
diff --git a/docs/content/primary-key-table/changelog-producer.md 
b/docs/content/primary-key-table/changelog-producer.md
index ffbe6f27b2..0ea9ff857c 100644
--- a/docs/content/primary-key-table/changelog-producer.md
+++ b/docs/content/primary-key-table/changelog-producer.md
@@ -58,7 +58,7 @@ By specifying `'changelog-producer' = 'input'`, Paimon 
writers rely on their inp
 
 ## Lookup
 
-If your input can’t produce a complete changelog but you still want to get rid 
of the costly normalized operator, you
+If your input can't produce a complete changelog but you still want to get rid 
of the costly normalized operator, you
 may consider using the `'lookup'` changelog producer.
 
 By specifying `'changelog-producer' = 'lookup'`, Paimon will generate 
changelog through `'lookup'` during compaction (You can also enable [Async 
Compaction]({{< ref "primary-key-table/compaction#asynchronous-compaction" 
>}})). By default, lookup compaction is performed before committing written 
data unless disabled by `write-only` property.
diff --git a/docs/content/primary-key-table/merge-engine/aggregation.md 
b/docs/content/primary-key-table/merge-engine/aggregation.md
index fd8414ca1d..20270bcede 100644
--- a/docs/content/primary-key-table/merge-engine/aggregation.md
+++ b/docs/content/primary-key-table/merge-engine/aggregation.md
@@ -308,7 +308,7 @@ public static class BitmapContainsUDF extends 
ScalarFunction {
 
   Use `fields.<field-name>.nested-key=pk0,pk1,...` to specify the primary keys 
of the nested table. If no keys, row will be appended to array<row>.
 
-  Use `fields.<field-name>.count-limit=<Interger>` to specify the maximum 
number of rows in the nested table. When no nested-key, it will select data
+  Use `fields.<field-name>.count-limit=<Integer>` to specify the maximum 
number of rows in the nested table. When no nested-key, it will select data
   sequentially up to limit; but if nested-key is specified, it cannot 
guarantee the correctness of the aggregation result. This option can be used to
   avoid abnormal input.
 
diff --git a/docs/content/primary-key-table/merge-engine/partial-update.md 
b/docs/content/primary-key-table/merge-engine/partial-update.md
index 5bd689f4a9..31b2adc70b 100644
--- a/docs/content/primary-key-table/merge-engine/partial-update.md
+++ b/docs/content/primary-key-table/merge-engine/partial-update.md
@@ -219,7 +219,7 @@ SELECT *
 FROM AGG;
 -- output 1, 3, 2, 2, "1", 1, 2
 
--- g_1, g_3 are smaller, a should not beupdated
+-- g_1, g_3 are smaller, a should not be updated
 INSERT INTO AGG
 VALUES (1, 3, 3, 2, '3', 3, 1);
 
diff --git a/docs/content/program-api/cpp-api.md 
b/docs/content/program-api/cpp-api.md
index e812bf1644..a122f3b4d7 100644
--- a/docs/content/program-api/cpp-api.md
+++ b/docs/content/program-api/cpp-api.md
@@ -35,7 +35,7 @@ format with maximum efficiency.
 
 [Paimon C++](https://github.com/alibaba/paimon-cpp.git) is currently governed 
under Alibaba open source
 community. You can checkout the 
[document](https://alibaba.github.io/paimon-cpp/getting_started.html)
-for more details about envinroment settings.
+for more details about environment settings.
 
 ```sh
 git clone https://github.com/alibaba/paimon-cpp.git
diff --git a/docs/content/project/contributing.md 
b/docs/content/project/contributing.md
index 3213343178..39a7edabc7 100644
--- a/docs/content/project/contributing.md
+++ b/docs/content/project/contributing.md
@@ -44,7 +44,7 @@ Contributing to Apache Paimon goes beyond writing code for 
the project. Below, w
   <tbody>
     <tr>
       <td><span class="glyphicon glyphicon-exclamation-sign" 
aria-hidden="true"></span> Report Bug</td>
-      <td>To report a problem with Paimon, open <a 
href="https://github.com/apache/paimon/issues";>Paimon’s issues</a>. <br/>
+      <td>To report a problem with Paimon, open <a 
href="https://github.com/apache/paimon/issues";>Paimon's issues</a>. <br/>
       Please give detailed information about the problem you encountered and, 
if possible, add a description that helps to reproduce the problem.</td>
     </tr>
     <tr>
diff --git a/docs/content/spark/dataframe.md b/docs/content/spark/dataframe.md
index b0e507793c..040508f428 100644
--- a/docs/content/spark/dataframe.md
+++ b/docs/content/spark/dataframe.md
@@ -75,7 +75,7 @@ data.write.format("paimon")
 You can achieve REPLACE TABLE semantics by setting the mode to `overwrite` 
with `saveAsTable` or `save`.
 
 It first drops the existing table and then create a new one,
-so you need to specify the table’s properties or partition columns if needed.
+so you need to specify the table's properties or partition columns if needed.
 
 ```scala
 val data: DataFrame = ...
diff --git a/docs/content/spark/procedures.md b/docs/content/spark/procedures.md
index 8770e75e42..089b0ea7fc 100644
--- a/docs/content/spark/procedures.md
+++ b/docs/content/spark/procedures.md
@@ -77,7 +77,7 @@ This section introduce all available spark procedures about 
paimon.
       <td>
          To expire partitions. Argument:
             <li>table: the target table identifier. Cannot be empty.</li>
-            <li>expiration_time: the expiration interval of a partition. A 
partition will be expired if it‘s lifetime is over this value. Partition time 
is extracted from the partition value.</li>
+            <li>expiration_time: the expiration interval of a partition. A 
partition will be expired if it's lifetime is over this value. Partition time 
is extracted from the partition value.</li>
             <li>timestamp_formatter: the formatter to format timestamp from 
string.</li>
             <li>timestamp_pattern: the pattern to get a timestamp from 
partitions.</li>
             <li>expire_strategy: specifies the expiration strategy for 
partition expiration, possible values: 'values-time' or 'update-time' , 
'values-time' as default.</li>
diff --git a/docs/content/spark/quick-start.md 
b/docs/content/spark/quick-start.md
index 79268feab1..58530ebcb7 100644
--- a/docs/content/spark/quick-start.md
+++ b/docs/content/spark/quick-start.md
@@ -107,7 +107,7 @@ Alternatively, you can copy `paimon-spark-3.5_2.12-{{< 
version >}}.jar` under `s
 
 {{< tab "Catalog" >}}
 
-When starting `spark-sql`, use the following command to register Paimon’s 
Spark catalog with the name `paimon`. Table files of the warehouse is stored 
under `/tmp/paimon`.
+When starting `spark-sql`, use the following command to register Paimon's 
Spark catalog with the name `paimon`. Table files of the warehouse is stored 
under `/tmp/paimon`.
 
 ```bash
 spark-sql ... \
@@ -133,7 +133,7 @@ can use the `spark_catalog.${database_name}.${table_name}` 
to access Spark table
 
 {{< tab "Generic Catalog" >}}
 
-When starting `spark-sql`, use the following command to register Paimon’s 
Spark Generic catalog to replace Spark
+When starting `spark-sql`, use the following command to register Paimon's 
Spark Generic catalog to replace Spark
 default catalog `spark_catalog`. (default warehouse is Spark 
`spark.sql.warehouse.dir`)
 
 Currently, it is only recommended to use `SparkGenericCatalog` in the case of 
Hive metastore, Paimon will infer
diff --git a/docs/content/spark/sql-ddl.md b/docs/content/spark/sql-ddl.md
index 1494216e35..83a4bd09f7 100644
--- a/docs/content/spark/sql-ddl.md
+++ b/docs/content/spark/sql-ddl.md
@@ -219,7 +219,7 @@ CREATE TABLE my_table (
 ```
 
 Furthermore, if there is already data stored in the specified location, you 
can create the table without explicitly specifying the fields, partitions and 
props or other information. 
-In this case, the new table will inherit them all from the existing table’s 
metadata. 
+In this case, the new table will inherit them all from the existing table's 
metadata.
 
 However, if you manually specify them, you need to ensure that they are 
consistent with those of the existing table (props can be a subset). Therefore, 
it is strongly recommended not to specify them.
 
diff --git a/docs/layouts/shortcodes/generated/core_configuration.html 
b/docs/layouts/shortcodes/generated/core_configuration.html
index 90419b609b..b592fea8bb 100644
--- a/docs/layouts/shortcodes/generated/core_configuration.html
+++ b/docs/layouts/shortcodes/generated/core_configuration.html
@@ -987,7 +987,7 @@ This config option does not affect the default filesystem 
metastore.</td>
             <td><h5>partition.expiration-time</h5></td>
             <td style="word-wrap: break-word;">(none)</td>
             <td>Duration</td>
-            <td>The expiration interval of a partition. A partition will be 
expired if it‘s lifetime is over this value. Partition time is extracted from 
the partition value.</td>
+            <td>The expiration interval of a partition. A partition will be 
expired if it's lifetime is over this value. Partition time is extracted from 
the partition value.</td>
         </tr>
         <tr>
             <td><h5>partition.idle-time-to-report-statistic</h5></td>
diff --git a/docs/layouts/shortcodes/generated/orc_configuration.html 
b/docs/layouts/shortcodes/generated/orc_configuration.html
index e48edb4dd0..4f1a469151 100644
--- a/docs/layouts/shortcodes/generated/orc_configuration.html
+++ b/docs/layouts/shortcodes/generated/orc_configuration.html
@@ -42,7 +42,7 @@ under the License.
             <td><h5>orc.timestamp-ltz.legacy.type</h5></td>
             <td style="word-wrap: break-word;">true</td>
             <td>Boolean</td>
-            <td>This option is used to be compatible with the paimon-orc‘s old 
behavior for the `timestamp_ltz` data type.</td>
+            <td>This option is used to be compatible with the paimon-orc's old 
behavior for the `timestamp_ltz` data type.</td>
         </tr>
     </tbody>
 </table>
diff --git a/paimon-api/src/main/java/org/apache/paimon/CoreOptions.java 
b/paimon-api/src/main/java/org/apache/paimon/CoreOptions.java
index 12799630cc..aa27621b21 100644
--- a/paimon-api/src/main/java/org/apache/paimon/CoreOptions.java
+++ b/paimon-api/src/main/java/org/apache/paimon/CoreOptions.java
@@ -1103,7 +1103,7 @@ public class CoreOptions implements Serializable {
                     .noDefaultValue()
                     .withDescription(
                             "The expiration interval of a partition. A 
partition will be expired if"
-                                    + " it‘s lifetime is over this value. 
Partition time is extracted from"
+                                    + " it's lifetime is over this value. 
Partition time is extracted from"
                                     + " the partition value.");
 
     public static final ConfigOption<Duration> 
PARTITION_EXPIRATION_CHECK_INTERVAL =
diff --git 
a/paimon-format/src/main/java/org/apache/paimon/format/OrcOptions.java 
b/paimon-format/src/main/java/org/apache/paimon/format/OrcOptions.java
index 3cd95aea01..c3ae041962 100644
--- a/paimon-format/src/main/java/org/apache/paimon/format/OrcOptions.java
+++ b/paimon-format/src/main/java/org/apache/paimon/format/OrcOptions.java
@@ -47,5 +47,5 @@ public class OrcOptions {
                     .booleanType()
                     .defaultValue(true)
                     .withDescription(
-                            "This option is used to be compatible with the 
paimon-orc‘s old behavior for the `timestamp_ltz` data type.");
+                            "This option is used to be compatible with the 
paimon-orc's old behavior for the `timestamp_ltz` data type.");
 }
diff --git 
a/tools/ci/paimon-ci-tools/src/main/java/org/apache/paimon/tools/ci/licensecheck/JarFileChecker.java
 
b/tools/ci/paimon-ci-tools/src/main/java/org/apache/paimon/tools/ci/licensecheck/JarFileChecker.java
index b8294353f6..386f2407b3 100644
--- 
a/tools/ci/paimon-ci-tools/src/main/java/org/apache/paimon/tools/ci/licensecheck/JarFileChecker.java
+++ 
b/tools/ci/paimon-ci-tools/src/main/java/org/apache/paimon/tools/ci/licensecheck/JarFileChecker.java
@@ -173,7 +173,7 @@ public class JarFileChecker {
                         // JSON License
                         "The Software shall be used for Good, not Evil.",
                         // can sometimes be found in "funny" licenses
-                        "Don’t be evil"));
+                        "Don't be evil"));
     }
 
     private static Collection<Pattern> asPatterns(String... texts) {

Reply via email to