This is an automated email from the ASF dual-hosted git repository.

turaga pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/airflow.git


The following commit(s) were added to refs/heads/main by this push:
     new e7898b8100f Fix duplicate words in airflow-core documentation (#62384)
e7898b8100f is described below

commit e7898b8100f92f1988edc35188339e37922813da
Author: Keisuke Kawamura <[email protected]>
AuthorDate: Tue Feb 24 11:45:30 2026 +0900

    Fix duplicate words in airflow-core documentation (#62384)
    
    Remove repeated words ("is is", "the the", "this this", "set set",
    "how how") and an extra article ("the a") across 8 documentation files.
    
    Co-authored-by: keisuke-kawamura <[email protected]>
    Co-authored-by: Claude Opus 4.6 <[email protected]>
---
 airflow-core/docs/administration-and-deployment/dagfile-processing.rst | 2 +-
 airflow-core/docs/authoring-and-scheduling/assets.rst                  | 2 +-
 airflow-core/docs/authoring-and-scheduling/deferring.rst               | 2 +-
 airflow-core/docs/authoring-and-scheduling/timetable.rst               | 2 +-
 airflow-core/docs/core-concepts/executor/index.rst                     | 2 +-
 airflow-core/docs/core-concepts/objectstorage.rst                      | 2 +-
 airflow-core/docs/howto/connection.rst                                 | 2 +-
 airflow-core/docs/installation/upgrading.rst                           | 2 +-
 8 files changed, 8 insertions(+), 8 deletions(-)

diff --git 
a/airflow-core/docs/administration-and-deployment/dagfile-processing.rst 
b/airflow-core/docs/administration-and-deployment/dagfile-processing.rst
index d2050e1df6a..b83c72171a2 100644
--- a/airflow-core/docs/administration-and-deployment/dagfile-processing.rst
+++ b/airflow-core/docs/administration-and-deployment/dagfile-processing.rst
@@ -120,7 +120,7 @@ There are several areas of resource usage that you should 
pay attention to:
 * Another solution to FileSystem performance, if it becomes your bottleneck, 
is to turn to alternative
   mechanisms of distributing your Dags. Embedding Dags in your image and 
GitSync distribution have both
   the property that the files are available locally for the Dag processor and 
it does not have to use a
-  distributed filesystem to read the files, the files are available locally 
for the the Dag processor and it is
+  distributed filesystem to read the files, the files are available locally 
for the Dag processor and it is
   usually as fast as it can be, especially if your machines use fast SSD disks 
for local storage. Those
   distribution mechanisms have other characteristics that might make them not 
the best choice for you,
   but if your problems with performance come from distributed filesystem 
performance, they might be the
diff --git a/airflow-core/docs/authoring-and-scheduling/assets.rst 
b/airflow-core/docs/authoring-and-scheduling/assets.rst
index de5913a59e8..b5b97b9979c 100644
--- a/airflow-core/docs/authoring-and-scheduling/assets.rst
+++ b/airflow-core/docs/authoring-and-scheduling/assets.rst
@@ -63,7 +63,7 @@ For pre-defined schemes (e.g., ``file``, ``postgres``, and 
``s3``), you must pro
     # invalid asset:
     must_contain_bucket_name = Asset("s3://")
 
-Do not use the ``airflow`` scheme, which is is reserved for Airflow's 
internals.
+Do not use the ``airflow`` scheme, which is reserved for Airflow's internals.
 
 Airflow always prefers using lower cases in schemes, and case sensitivity is 
needed in the host part of the URI to correctly distinguish between resources.
 
diff --git a/airflow-core/docs/authoring-and-scheduling/deferring.rst 
b/airflow-core/docs/authoring-and-scheduling/deferring.rst
index 1cfa2384144..9344f07aa60 100644
--- a/airflow-core/docs/authoring-and-scheduling/deferring.rst
+++ b/airflow-core/docs/authoring-and-scheduling/deferring.rst
@@ -506,7 +506,7 @@ Additionally, queue assignment is currently only compatible 
with the subset of t
 
+------------------------------------------------------------------------------------------------------+----------------------+----------------------------------------------------------------------------------+
 |          Trigger Type                                                        
                        |   Supports queues?   |  Triggerer assignment when 
:ref:`config:triggerer__queues_enabled` is ``True``   |
 
+======================================================================================================+======================+==================================================================================+
-| Task-created Trigger instances                                               
                        |   Yes                |  Any triggerer with the the 
task queue present in its ``--queues`` option        |
+| Task-created Trigger instances                                               
                        |   Yes                |  Any triggerer with the task 
queue present in its ``--queues`` option            |
 
+------------------------------------------------------------------------------------------------------+----------------------+----------------------------------------------------------------------------------+
 | :doc:`Event-Driven Triggers<../authoring-and-scheduling/event-scheduling>`   
                        |   No                 |  Any triggerer running without 
the ``--queues`` option                           |
 
+------------------------------------------------------------------------------------------------------+----------------------+----------------------------------------------------------------------------------+
diff --git a/airflow-core/docs/authoring-and-scheduling/timetable.rst 
b/airflow-core/docs/authoring-and-scheduling/timetable.rst
index e92fc66130e..28957370a47 100644
--- a/airflow-core/docs/authoring-and-scheduling/timetable.rst
+++ b/airflow-core/docs/authoring-and-scheduling/timetable.rst
@@ -325,7 +325,7 @@ By default, ``catchup`` is set to ``False``. This prevents 
running unnecessary D
 - If you create a new Dag with a start date in the past, and don't want to run 
Dags for the past. If ``catchup`` is ``True``, Airflow runs all Dags that would 
have run in that time interval.
 - If you pause an existing Dag, and then restart it at a later date, 
``catchup`` being ``False`` means that Airflow does not run the Dags that would 
have run during the paused period.
 
-In these scenarios, the ``logical_date`` in the ``run_id`` are based on how 
how the timetable handles the data
+In these scenarios, the ``logical_date`` in the ``run_id`` are based on how 
the timetable handles the data
 interval.
 
 You can change the default ``catchup`` behavior using the Airflow config 
``[scheduler] catchup_by_default``.
diff --git a/airflow-core/docs/core-concepts/executor/index.rst 
b/airflow-core/docs/core-concepts/executor/index.rst
index e3178bb48e2..3a0d593a49b 100644
--- a/airflow-core/docs/core-concepts/executor/index.rst
+++ b/airflow-core/docs/core-concepts/executor/index.rst
@@ -354,7 +354,7 @@ Logging
 ^^^^^^^
 
 Executors may vend log messages which will be included in the Airflow task 
logs by implementing the ``get_task_logs`` method. This can be helpful if the 
execution environment has extra context in the case of task failures, which may 
be due to the execution environment itself rather than the Airflow task code. 
It can also be helpful to include setup/teardown logging from the execution 
environment.
-The ``KubernetesExecutor`` leverages this this capability to include logs from 
the pod which ran a specific Airflow task and display them in the logs for that 
Airflow task. A pseudo-code example of implementing task log vending from an 
executor can be seen below:
+The ``KubernetesExecutor`` leverages this capability to include logs from the 
pod which ran a specific Airflow task and display them in the logs for that 
Airflow task. A pseudo-code example of implementing task log vending from an 
executor can be seen below:
 
 .. code-block:: python
 
diff --git a/airflow-core/docs/core-concepts/objectstorage.rst 
b/airflow-core/docs/core-concepts/objectstorage.rst
index 4478980aeda..aabdd019060 100644
--- a/airflow-core/docs/core-concepts/objectstorage.rst
+++ b/airflow-core/docs/core-concepts/objectstorage.rst
@@ -314,7 +314,7 @@ External Integrations
 ---------------------
 
 Many other projects, like DuckDB, Apache Iceberg etc, can make use of the 
object storage abstraction. Often this is
-done by passing the underlying ``fsspec`` implementation. For this this 
purpose ``ObjectStoragePath`` exposes
+done by passing the underlying ``fsspec`` implementation. For this purpose 
``ObjectStoragePath`` exposes
 the ``fs`` property. For example, the following works with ``duckdb`` so that 
the connection details from Airflow
 are used to connect to s3 and a parquet file, indicated by a 
``ObjectStoragePath``, is read:
 
diff --git a/airflow-core/docs/howto/connection.rst 
b/airflow-core/docs/howto/connection.rst
index e58d0260db4..0b9c352c0c4 100644
--- a/airflow-core/docs/howto/connection.rst
+++ b/airflow-core/docs/howto/connection.rst
@@ -113,7 +113,7 @@ If serializing with Airflow URI:
 
     export 
AIRFLOW_CONN_MY_PROD_DATABASE='my-conn-type://login:password@host:port/schema?param1=val1&param2=val2'
 
-See :ref:`Connection URI format <connection-uri-format>` for more details on 
how to generate the a valid URI.
+See :ref:`Connection URI format <connection-uri-format>` for more details on 
how to generate a valid URI.
 
 Visibility in UI and CLI
 ^^^^^^^^^^^^^^^^^^^^^^^^^
diff --git a/airflow-core/docs/installation/upgrading.rst 
b/airflow-core/docs/installation/upgrading.rst
index f2be39dd427..990d97ed1fe 100644
--- a/airflow-core/docs/installation/upgrading.rst
+++ b/airflow-core/docs/installation/upgrading.rst
@@ -132,7 +132,7 @@ Get familiar with the internal Database structure of 
Airflow which you might fin
 
 Make sure to copy the output. You will need it in the last step. Your
 ``dag_id``, ``run_id``, ``task_id`` and ``key`` columns should have ``utf8`` 
or ``utf8mb3`` character
-set set explicitly, similar to:
+set explicitly, similar to:
 
 .. code-block:: text
 

Reply via email to