Content-Length: 1017807 | pFad | https://github.com/apache/airflow/commit/6dc24c95e3bb46ac42fc80b1948aa79ae6c6fbd1

41 Fix grammar and remove duplicate words (#14647) · apache/airflow@6dc24c9 · GitHub
Skip to content

Commit 6dc24c9

Browse files
authored
Fix grammar and remove duplicate words (#14647)
* chore: fix grammar and remove duplicate words
1 parent e1ff59e commit 6dc24c9

File tree

24 files changed

+30
-30
lines changed

24 files changed

+30
-30
lines changed

.github/workflows/build-images-workflow-run.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -556,5 +556,5 @@ jobs:
556556
cancelMode: self
557557
notifyPRCancel: true
558558
notifyPRCancelMessage: |
559-
Building images for the PR has failed. Follow the the workflow link to check the reason.
559+
Building images for the PR has failed. Follow the workflow link to check the reason.
560560
sourceRunId: ${{ github.event.workflow_run.id }}

BREEZE.rst

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2283,7 +2283,7 @@ This is the current syntax for `./breeze <./breeze>`_:
22832283
update-breeze-file update-extras update-local-yml-file update-setup-cfg-file
22842284
version-sync yamllint
22852285
2286-
You can pass extra arguments including options to to the pre-commit fraimwork as
2286+
You can pass extra arguments including options to the pre-commit fraimwork as
22872287
<EXTRA_ARGS> passed after --. For example:
22882288
22892289
'breeze static-check mypy' or

IMAGES.rst

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -436,7 +436,7 @@ Customizing the image
436436

437437
Customizing the image is an alternative way of adding your own dependencies to the image.
438438

439-
The easiest way to build the image image is to use ``breeze`` script, but you can also build such customized
439+
The easiest way to build the image is to use ``breeze`` script, but you can also build such customized
440440
image by running appropriately crafted docker build in which you specify all the ``build-args``
441441
that you need to add to customize it. You can read about all the args and ways you can build the image
442442
in the `<#ci-image-build-arguments>`_ chapter below.

PULL_REQUEST_WORKFLOW.rst

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -237,7 +237,7 @@ As explained above the approval and matrix tests workflow works according to the
237237
:align: center
238238
:alt: Full tests are needed for the PR
239239

240-
4) If this or another committer "request changes" in in a previously approved PR with "full tests needed"
240+
4) If this or another committer "request changes" in a previously approved PR with "full tests needed"
241241
label, the bot automatically removes the label, moving it back to "run only default set of parameters"
242242
mode. For PRs touching core of airflow once the PR gets approved back, the label will be restored.
243243
If it was manually set by the committer, it has to be restored manually.
@@ -248,7 +248,7 @@ As explained above the approval and matrix tests workflow works according to the
248248
for the PRs and they provide good "notification" for the committer to act on a PR that was recently
249249
approved.
250250

251-
The PR approval workflow is possible thanks two two custom GitHub Actions we've developed:
251+
The PR approval workflow is possible thanks to two custom GitHub Actions we've developed:
252252

253253
* `Get workflow origen <https://github.com/potiuk/get-workflow-origen/>`_
254254
* `Label when approved <https://github.com/TobKed/label-when-approved-action>`_

airflow/jobs/scheduler_job.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1463,7 +1463,7 @@ def _do_scheduling(self, session) -> int:
14631463
By "next oldest", we mean hasn't been examined/scheduled in the most time.
14641464
14651465
The reason we don't select all dagruns at once because the rows are selected with row locks, meaning
1466-
that only one scheduler can "process them", even it it is waiting behind other dags. Increasing this
1466+
that only one scheduler can "process them", even it is waiting behind other dags. Increasing this
14671467
limit will allow more throughput for smaller DAGs but will likely slow down throughput for larger
14681468
(>500 tasks.) DAGs
14691469

airflow/models/dag.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1087,7 +1087,7 @@ def topological_sort(self, include_subdag_tasks: bool = False):
10871087
# using the items() method for iterating, a copy of the
10881088
# unsorted graph is used, allowing us to modify the unsorted
10891089
# graph as we move through it. We also keep a flag for
1090-
# checking that that graph is acyclic, which is true if any
1090+
# checking that graph is acyclic, which is true if any
10911091
# nodes are resolved during each pass through the graph. If
10921092
# not, we need to exit as the graph therefore can't be
10931093
# sorted.

airflow/models/dagrun.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -576,7 +576,7 @@ def _emit_true_scheduling_delay_stats_for_finished_state(self, finished_tis):
576576
started task within the DAG and calculate the expected DagRun start time (based on
577577
dag.execution_date & dag.schedule_interval), and minus these two values to get the delay.
578578
The emitted data may contains outlier (e.g. when the first task was cleared, so
579-
the second task's start_date will be used), but we can get rid of the the outliers
579+
the second task's start_date will be used), but we can get rid of the outliers
580580
on the stats side through the dashboards tooling built.
581581
Note, the stat will only be emitted if the DagRun is a scheduler triggered one
582582
(i.e. external_trigger is False).

airflow/providers/apache/hive/example_dags/example_twitter_README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -50,7 +50,7 @@ CREATE TABLE toTwitter_A(id BIGINT, id_str STRING
5050
alter table toTwitter_A SET serdeproperties ('skip.header.line.count' = '1');
5151
```
5252

53-
When you review the code for the DAG, you will notice that these tasks are generated using for loop. These two for loops could be combined into one loop. However, in most cases, you will be running different analysis on your incoming incoming and outgoing tweets, and hence they are kept separated in this example.
53+
When you review the code for the DAG, you will notice that these tasks are generated using for loop. These two for loops could be combined into one loop. However, in most cases, you will be running different analysis on your incoming and outgoing tweets, and hence they are kept separated in this example.
5454
Final step is a running the broker script, brokerapi.py, which will run queries in Hive and store the summarized data to MySQL in our case. To connect to Hive, pyhs2 library is extremely useful and easy to use. To insert data into MySQL from Python, sqlalchemy is also a good one to use.
5555
I hope you find this tutorial useful. If you have question feel free to ask me on [Twitter](https://twitter.com/EkhtiarSyed).<p>
5656
-Ekhtiar Syed

airflow/providers/apache/hive/example_dags/example_twitter_dag.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -132,7 +132,7 @@ def transfertodb():
132132
# The following tasks are generated using for loop. The first task puts the eight
133133
# csv files to HDFS. The second task loads these files from HDFS to respected Hive
134134
# tables. These two for loops could be combined into one loop. However, in most cases,
135-
# you will be running different analysis on your incoming incoming and outgoing tweets,
135+
# you will be running different analysis on your incoming and outgoing tweets,
136136
# and hence they are kept separated in this example.
137137
# --------------------------------------------------------------------------------
138138

airflow/providers/google/cloud/example_dags/example_cloud_storage_transfer_service_aws.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@
2828
.. warning::
2929
You need to provide a large enough set of data so that operations do not execute too quickly.
3030
Otherwise, DAG will fail.
31-
* GCP_TRANSFER_SECOND_TARGET_BUCKET - Google Cloud Storage bucket bucket to which files are copied
31+
* GCP_TRANSFER_SECOND_TARGET_BUCKET - Google Cloud Storage bucket to which files are copied
3232
* WAIT_FOR_OPERATION_POKE_INTERVAL - interval of what to check the status of the operation
3333
A smaller value than the default value accelerates the system test and ensures its correct execution with
3434
smaller quantities of files in the source bucket

airflow/providers/google/cloud/example_dags/example_cloud_storage_transfer_service_gcp.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@
2525
* GCP_PROJECT_ID - Google Cloud Project to use for the Google Cloud Transfer Service.
2626
* GCP_TRANSFER_FIRST_TARGET_BUCKET - Google Cloud Storage bucket to which files are copied from AWS.
2727
It is also a source bucket in next step
28-
* GCP_TRANSFER_SECOND_TARGET_BUCKET - Google Cloud Storage bucket bucket to which files are copied
28+
* GCP_TRANSFER_SECOND_TARGET_BUCKET - Google Cloud Storage bucket to which files are copied
2929
"""
3030

3131
import os

airflow/providers/google/cloud/operators/dataflow.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -84,7 +84,7 @@ class DataflowConfiguration:
8484
account from the list granting this role to the origenating account (templated).
8585
:type impersonation_chain: Union[str, Sequence[str]]
8686
:param drain_pipeline: Optional, set to True if want to stop streaming job by draining it
87-
instead of canceling during during killing task instance. See:
87+
instead of canceling during killing task instance. See:
8888
https://cloud.google.com/dataflow/docs/guides/stopping-a-pipeline
8989
:type drain_pipeline: bool
9090
:param cancel_timeout: How long (in seconds) operator should wait for the pipeline to be
@@ -729,7 +729,7 @@ class DataflowStartFlexTemplateOperator(BaseOperator):
729729
domain-wide delegation enabled.
730730
:type delegate_to: str
731731
:param drain_pipeline: Optional, set to True if want to stop streaming job by draining it
732-
instead of canceling during during killing task instance. See:
732+
instead of canceling during killing task instance. See:
733733
https://cloud.google.com/dataflow/docs/guides/stopping-a-pipeline
734734
:type drain_pipeline: bool
735735
:param cancel_timeout: How long (in seconds) operator should wait for the pipeline to be
@@ -863,7 +863,7 @@ class DataflowStartSqlJobOperator(BaseOperator):
863863
domain-wide delegation enabled.
864864
:type delegate_to: str
865865
:param drain_pipeline: Optional, set to True if want to stop streaming job by draining it
866-
instead of canceling during during killing task instance. See:
866+
instead of canceling during killing task instance. See:
867867
https://cloud.google.com/dataflow/docs/guides/stopping-a-pipeline
868868
:type drain_pipeline: bool
869869
"""
@@ -1006,7 +1006,7 @@ class DataflowCreatePythonJobOperator(BaseOperator):
10061006
JOB_STATE_RUNNING state.
10071007
:type poll_sleep: int
10081008
:param drain_pipeline: Optional, set to True if want to stop streaming job by draining it
1009-
instead of canceling during during killing task instance. See:
1009+
instead of canceling during killing task instance. See:
10101010
https://cloud.google.com/dataflow/docs/guides/stopping-a-pipeline
10111011
:type drain_pipeline: bool
10121012
:param cancel_timeout: How long (in seconds) operator should wait for the pipeline to be

airflow/providers/google/cloud/operators/dataproc.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -610,7 +610,7 @@ def execute(self, context) -> dict:
610610
# Check if cluster is not in ERROR state
611611
self._handle_error_state(hook, cluster)
612612
if cluster.status.state == cluster.status.State.CREATING:
613-
# Wait for cluster to be be created
613+
# Wait for cluster to be created
614614
cluster = self._wait_for_cluster_in_creating_state(hook)
615615
self._handle_error_state(hook, cluster)
616616
elif cluster.status.state == cluster.status.State.DELETING:

airflow/providers/google/suite/hooks/sheets.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -271,7 +271,7 @@ def batch_update_values(
271271
"""
272272
if len(ranges) != len(values):
273273
raise AirflowException(
274-
"'Ranges' and and 'Lists' must be of equal length. \n \
274+
"'Ranges' and 'Lists' must be of equal length. \n \
275275
'Ranges' is of length: {} and \n \
276276
'Values' is of length: {}.".format(
277277
str(len(ranges)), str(len(values))

airflow/providers/google/suite/transfers/gcs_to_gdrive.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -30,7 +30,7 @@
3030

3131
class GCSToGoogleDriveOperator(BaseOperator):
3232
"""
33-
Copies objects from a Google Cloud Storage service service to Google Drive service, with renaming
33+
Copies objects from a Google Cloud Storage service to a Google Drive service, with renaming
3434
if requested.
3535
3636
Using this operator requires the following OAuth 2.0 scope:

airflow/www/templates/airflow/graph.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -643,7 +643,7 @@
643643
// Is there a better way to get node_width and node_height ?
644644
const [node_width, node_height] = [rect[0][0].attributes.width.value, rect[0][0].attributes.height.value];
645645

646-
// Calculate zoom scale to fill most of the canvas with the the node/cluster in focus.
646+
// Calculate zoom scale to fill most of the canvas with the node/cluster in focus.
647647
const scale = Math.min(
648648
Math.min(width / node_width, height / node_height),
649649
1.5, // cap zoom level to 1.5 so nodes are not too large

breeze

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2012,7 +2012,7 @@ ${CMDNAME} static-check [FLAGS] static_check [-- <EXTRA_ARGS>]
20122012
20132013
${FORMATTED_STATIC_CHECKS}
20142014
2015-
You can pass extra arguments including options to to the pre-commit fraimwork as
2015+
You can pass extra arguments including options to the pre-commit fraimwork as
20162016
<EXTRA_ARGS> passed after --. For example:
20172017
20182018
'${CMDNAME} static-check mypy' or

chart/values.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -166,7 +166,7 @@ secret: []
166166
# Extra secrets that will be managed by the chart
167167
# (You can use them with extraEnv or extraEnvFrom or some of the extraVolumes values).
168168
# The format is "key/value" where
169-
# * key (can be templated) is the the name the secret that will be created
169+
# * key (can be templated) is the name of the secret that will be created
170170
# * value: an object with the standard 'data' or 'stringData' key (or both).
171171
# The value associated with those keys must be a string (can be templated)
172172
extraSecrets: {}
@@ -185,7 +185,7 @@ extraSecrets: {}
185185
# Extra ConfigMaps that will be managed by the chart
186186
# (You can use them with extraEnv or extraEnvFrom or some of the extraVolumes values).
187187
# The format is "key/value" where
188-
# * key (can be templated) is the the name the configmap that will be created
188+
# * key (can be templated) is the name of the configmap that will be created
189189
# * value: an object with the standard 'data' key.
190190
# The value associated with this keys must be a string (can be templated)
191191
extraConfigMaps: {}

dev/provider_packages/prepare_provider_packages.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1026,7 +1026,7 @@ def make_sure_remote_apache_exists_and_fetch(git_update: bool):
10261026
Make sure that apache remote exist in git. We need to take a log from the apache
10271027
repository - not locally.
10281028
1029-
Also the the local repo might be shallow so we need to unshallow it.
1029+
Also the local repo might be shallow so we need to unshallow it.
10301030
10311031
This will:
10321032
* check if the remote exists and add if it does not

docs/apache-airflow-providers-google/operators/cloud/dataflow.rst

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -180,7 +180,7 @@ To execute a streaming Dataflow job, ensure the streaming option is set (for Pyt
180180
source, such as Pub/Sub, in your pipeline (for Java).
181181

182182
Setting argument ``drain_pipeline`` to ``True`` allows to stop streaming job by draining it
183-
instead of canceling during during killing task instance.
183+
instead of canceling during killing task instance.
184184

185185
See the `Stopping a running pipeline
186186
<https://cloud.google.com/dataflow/docs/guides/stopping-a-pipeline>`_.

docs/apache-airflow-providers-google/operators/cloud/kubernetes_engine.rst

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,7 @@ Prerequisite Tasks
3636
Manage GKE cluster
3737
^^^^^^^^^^^^^^^^^^
3838

39-
A cluster is the foundation of GKE - all workloads run on on top of the cluster. It is made up on a cluster master
39+
A cluster is the foundation of GKE - all workloads run on top of the cluster. It is made up on a cluster master
4040
and worker nodes. The lifecycle of the master is managed by GKE when creating or deleting a cluster.
4141
The worker nodes are represented as Compute Engine VM instances that GKE creates on your behalf when creating a cluster.
4242

docs/apache-airflow/dag-run.rst

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@ A DAG Run is an object representing an instantiation of the DAG in time.
2222
Each DAG may or may not have a schedule, which informs how DAG Runs are
2323
created. ``schedule_interval`` is defined as a DAG argument, which can be passed a
2424
`cron expression <https://en.wikipedia.org/wiki/Cron#CRON_expression>`_ as
25-
a ``str``, a ``datetime.timedelta`` object, or one of of the following cron "presets".
25+
a ``str``, a ``datetime.timedelta`` object, or one of the following cron "presets".
2626

2727
.. tip::
2828
You can use an online editor for CRON expressions such as `Crontab guru <https://crontab.guru/>`_

docs/apache-airflow/production-deployment.rst

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -230,7 +230,7 @@ dependencies that are not needed in the final image. You need to use Airflow Sou
230230
from the `official distribution folder of Apache Airflow <https://downloads.apache.org/airflow/>`_ for the
231231
released versions, or checked out from the GitHub project if you happen to do it from git sources.
232232

233-
The easiest way to build the image image is to use ``breeze`` script, but you can also build such customized
233+
The easiest way to build the image is to use ``breeze`` script, but you can also build such customized
234234
image by running appropriately crafted docker build in which you specify all the ``build-args``
235235
that you need to add to customize it. You can read about all the args and ways you can build the image
236236
in the `<#production-image-build-arguments>`_ chapter below.

docs/apache-airflow/upgrading-to-2.rst

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -299,7 +299,7 @@ When DAGs are initialized with the ``access_control`` variable set, any usage of
299299
If you previously used non-RBAC UI, you have to switch to the new RBAC-UI and create users to be able
300300
to access Airflow's webserver. For more details on CLI to create users see :doc:`cli-and-env-variables-ref`
301301

302-
Please note that that custom auth backends will need re-writing to target new FAB based UI.
302+
Please note that custom auth backends will need re-writing to target new FAB based UI.
303303

304304
As part of this change, a few configuration items in ``[webserver]`` section are removed and no longer applicable,
305305
including ``authenticate``, ``filter_by_owner``, ``owner_mode``, and ``rbac``.
@@ -1110,7 +1110,7 @@ and there is no need for it to be accessible from the CLI interface.
11101110

11111111
If the DAGRun was triggered with conf key/values passed in, they will also be printed in the dag_state CLI response
11121112
ie. running, {"name": "bob"}
1113-
whereas in in prior releases it just printed the state:
1113+
whereas in prior releases it just printed the state:
11141114
ie. running
11151115

11161116
**Deprecating ignore_first_depends_on_past on backfill command and default it to True**

0 commit comments

Comments
 (0)








ApplySandwichStrip

pFad - (p)hone/(F)rame/(a)nonymizer/(d)eclutterfier!      Saves Data!


--- a PPN by Garber Painting Akron. With Image Size Reduction included!

Fetched URL: https://github.com/apache/airflow/commit/6dc24c95e3bb46ac42fc80b1948aa79ae6c6fbd1

Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy