Skip to content

Add typing to amazon provider EMR#9969

Closed
Swalloow wants to merge 133 commits intoapache:masterfrom
Swalloow:typing-aws-emr
Closed

Add typing to amazon provider EMR#9969
Swalloow wants to merge 133 commits intoapache:masterfrom
Swalloow:typing-aws-emr

Conversation

@Swalloow
Copy link
Contributor

related: #9708
Adding type annotations to amazon/emr.


^ Add meaningful description above

Read the Pull Request Guidelines for more information.
In case of fundamental code change, Airflow Improvement Proposal (AIP) is needed.
In case of a new dependency, check compliance with the ASF 3rd Party License Policy.
In case of backwards incompatible changes please leave a note in UPDATING.md.

@boring-cyborg boring-cyborg bot added the provider:amazon AWS/Amazon - related issues label Jul 24, 2020
Swalloow and others added 16 commits July 25, 2020 06:59
mysqlclient was pinned as the functionality of Connection object of MySQL had changed and that it dropped python2.

mysqlclient has added context manager interface to Connection which closes the connection on `__exit__` in 2.0.0 and Airflow Master no longer uses Python 2
`startd_ate_lte` -> `start_date_lte`
apache#9993)

Sharing session with RTIF leads to idle-in-transaction timeout error when DAG serialization is enabled and task running duration exceeds the idle-in-transaction timeout setting of the database.
Unfortunately cancelling workflows does not work from the workflows
executed by forks because their tokens do not allow doing that
(they are read only). So we have to run a separate cron-triggered
action (in the context of apache/airflow repository to cancel
all duplicate workflows. This action stops all the workflows
running from the same fork/branch except the last one.
j-y-matsubara and others added 8 commits August 6, 2020 13:08
…gger_rule (apache#8867)

Fixes the BaseSensorOperator to make respect the trigger_rule in downstream tasks, when setting soft_fail="True".
Co-authored-by: Jarek Potiuk <jarek@potiuk.com>

Co-authored-by: Jarek Potiuk <jarek@potiuk.com>
Signed-off-by: Raymond Etornam <retornam@users.noreply.github.com>
* add correct signature to operators in providers package

* add keyword only to operators and sensors outside provider package

* remove unused type ignore
@Swalloow
Copy link
Contributor Author

Swalloow commented Aug 7, 2020

@mik-laj Any updates on reviews?

potiuk and others added 7 commits August 7, 2020 10:29
…he#10209)

We are using newly added feature of GitHub to add manually triggered
workflow to enable manually-triggered force-syncing of your fork
with apache/airflow.
* Pylint checks should be way faster now

Instead of running separate pylint checks for tests and main source
we are running a single check now. This is possible thanks to a
nice hack - we have pylint plugin that injects the right
"# pylint: disable=" comment for all test files while reading
the file content by astroid (just before tokenization)

Thanks to that we can also separate out pylint checks
to a separate job in CI - this way all pylint checks will
be run in parallel to all other checks effectively halfing
the time needed to get the static check feedback and potentially
cancelling other jobs much faster.

* fixup! Pylint checks should be way faster now
Replace fixed namespace "airflow" with variable {{ .Release.Namespace }}
)

* Improves stability of reported coverage and makes it nicer

With this change we only upload coverage report in the case
when all tests were successuful and actually executed. This means
that coverage report will not be run when the job gets canceled.

Currently a lof of coverage reports gathered contain far less
coverage because when static check fails or docs some test jobs
could already submit their coverage - resulting in partial coverage
reports.

With this change we also remove comment from coverage report
and replace it with (for now) informational status message published
to github. If we see that it works, we can change it to a
PR-failing status if coverage drops for a given PR.

This way we might get our coverage monotonously increasing :).
While doing a trigger_dag from UI, DagRun gets created first and then WebServer starts creating TIs. Meanwhile, Scheduler also picks up the DagRun and starts creating the TIs, which results in IntegrityError as the Primary key constraint gets violated. This happens when a DAG has a good number of tasks.

Also, changing the TIs array with a set for faster lookups for Dags with too many tasks.
@mik-laj
Copy link
Member

mik-laj commented Aug 7, 2020

LGTM. Can you do a rebase?

@Swalloow
Copy link
Contributor Author

Swalloow commented Aug 8, 2020

@mik-laj During the rebase process, all previous history has been pushed to this PR.
Would it be okay or How can I fix it?

@potiuk
Copy link
Member

potiuk commented Aug 8, 2020

You need to just get your commit from the top and rebase it on top of master. See the CONTRIBUTING.rst - there is a section about it in the docs and some useful links explaining what to do.

@Swalloow
Copy link
Contributor Author

Replace to #10910

@Swalloow Swalloow closed this Sep 13, 2020
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

provider:amazon AWS/Amazon - related issues

Projects

None yet

Development

Successfully merging this pull request may close these issues.