Skip to content

Conversation

@JCoder01
Copy link
Contributor

@JCoder01 JCoder01 commented May 22, 2019

Make sure you have checked all steps below.

Jira

  • My PR addresses the following Airflow Jira issues and references them in the PR title. For example, "[AIRFLOW-XXX] My Airflow PR"
    • https://issues.apache.org/jira/browse/AIRFLOW-4479
    • In case you are fixing a typo in the documentation you can prepend your commit with [AIRFLOW-XXX], code changes always need a Jira issue.
    • In case you are proposing a fundamental code change, you need to create an Airflow Improvement Proposal (AIP).
    • In case you are adding a dependency, check if the license complies with the ASF 3rd Party License Policy.

Description

  • Here are some details about my PR, including screenshots of any UI changes:
    Fixes ImapAttachementToS3Operator s3_overwrite. Argument was present but not called.

Tests

  • My PR adds the following unit tests OR does not need testing for this extremely good reason:
    adjusts existing tests

Commits

  • My commits all reference Jira issues in their subject lines, and I have squashed multiple commits if they address the same issue. In addition, my commits follow the guidelines from "How to write a good git commit message":
    1. Subject is separated from body by a blank line
    2. Subject is limited to 50 characters (not including Jira issue reference)
    3. Subject does not end with a period
    4. Subject uses the imperative mood ("add", not "adding")
    5. Body wraps at 72 characters
    6. Body explains "what" and "why", not "how"

Documentation

  • In case of new functionality, my PR adds documentation that describes how to use it.
    • All the public functions and the classes in the PR contain docstrings that explain what it does
    • If you implement backwards incompatible changes, please leave a note in the Updating.md so we can assign it to a appropriate release

Code Quality

  • Passes flake8

aam-jacoder and others added 30 commits May 3, 2019 13:43
Add assert to test_sync_to_db to confirm subdag.fileloc == dag.fileloc
Add new test test_dag_details_subdag to check dag_details endpoint for subdags.
Add replace=self.s3_overwrite to s3_hook.load_bytes
…_wired_up

Revert "Imap to s3 overwrite not wired up"
…5233)

Our new environment already sets this env var but we aren't using this
yet
Remove _future_ should done in
apache#5020, but in
the same time our code base refactor models and
move some code out of _init.py. This PR to remove
__future__ in models.
…pache#5242)

The issue raised in https://issues.apache.org/jira/browse/AIRFLOW-2522 was resolved.
Scope is not ignored when default credentials are used. This note should be deleted.

Scope is read in:
https://github.com/apache/airflow/blob/master/airflow/contrib/hooks/gcp_api_base_hook.py#L88-L92
When default credentails is used, then this code is used:
https://github.com/apache/airflow/blob/master/airflow/contrib/hooks/gcp_api_base_hook.py#L94-L97
so scope is passed to the external library.
…#5147)

models.baseoperator get relatives by
attribute, but _downstream_task_ids
and _upstream_task_ids is private member
This patch will get relatives from
exists function get_direct_relative_ids
apache#5248)

* [AIRFLOW-4467] Add dataproc_jars to templated fields in Dataproc operators



apache#5192 edit docs about:
dataproc_pig_jars  in DataProcPigOperator
dataproc_hive_jars in DataProcHiveOperator
dataproc_spark_jars in DataProcSparkSqlOperator
dataproc_hadoop_jars in DataProcHadoopOperator
The edit mentioned that these fields are tempated but they are not.
Raising PR to make them templated as doc suggest.

* Fix flake8
…pache#4050)

If we are running tasks via sudo then AIRFLOW__ config env vars won't be
visible anymore (without them showing up in `ps`) and we likely might
not have permission to run the _cmd's specified to find the passwords.

But if we are running as the same user then there is no need to "bake"
those options in to the temporary config file -- if the operator decided
they didn't want those values appearing in a config file on disk, then
lets do our best to respect that.
…he#5253)

jinja2 cannot use dict/lists as templates hence converting
it to json solves this while keeping complexity down.
* [AIRFLOW-4468] add sql_alchemy_max_overflow parameter
… on task pods (apache#4551)

* [AIRFLOW-2955] Fix kubernetes pod operator to set requests and limits on task pods.

* [AIRFLOW-2955] Remove bare except to follow flake8.

* [AIRFLOW-2955] Remove unused library.

* [AIRFLOW-2955] Fix kubernetes pod operator to set requests and limits on task pods.

* [AIRFLOW-2955] Remove bare except to follow flake8.

* [AIRFLOW-2955] Remove unused library.

* Resolved conflicts.

* [AIRFLOW-2955] Fix kubernetes pod operator to set requests and limits on task pods.

* [AIRFLOW-2955] Remove bare except to follow flake8.

* [AIRFLOW-2955] Remove unused library.

* Resolved conflicts.

* Resolve conflicts.

* [AIRFLOW-2955] Remove bare except to follow flake8.

* [AIRFLOW-2955] Remove unused library.

* [AIRFLOW-2955] clear up commits.

* Resolve nits form @galuszkak and @dimberman.
milton0825 and others added 22 commits May 14, 2019 06:36
Ensure that backfill respects task_concurrency.
That is, the number of concurrent running tasks
across DAG runs should not exceed task_concurrency.
* added example of a function returning a dag object
Add papermill operator to productize python notebooks.
…pache#5261)

* Move k8s executor from contrib folder

Considering that the k8s executor is now fully supported by core
committers, we should move it from contrib to the primary executor
directory.
* HA for Metastore

* [AIRFLOW-3888] HA for metastore connection

Creating a connection to a metasotor with two hosts for high avitablity (eg connection 1, connection 2) is not possible because the entire value entered is taken. For our needs, it is necessary to go through subsequent hosts and connect to the first working.

This change allows you to check and then connect to a working metastor.

* add function to base_hook

* update webhdfs_hook

* back to original version

* back to original version

* Update hive_hooks.py

Thank you. I made a few changes because during the tests I detected several errors.

I have a question, when I do marge to my pull it will be  still possible to land it in the airflow main branch?

* [AIRFLOW-3888] HA for metastore connection 

flake8 code repair

* [AIRFLOW-3888] HA for metastore connection 

Flake8 repair

* [AIRFLOW-3888] HA for metastore connection

Code behavior improvements

* [AIRFLOW-3888] HA for metastore connection 

Add test

* [AIRFLOW-3888] HA for metastore connection

test improvement

* [AIRFLOW-3888] HA for metastore connection

Add test

[AIRFLOW-3888] HA for metastore connection

test improvement

* [AIRFLOW-3888] HA for metastore connection

Add test

[AIRFLOW-3888] HA for metastore connection

test improvement

[AIRFLOW-3888] HA for metastore connection

test improvement

* [AIRFLOW-3888] HA for metastore connection

Improving the typo in the variable name

*  [AIRFLOW-3888] HA for metastore connection

Mock return_value edit

* [AIRFLOW-3888] HA for metastore connection

Flake8 repair

* [AIRFLOW-3888] HA for metastore connection

Test repair

* [AIRFLOW-3888] HA for metastore connection

Flake8 repair

[AIRFLOW-3888] HA for metastore connection

Test repair
…che#4923)

* [AIRFLOW-4092] Add gRPCOperator, unit test and added to auto doc

* [AIRFLOW-4092] fix documentation errors

* [AIRFLOW-4092] remove hook dispatcher and auth_type as we don't use it now
* AIRFLOW-4174 Fix run with backoff

* AIRFLOW-4174 Fix flake 8 issues
…e#5264)

* [AIRFLOW-4457] Enhance Task logs by providing the task context
update test pass in s3_overwrite kwarg
[AIRFLOW-4479] Imap to s3 overwrite not wired up
Add replace=self.s3_overwrite to s3_hook.load_bytes
update test pass in s3_overwrite kwarg
@JCoder01 JCoder01 closed this May 22, 2019
@JCoder01 JCoder01 deleted the AIRFLOW-4479 branch May 22, 2019 02:35
@codecov-io
Copy link

Codecov Report

Merging #5311 into master will increase coverage by <.01%.
The diff coverage is 100%.

Impacted file tree graph

@@            Coverage Diff             @@
##           master    #5311      +/-   ##
==========================================
+ Coverage   78.92%   78.92%   +<.01%     
==========================================
  Files         479      479              
  Lines       30098    30098              
==========================================
+ Hits        23755    23756       +1     
+ Misses       6343     6342       -1
Impacted Files Coverage Δ
...ontrib/operators/imap_attachment_to_s3_operator.py 100% <100%> (ø) ⬆️
airflow/models/taskinstance.py 92.59% <0%> (+0.17%) ⬆️

Continue to review full report at Codecov.

Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update b8d8371...9ea8039. Read the comment docs.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.