Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Installation : Extra Packages #21

Closed
afarbos opened this issue Jun 12, 2015 · 7 comments
Closed

Installation : Extra Packages #21

afarbos opened this issue Jun 12, 2015 · 7 comments

Comments

@afarbos
Copy link

afarbos commented Jun 12, 2015

Hi,

I see in your documentation some subpackage (http://pythonhosted.org/airflow/installation.html#extra-packages). But I never heard about that with pip, moreover I don't see them using pip search airflow.

@artwr
Copy link
Contributor

artwr commented Jun 12, 2015

Hi @glodus,
Packages extra are actually additional optional dependencies to enable optional functionality. Since airflow can connect with different databases, but we do not expect you to have all of them, you have the option of only installing the requirements you need.

You can see the detail here

@mistercrunch
Copy link
Member

It's part of setuptools and supported by pip though subpackages don't seem to show when using pip search. You can dig for extras_require in here for more info: https://pythonhosted.org/setuptools/setuptools.html#declaring-dependencies

For the full details of the packages and what they include, it's all in here:
https://github.com/airbnb/airflow/blob/master/setup.py#L56

The pip calls as documented should work, provided that you have the requirements for the specified packages to install. The reason why we created subpackages is that some packages are tricky to install and require some os level libs and we wanted to make it more modular where you just install what you need.

Note that the ipython project also uses extras_require

@afarbos
Copy link
Author

afarbos commented Jun 12, 2015

Thanks for the answer, I all ready try the setup.py script but I have still so many warnings :

  • WARNING - Couldn't import module postgres_hook
  • WARNING - Couldn't import module hive_hooks
  • WARNING - Couldn't import module presto_hook
  • WARNING - Couldn't import module samba_hook
  • WARNING - Couldn't import module hive_to_samba_operator
  • WARNING - Couldn't import module hive_to_mysql
  • WARNING - Couldn't import module mysql_to_hive
  • WARNING - Couldn't import module hive_operator
  • WARNING - Couldn't import module hive_stats_operator
  • WARNING - Couldn't import module s3_to_hive_operator
  • WARNING - Couldn't import module postgres_operator
  • WARNING - Couldn't import module presto_check_operator

The setup.py should fix that no ? I try to do setup.py install.

Just like that, don't forget to add a shebang to setup.py (#!/usr/bin/env python2.7).

@afarbos
Copy link
Author

afarbos commented Jun 12, 2015

I fix most of them pip -r requierements.txt.

If I have other question should I open new issues or do you have an ml or something like that?

@afarbos afarbos closed this as completed Jun 12, 2015
@afarbos afarbos reopened this Jun 12, 2015
@mistercrunch
Copy link
Member

@glodus , the setup.py script doesn't install extras_require by default, I looked for switched before to try to make it work but couldn't find any. If you want all dependencies, you can do:
pip install -r requirements.txt

Some of the libs in there require os packages to be installed. You could also just follow the documentation and pip install airflow[subpackage]

@statwonk
Copy link
Contributor

@mistercrunch @artwr I ran into this issue, too. I struggled until it occurred to me to look at the closed issues. I understand the reason you don't include the dependencies, yagni principle. That said, it might be good to find a way to suppress the warnings. As someone new to Airflow, the warnings were a huge distraction and I do believe they might dissuade others from going further.

@mistercrunch
Copy link
Member

#28 addresses the warnings

aoen pushed a commit to aoen/incubator-airflow that referenced this issue Apr 10, 2020
* 99ee040: CP from 1.10+twtr

* 2e01c24: CP from 1.10.4 ([TWTR][AIRFLOW-4939] Fixup use of fallback kwarg in conf.getint)

* 00cb4ae: [TWTR][AIRFLOW-XXXX] Cherry-pick d4a83bc and bump version (apache#21)

* CP 51b1aee: Relax version requiremets (apache#24)

* CP 67a4d1c: [CX-16266] Change with reference to 1a4c164 commit in open source (apache#25)

* CP 54bd095: [TWTR][CX-17516] Queue tasks already being handled by the executor (apache#26)

* CP 87fcc1c: [TWTR][CX-17516] Requeue tasks in the queued state (apache#27)

* CP 98a1ca9: [AIRFLOW-6625] Explicitly log using utf-8 encoding (apache#7247) (apache#31)

* fixing models.py and jobs.py file fix after CP

* fixing typo and version bump

Co-authored-by: Vishesh Jain <visheshj@twitter.com>
mobuchowski pushed a commit to mobuchowski/airflow that referenced this issue Jan 4, 2022
* Adding new version of marquez-python and the resulting api changes in the test and integration

* Remove coverage from being checked in and adjust gitignore

* Remove requests from dependencies
mobuchowski pushed a commit to mobuchowski/airflow that referenced this issue Jan 4, 2022
Signed-off-by: Julien Le Dem <julien@apache.org>
rajatsri28 pushed a commit to rajatsri28/airflow that referenced this issue Jan 25, 2022
* EWT-569 : Initial Commit for migrations

* [EWT-569] Airflow Upgrade to 1.10.14, Cherry-Pick  76fe7ac from 1.10.4

* CP Contains fb64f2e: [TWTR][AIRFLOW-XXX] Twitter Airflow Customizations + Fixup job scheduling without explicit_defaults_for_timestamp

* [EWT-569] Airflow Upgrade to 1.10.14, Cherry-Pick 91d2b00
[CP][EWT-548][AIRFLOW-6527] Make send_task_to_executor timeout configurable (apache#63)

* [EWT-569] Airflow Upgrade to 1.10.14, Cherry-Pick 91d2b00
CP contains [EWT-16]: Airflow fix for manual trigger during version upgrade (apache#13)

* [EWT-16]: Airflow fix for manual trigger during version upgrade

* [EWT-569] Airflow Upgrade to 1.10.14, Cherry-Pick 91d2b00
[CP][EWT-548][AIRFLOW-6527] Make send_task_to_executor timeout configurable (apache#63)

CP of f757a54

* CP(55bb579) [AIRFLOW-5597] Linkify urls in task instance log (apache#16)

* [EWT-569] Airflow Upgrade to 1.10.14 [CP] from 1.10.4+twtr : 94cdcf6
[CP] Contains [AIRFLOW-5597] Linkify urls in task instance log

CP of f757a54

* [EWT-569] Airflow Upgrade to 1.10.14, Cherry-Pick  4ce8d4c from 1.10.4
CP contains [TWTTR] Fix for rendering code on UI (apache#34)

* [EWT-569] Airflow Upgrade to 1.10.14, Cherry-Pick  299b4d8 from 1.10.4
CP contains [TWTR] CP from 1.10+twtr (apache#35)

* 99ee040: CP from 1.10+twtr

* 2e01c24: CP from 1.10.4 ([TWTR][AIRFLOW-4939] Fixup use of fallback kwarg in conf.getint)

* 00cb4ae: [TWTR][AIRFLOW-XXXX] Cherry-pick d4a83bc and bump version (apache#21)

* CP 51b1aee: Relax version requiremets (apache#24)

* CP 67a4d1c: [CX-16266] Change with reference to 1a4c164 commit in open source (apache#25)

* CP 54bd095: [TWTR][CX-17516] Queue tasks already being handled by the executor (apache#26)

* CP 87fcc1c: [TWTR][CX-17516] Requeue tasks in the queued state (apache#27)

* CP 98a1ca9: [AIRFLOW-6625] Explicitly log using utf-8 encoding (apache#7247) (apache#31)

* [EWT-569] Airflow Upgrade to 1.10.14 [CP] from 1.10.4+twtr : f7050fb
CP Contains Experiment API path fix (apache#37)

* [EWT-569] Airflow Upgrade to 1.10.14, Cherry-Pick  8a689af from 1.10.4
CP Contains Export scheduler env variable into worker pods. (apache#38)

* [EWT-569] Airflow Upgrade to 1.10.14, Cherry-Pick  5875a15 from 1.10.4
Cp Contains [EWT-115][EWT-118] Initialise dag var to None and fix for DagModel.fileloc (missed in EWT-16) (apache#39)

* [EWT-569] Airflow Upgrade to 1.10.14, Cherry-Pick  a68e2b3 from 1.10.4
[CX-16591] Fix regex to work with impersonated clusters like airflow_scheduler_ddavydov (apache#42)

* [EWT-569] Airflow Upgrade to 1.10.14 [CP] from 1.10.4+twtr : e9642c2
[CP][EWT-128] Fetch task logs from worker pods (19ac45a) (apache#43)

* [EWT-569] Airflow Upgrade to 1.10.14 [CP] from 1.10.4+twtr : d5d0a07
[CP][AIRFLOW-6561][EWT-290]: Adding priority class and default resource for worker pod. (apache#47)

* [EWT-569] Airflow Upgrade to 1.10.14 [CP] from 1.10.4+twtr : 9b58c88
[CP][EWT-302]Patch Pool.DEFAULT_POOL_NAME in BaseOperator (apache#8587) (apache#49)

Open source commit id: b37ce29

* [EWT-569] Airflow Upgrade to 1.10.14 [CP] from 1.10.4+twtr : 7b52a71
[CP][AIRFLOW-3121] Define closed property on StreamLogWriter (apache#3955) (apache#52)

CP of 2d5b8a5

* [EWT-361] Fix broken regex pattern for extracting dataflow job id (apache#51)

Update the dataflow URL regex as per AIRFLOW-9323

* [EWT-569] Airflow Upgrade to 1.10.14 [CP] from 1.10.4+twtr : 4b5b977
EWT-370: Use python3 to launch the dataflow job. (apache#53)

* [EWT-569] Airflow Upgrade to 1.10.14 [CP] from 1.10.4+twtr : 596e24f
* [EWT-450] fixing sla miss triggering duplicate alerts every minute (apache#56)

* [EWT-569] Airflow Upgrade to 1.10.14 [CP] from 1.10.4+twtr : b3d7fb4
[CP] Handle IntegrityErrors for trigger dagruns & add Stacktrace when DagFileProcessorManager gets killed (apache#57)

CP of faaf179 - from master
CP of 2102122 - from 1.10.12

* [EWT-569] Airflow Upgrade to 1.10.14 [CP] from 1.10.4+twtr : bac4acd
[TWTR][EWT-472] Add lifecycle support while launching worker pods (apache#59)

* [EWT-569] Airflow Upgrade to 1.10.14 [CP] from 1.10.4+twtr : 6162402
[TWTTR] Don't enqueue tasks again if already queued for K8sExecutor(apache#60)

Basically reverting commit 87fcc1c  and making changes specifically into the Celery Executor class only.

* [EWT-569] Airflow Upgrade to 1.10.14 [CP] from 1.10.4+twtr : 1991419
[CP][TWTR][EWT-377] Fix DagBag bug when a Dag has invalid schedule_interval (apache#61)

CP of 5605d10 & apache#11462

* [EWT-569] Airflow Upgrade to 1.10.14 [CP] from 1.10.4+twtr : 48be0f9
[TWTR][EWT-350] Reverting the last commit partially (apache#62)

* [EWT-569] Airflow Upgrade to 1.10.14 [CP] from 1.10.4+twtr : d8c473e
[CP][EWT-548][AIRFLOW-6527] Make send_task_to_executor timeout configurable (apache#63)

CP of f757a54
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants