-
Notifications
You must be signed in to change notification settings - Fork 13.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
State of this instance has been externally set to removed. Taking the poison pill. #8087
Comments
Thanks for opening your first issue here! Be sure to follow the issue template! |
I had the same problem. 🤔 |
any update on this? |
@K-7 Can you test it with Airflow 2.0.2 and see if you still see this error, in which case can you also provide steps for reproduction please |
we are seeing this from time to time on
|
I am using 1.10.12 and facing same issue due to which my GlueJob called two times, below are the highlighted log where glue job called two times. Scenario : I have used one dummy operator and success of that task i have called glue job using 'on_success_callback' callback function Note : This is coming intermittently, not everytime. [2021-05-15 03:30:10,863] {{taskinstance.py:901}} INFO - Executing <Task(DummyOperator): insert_run_control> on 2021-05-14T03:30:00+00:00 |
I am having the same issue. Atleast no poison pill |
#16637 Seems to be the same issue |
Hi All, |
Airflow 1.10 reached end-of-life on June 17th 2021 and it will not receive any updates any more. Please upgrade to Airflow 2 at the earliest convenient time. Note that there are already security fixes addressed in 2+ Airflow series that are not fixed in 1.10 so you are putting yourself in high risk by not upgrading. You can find out more about it by watching the recording of the discussion panel we just had at the Airflow Summit 2021: https://www.crowdcast.io/e/airflowsummit2021/3 |
We have tried AWS + Airflow 2.0, but Airflow 2.0 has multiple issue. Main Issue in Airflow 2.0 :
|
Airflow is now at 2.1.1 and 2.1.2 will be released next week. And a number of stability improvements have been already released in 2.1 line. If this is limitation of the MWAA to only support 2.0.0, then you should raise issue there. Also if you have an SLA with managed service to support 1.10 for longer, then you should raise the issue in that managed service. Maybe they can diagnose/fix it. The 1.10 line is no longer supported here, sorry - that's the policy that we introduced already more than a year ago. |
Sorry, this seems to be a very old ticket that I had raised. Since I dint not get any response from Airflow back those days even i have forgotten about it. I dont remember how exactly I fixed this issue, but one thing I am very sure of is Airflow needs to do a better job in error handling. The error messages dont clearly indicate the root cause of the problem. Recently when I was trying the latest version of Airflow I remember Any way this ticket can be closed since I am not able to recall the exact steps to reproduce this issue |
hey @K-7 - Airflow is a community managed product that you get for free. I think if you are aware of such issue, absolutely the best what you can do is to raise pull request fixing it ! This is best way to thank community for all the work they to so that you can use the software for free. Many of our users do that and we even yesterday run a workshop for first time contributors to help them to learn how to do it. It seems that fixing error message seems to be an easy task to do. Would you be willing to fix it @K-7 ? this way you could help others to get better diagnostics. |
Yup I do that, I remember I had raised a PR for supporting some AWS operator. But not sure whether that PR got merged or what happened with it |
Apache Airflow version: 1.10.2
Environment: Linux EC2 Machine
Cloud provider or hardware configuration: AWS
What happened:
AIiflow Tasks are killed by passing SIGTERM signal
What you expected to happen:
ECSOperator triggered from Airflow has to work smoothly when the AWS policies are attached correctly to the role.
How to reproduce it:
Using Airflow Dags when I run ECS tasks using ECSOperator, the tasks are first triggered and a response of 200 is received which I see in the logs. But the very next log message is
State of this instance has been externally set to removed. Taking the poison pill
This goes and kills the ECS task and a response of 'desiredStatus': 'STOPPED' is returned back. The log messages does not clearly say as why the task was killed.
Under airflow.cfg the configurations are as follow
parallelism = 32
dag_concurrency = 16
dags_are_paused_at_creation = True
max_active_runs_per_dag = 16
non_pooled_task_slot_count = 128
Kindly change the log messages so that we understand what is the root cause of the issue
The text was updated successfully, but these errors were encountered: