-
Notifications
You must be signed in to change notification settings - Fork 3.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Workflow - Could not get container status #5251
Comments
How odd. You can clearly see |
Signed-off-by: Alex Collins <alex_collins@intuit.com>
Signed-off-by: Alex Collins <alex_collins@intuit.com>
Signed-off-by: Alex Collins <alex_collins@intuit.com>
Signed-off-by: Alex Collins <alex_collins@intuit.com>
Signed-off-by: Alex Collins <alex_collins@intuit.com>
Signed-off-by: Alex Collins <alex_collins@intuit.com>
Signed-off-by: Alex Collins <alex_collins@intuit.com>
Signed-off-by: Alex Collins <alex_collins@intuit.com>
Signed-off-by: Alex Collins <alex_collins@intuit.com>
Signed-off-by: Alex Collins <alex_collins@intuit.com>
I've created an engineer build for you that has better logging |
Let me try that - Give me a few minutes |
What did I do? Changed executor image (Please let me know if I need to change something else) spec:
containers:
- args:
- --configmap
- workflow-controller-configmap
- --executor-image
- argoproj/argoexec:dev-status
command:
- workflow-controller
env:
- name: LEADER_ELECTION_IDENTITY
valueFrom:
fieldRef:
apiVersion: v1
fieldPath: metadata.name
image: argoproj/workflow-controller:v3.0.0-rc3
imagePullPolicy: IfNotPresent Reran test workflow All failed tasks returned the same error message Container main
Container wait
Controller
|
That's progress. Can you add |
Signed-off-by: Alex Collins <alex_collins@intuit.com>
Signed-off-by: Alex Collins <alex_collins@intuit.com>
I've push a new version with the changes to address short-running pods. Can you test please? |
Should I use |
Yes please. |
@alexec Perfect! I know that we're stressing the short running scenario here, but the new release candidate version has issues with DAG tasks even when it's not short running. I just thought was a good idea to share this with you right now, please find below one example The logs are pretty the same
|
Do you mean |
The logs look older to me. |
|
Yeah. I'm not running this version on production anymore, just captured a little bit before the downgrade. |
Signed-off-by: Alex Collins <alex_collins@intuit.com>
Signed-off-by: Alex Collins <alex_collins@intuit.com>
Signed-off-by: Alex Collins <alex_collins@intuit.com>
Signed-off-by: Alex Collins <alex_collins@intuit.com>
I've another build to test. I fixed a bug, but found a new bug. Can you try |
Signed-off-by: Alex Collins <alex_collins@intuit.com>
Signed-off-by: Alex Collins <alex_collins@intuit.com>
@alexec Sorry about the late, busy day 😃 . What did I do? Pulled the latest version into my environment again wait container logs
|
Signed-off-by: Alex Collins <alex_collins@intuit.com>
Summary
What happened/what you expected to happen?
The workflow should successfully terminate all of your triggered tasks. It wasn't easy to reproduce the anomalous behaviour outside of the cluster and also not happen every single time, so please find additional configurations that might help to reproduce it
Diagnostics
What Kubernetes provider are you using?
GKE or Kind
What version of Argo Workflows are you running?
Workflow template
Controller logs
kubectl logs ci-t8vq2-1533362641 -n argo -c wait
Message from the maintainers:
Impacted by this bug? Give it a 👍. We prioritise the issues with the most 👍.
The text was updated successfully, but these errors were encountered: