Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

v3.2.6 Argo UI failed to render pod log #7595

Closed
Tracked by #7589
liuzqt opened this issue Jan 20, 2022 · 9 comments · Fixed by #7605
Closed
Tracked by #7589

v3.2.6 Argo UI failed to render pod log #7595

liuzqt opened this issue Jan 20, 2022 · 9 comments · Fixed by #7605
Assignees

Comments

@liuzqt
Copy link

liuzqt commented Jan 20, 2022

Summary

What happened/what you expected to happen?
image-2022-01-19-22-04-27-808

What version of Argo Workflows are you running?
3.2.6 (we just upgrade it from 3.2.3 today, so I guess this issue might be introduced between 3.2.3 and 3.2.6. We'll try to roll back to 3.2.3 and see if this issue still exist)

Diagnostics

I'm still not able to find a minimal reproducing example, since this issue only happened in some of our pipelines, and not sure which part of the template triggered this issue.


Message from the maintainers:

Impacted by this bug? Give it a 👍. We prioritise the issues with the most 👍.

@alexec
Copy link
Contributor

alexec commented Jan 20, 2022

Most likely introduced by this commit:

f5f6899 fix: Add pod name format annotation. Fixes #6962 and #6989 (#6982)

@JPZ13 would you be able to look into this please?

@alexec alexec changed the title Argo UI failed to render pod log v3.2.6 Argo UI failed to render pod log Jan 20, 2022
@alexec alexec removed the triage label Jan 20, 2022
@JPZ13
Copy link
Member

JPZ13 commented Jan 20, 2022

Most likely introduced by this commit:

f5f6899 fix: Add pod name format annotation. Fixes #6962 and #6989 (#6982)

@JPZ13 would you be able to look into this please?

On it

@alexec alexec added this to To do in Run The Business (incl. bugs) via automation Jan 20, 2022
@JPZ13
Copy link
Member

JPZ13 commented Jan 20, 2022

@liuzqt Can you share what the pod names are in kubernetes for this workflow that are resulting in the error on the UI? I'm working on a fix and need to make sure there's a 1:1 match between the pod names as they appear in k8s and what our UI pod name function returns

Thanks for reporting the issue!

@liuzqt
Copy link
Author

liuzqt commented Jan 20, 2022

@JPZ13 the Pod UI is not working, I tried to find their information from the json meta, hopefully this can help:

"patch-processing-pipeline-ksp78-1623891970": {
                "id": "patch-processing-pipeline-ksp78-1623891970",
                "name": "patch-processing-pipeline-ksp78.retriable-map-authoring-initializer",
                "displayName": "retriable-map-authoring-initializer",
                "type": "Pod",
                "templateScope": "local/",
                "phase": "Succeeded",
                "boundaryID": "patch-processing-pipeline-ksp78",
                "startedAt": "2022-01-20T01:45:35Z",
                "finishedAt": "2022-01-20T01:45:39Z",
                "progress": "1/1",
                "resourcesDuration": {
                    "cpu": 3,
                    "memory": 3
                },
                # ...
"patch-processing-pipeline-ksp78-1769534216": {
                "id": "patch-processing-pipeline-ksp78-1769534216",
                "name": "patch-processing-pipeline-ksp78.NODE-RETRIABLE-MAP-AUTHORING(1).NODE-GENERATE-EMLANE.pre-emlane-map-builder",
                "displayName": "pre-emlane-map-builder",
                "type": "Pod",
                "templateRef": {
                    "name": "map-builder-task",
                    "template": "map-builder"
                },
                "templateScope": "namespaced/node-generate-emlane-task",
                "phase": "Failed",
                "boundaryID": "patch-processing-pipeline-ksp78-242024017",
                "message": "Error (exit code 1)",
                "startedAt": "2022-01-20T01:53:05Z",
                "finishedAt": "2022-01-20T01:53:08Z",
                "progress": "1/1",
                "resourcesDuration": {
                    "cpu": 4,
                    "memory": 4
                },
                # ...

@JPZ13
Copy link
Member

JPZ13 commented Jan 20, 2022

That is super helpful @liuzqt. Thank you for the extra info. I'll tag you on the PR for the fix once it's ready

JPZ13 added a commit to JPZ13/argo-workflows that referenced this issue Jan 20, 2022
Signed-off-by: J.P. Zivalich <jp@pipekit.io>
@liuzqt
Copy link
Author

liuzqt commented Jan 20, 2022

@JPZ13 sorry actually I'm just an Argo user in our company and I have to ask our Argo team to help verify this fix later. I'll update once I get feedback from them.

@JPZ13
Copy link
Member

JPZ13 commented Jan 20, 2022

@JPZ13 sorry actually I'm just an Argo user in our company and I have to ask our Argo team to help verify this fix later. I'll update once I get feedback from them.

Sounds good. Thank you @liuzqt

@terrytangyuan
Copy link
Member

terrytangyuan commented Jan 21, 2022

@JPZ13 I am in touch with @liuzqt’s team. Let me know if there’s any other information you need.

@JPZ13
Copy link
Member

JPZ13 commented Jan 21, 2022

Thanks @terrytangyuan. I'm looking for the mapping of the json meta of NodeStatuses that @liuzqt provided above to the pod names as they appear in the kubernetes cluster. Also, if they find a minimal reproducible example, that would be super helpful too

@alexec alexec mentioned this issue Feb 2, 2022
28 tasks
alexec pushed a commit that referenced this issue Feb 4, 2022
#7605)

Signed-off-by: J.P. Zivalich <jp@pipekit.io>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging a pull request may close this issue.

5 participants