Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Bug] Grafana Operator Controller Manager pod failing due to Readiness and Liveness Probe Failing #1294

Closed
TehreemNisa opened this issue Nov 3, 2023 · 5 comments
Labels
bug Something isn't working stale triage/needs-information Indicates an issue needs more information in order to work on it.

Comments

@TehreemNisa
Copy link

**Grafana Operator Controller Pod Restarting **
When the operator is installed it goes intro continous restarts because the Readiness probes and Liveness Probes fail. The issue is resolved if we set the InitialDelaySeconds to 60 for the controller pod. There is no way available right now to do this via subscription resource.

Version
v4.10.1

To Reproduce
Steps to reproduce the behavior:

  1. Go to 'Openshift console -> Operators -> OperatorHub"
  2. Search for grafana operator
  3. Click on 'Install'
  4. Select channel 'v4' and namespace you want to install it in
  5. Click 'Install'
  6. Go to the namespace it is installed in
  7. The 'grafana-operator-controller-manager-*' pod is not in ready state with restarts ,See events of pod it shows liveness and readiness probe failing
    etc.

Expected behavior
The operator should install and the grafana-operator-controller-manager pod should be in running phase

Suspect component/Location where the bug might be occurring
We are installing it on Openshift 4.12.39

Screenshots
image

Runtime:

  • OS: Linux
  • Grafana Operator Version: v4.10.1
  • Environment: Openshift
  • Deployment type: deployed
@TehreemNisa TehreemNisa added bug Something isn't working needs triage Indicates an issue or PR lacks a `triage/foo` label and requires one. labels Nov 3, 2023
@pb82
Copy link
Collaborator

pb82 commented Nov 7, 2023

@TehreemNisa the version you are using is quite old (v4.10.1). Could you please try with the latest release? If there is still a problem, please report back.

@NissesSenap NissesSenap added triage/needs-information Indicates an issue needs more information in order to work on it. and removed needs triage Indicates an issue or PR lacks a `triage/foo` label and requires one. labels Nov 19, 2023
Copy link

This issue hasn't been updated for a while, marking as stale, please respond within the next 7 days to remove this label

@github-actions github-actions bot added the stale label Dec 20, 2023
@github-actions github-actions bot closed this as not planned Won't fix, can't repro, duplicate, stale Dec 27, 2023
@zetti12345
Copy link

I have the same problem with 5.6.0

@HVBE
Copy link
Collaborator

HVBE commented Jan 30, 2024

Hi @zetti12345 , can you share your Grafana CR configuration? we don't get a lot of these reports so its hard for us to diagnose and/or reproduce the issue on our clusters.
Does deleting the deployment/pods remediate the issue?

@zetti12345
Copy link

zetti12345 commented Jan 30, 2024

Hi @HubertStefanski,

no deleting doesn't help. I also reinstalled the operator, but nothing helps and the log looks normal. Somehow the readiness probe of the operator controller manager fails as well as the grafana instance. I use the olm subscription in openshift marketplace. Please find details below:

grafana.zip

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working stale triage/needs-information Indicates an issue needs more information in order to work on it.
Projects
None yet
Development

No branches or pull requests

5 participants