Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Log 1098 - Playbook for Critical Alerts #673

Merged

Conversation

sasagarw
Copy link
Contributor

Description

This PR:

  • Ensures that critical alerts have proper diagnostic steps and action steps.

/cc @lukas-vlcek
/assign @ewolinetz

Links

@openshift-ci-robot openshift-ci-robot added the do-not-merge/work-in-progress Indicates that a PR should not merge because it is a work in progress. label Mar 11, 2021
@sasagarw sasagarw changed the title [WIP] Log 1098 - Playbook for Critical Alerts Log 1098 - Playbook for Critical Alerts Mar 22, 2021
@openshift-ci-robot openshift-ci-robot removed the do-not-merge/work-in-progress Indicates that a PR should not merge because it is a work in progress. label Mar 22, 2021
docs/alerts.md Outdated Show resolved Hide resolved
docs/alerts.md Outdated Show resolved Hide resolved
docs/alerts.md Outdated Show resolved Hide resolved
docs/alerts.md Outdated Show resolved Hide resolved
docs/alerts.md Outdated Show resolved Hide resolved
```
oc logs <elasticsearch_node_name> -c elasticsearch -n openshift-logging
```

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

We probably should have a follow up step here... but this starts to get really tricky... @lukas-vlcek can you think of some steps we can take here? Do we just want to try to restart the nodes that haven't joined? but if there's a cert issue we need to figure out which one has the correct certs... also the operator should be doing something there already...

@ewolinetz
Copy link
Contributor

@jcantrill can you also try to look through some of these steps based on your past customer experiences?

```
oc exec -n openshift-logging -c elasticsearch <elasticsearch_pod_name> -- es_util --query=<elasticsearch_index_name> -X DELETE
```

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

i think we also want to add a step to unlock all the indices now that the watermark level is below the threshold. ES will lock the indices automatically but it will not unlock them for ES6

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@ewolinetz what I had found is that ES will lock the indices only on reaching flood watermark level and not on low or high. That's why I have mentioned a step to unlock the indices in flood watermark troubleshooting.

docs/alerts.md Outdated Show resolved Hide resolved
oc exec -n openshift-logging -c elasticsearch <elasticsearch_pod_name> -- es_util --query=<elasticsearch_index_name>/_settings?pretty
```
- Identify the number of replicas from the output of the above command.
- Lower the number of replicas if possible:
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

this would need to be cluster-wide... the operator would try to adjust them afterwards I believe

@ewolinetz
Copy link
Contributor

@openshift/sre-alert-sme could you also take a look through some of these and comment?

@RiRa12621
Copy link

/assign @RiRa12621

I'll check it out tomorrow, unless one of the APAC folks has time before that

@RiRa12621
Copy link

sorry for the delay, lgtm from my perspective.
This should give users a good path to fix given problems

@sasagarw
Copy link
Contributor Author

/retest

@openshift-ci
Copy link
Contributor

openshift-ci bot commented Apr 12, 2021

@sasagarw: The following test failed, say /retest to rerun all failed tests:

Test name Commit Details Rerun command
ci/prow/e2e-upgrade 8324809 link /test e2e-upgrade

Full PR test history. Your PR dashboard.

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. I understand the commands that are listed here.

@ewolinetz
Copy link
Contributor

/lgtm

@openshift-ci-robot openshift-ci-robot added the lgtm Indicates that a PR is ready to be merged. label Apr 12, 2021
@openshift-ci-robot
Copy link

[APPROVALNOTIFIER] This PR is APPROVED

This pull-request has been approved by: ewolinetz, sasagarw

The full list of commands accepted by this bot can be found here.

The pull request process is described here

Needs approval from an approver in each of these files:

Approvers can indicate their approval by writing /approve in a comment
Approvers can cancel approval by writing /approve cancel in a comment

@openshift-ci-robot openshift-ci-robot added the approved Indicates a PR has been approved by an approver from all required OWNERS files. label Apr 12, 2021
@openshift-merge-robot openshift-merge-robot merged commit 5655fe4 into openshift:master Apr 12, 2021
@jeremyeder
Copy link

Thank you @ewolinetz !

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
approved Indicates a PR has been approved by an approver from all required OWNERS files. lgtm Indicates that a PR is ready to be merged.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

6 participants