Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

cephadm: persist the grafana.db file #40537

Merged
merged 1 commit into from Apr 10, 2021
Merged

cephadm: persist the grafana.db file #40537

merged 1 commit into from Apr 10, 2021

Conversation

pcuzner
Copy link
Contributor

@pcuzner pcuzner commented Apr 1, 2021

This patch persists the grafana.db file, so the user can
create there own dashboard referencing ceph and node
metrics and not lose them on grafana restart! This also
ensures changes to users/passwords within grafana are
persisted.

Fixes: https://tracker.ceph.com/issues/49954

Signed-off-by: Paul Cuzner pcuzner@redhat.com

This patch persists the grafana.db file, so the user can
create there own dashboard referencing ceph and node
metrics and not lose them on grafana restart! This also
ensures changes to users/passwords within grafana are
persisted.

Fixes: https://tracker.ceph.com/issues/49954

Signed-off-by: Paul Cuzner <pcuzner@redhat.com>
@pcuzner pcuzner added the cephadm label Apr 1, 2021
@pcuzner pcuzner requested a review from a team as a code owner April 1, 2021 04:42
@sebastian-philipp
Copy link
Contributor

I think we have to do more than just mounting a local path.

Imagine we have two grafana daemons deployed on host1 and host2. How do we make sure those grafana.dbs are equal? How do we make sure a newly deployed third daemon will have the same users?

If we don't properly manage the database file on the host, we're going to expose an inconsistent state across multiple daemons of the same service. I know this is also true for Prometheus to some degree, but in contrast to Grafana, Prometheus instances are supposed to be somewhat independent.

Also, we can't store the gafana.db in RADOS, as this would make the monitoring unavailable in case of a cluster error.

@liewegas liewegas changed the title cephadm:persist the grafana.db file cephadm: persist the grafana.db file Apr 1, 2021
@pcuzner
Copy link
Contributor Author

pcuzner commented Apr 5, 2021

@sebastian-philipp - this is a regression from how a ceph-ansible deployment of grafana worked. This patch addresses this regression.

Your points about HA of the grafana and prometheus are valid but HA of these 3rd party services will require orch and dashboard changes which is perhaps better targeted for Quincy. My goal here is to fix the regression for the backport to Pacific.

@liewegas
Copy link
Member

liewegas commented Apr 5, 2021

@sebastian-philipp - this is a regression from how a ceph-ansible deployment of grafana worked. This patch addresses this regression.

Your points about HA of the grafana and prometheus are valid but HA of these 3rd party services will require orch and dashboard changes which is perhaps better targeted for Quincy. My goal here is to fix the regression for the backport to Pacific.

Sounds reasonable to me. This is the same durability model that we have for prometheus.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
3 participants