Questions about performance query and storage cost #2345
Replies: 1 comment 5 replies
-
The answers will depend on your setup but I might be able to give you some hints (hopefully corrected if wrong). First off there's some capacity planning tips here - https://grafana.com/docs/mimir/v2.1.x/operators-guide/running-production-environment/planning-capacity/ Both solutions use Prometheus Storage Engine (2). Unlike solutions such as VictoriaMetrics. So there is no compression as such, there is a compactor component (https://grafana.com/docs/mimir/v2.1.x/operators-guide/architecture/components/compactor/) that will make the LTS more efficient. But, if you're already running Prometheus you could have a rough estimate by multiplying by your desired retention period, eg if you're storing 15 days in Prometheus right now and you want a year you could get the 'worst case scenario'. As for performance for the LTS, you can configure the length of time your queries should use LTS. In other words your queries hit the ingesters first, these values are configurable so if you find most of your queries are for 8 hours, you could easily have those satisfied by the ingesters which have their own persistent storage or use memcached with the query frontend depending on cost concerns. I believe Grafana found most of their queries were under 2 hours though. Caching the query data - https://grafana.com/docs/mimir/v2.1.x/operators-guide/architecture/components/compactor/ - I don't think Thanos offers this - Thanos can cache metadeta but not the query results. There's also a tool for comparing query performance - https://grafana.com/docs/mimir/v2.1.x/operators-guide/tools/query-tee/ so if you do setup a PoC you could confirm your findings. |
Beta Was this translation helpful? Give feedback.
-
Hello
In my organization, we are looking at solutions for long term storage and global view with our Prometheus infrastructures. Among the solutions we are looking closely at Mimir and Thanos.
There are two major problems but we don't necessarily have the answers:
Are there any calculation rules available to evaluate the amount of disk needed to store historical data for a given period of time, depending on the number of samples ingested in Prometheus? Does Mimir apply a compression rate to the data?
For performance, are there any benchmarks or reports to evaluate the execution time of queries with Mimir on data with long term storage?
Without these elements, it is difficult to know if Mimir (or Thanos) is a good candidate for our needs.
Best regards
Beta Was this translation helpful? Give feedback.
All reactions