-
Notifications
You must be signed in to change notification settings - Fork 490
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
REPLICATION_FACTOR send metrics over carbon-caches on the same machine. #333
Comments
@toni-moreno, |
@deniszh , we have already done this scheme but is not good enough to scale. What happens if you would like split load in 2/3/4 relays on each machine in any of the tiers ? how can I split load across several relays and also maintain replication factor on different machines? |
Scaling graphite is painful, agreed.
Yep, that's sucks for normal relay. We are using https://github.com/grobian/carbon-c-relay, pretty fast and scalable across CPU cores.
Sorry, did not understand a question. Or it is related with previous one? |
@deniszh my wished scenario would be the following.
I prefer only one Tier to avoid a lot of queued data on several tiers if possible. And also because could be needed to add an extra aggregation tier. do you think Could the carbon-c-relay do this job ? |
@toni-moreno, |
Ok .. What capacity do you have? Is carbon-c-relay instrumented as carbon-relay.py is ? |
I can give you numbers, but YMMV, of course. You need to test your cluster with some load tool. |
Finally using carbon-c-relay as a proxy to other carbon servers. Thank you very much @deniszh. |
We are testing the following graphite scheme, Our goal is to have all data on graphite1 replicated ( as a backup server in graphite2) . After some test, we can see some metrics only on one machine, ( we are using carbon-lookup tool from carbonate also to check where it is sending metrics the relay) , after some test We have seen that REPLICATION_FACTOR + consistent_hashing doesn't know anything about hosts machines only carbon destinations. This is , the relay is sending the same metric twice over the same machine with the following results.
how can configure this scheme avoiding this result?
The text was updated successfully, but these errors were encountered: