-
Notifications
You must be signed in to change notification settings - Fork 183
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Byi more replicas #23
Conversation
@@ -155,7 +155,7 @@ spec: | |||
selector: | |||
matchLabels: | |||
k8s-app: fluentd-sumologic | |||
replicas: 1 | |||
replicas: 3 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Is there any data duplication with 3 replicas of Fluentd running? Or when Prometheus writes to the remote write endpoint do the metrics only get sent to one of the 3 replicas?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
No data duplication. the idea is much similar as our receiver nodes. from outside (e.g. Prometheus) there is only single endpoint (which we defined as a kubernetes service). The traffic will be distributed to 3 nodes. We don't have any session or partition concept here so basically it's just randomly pick anyone
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Nice, thanks for confirming
* start creating test account resources * add sumo env var * move account config into tf config * tf fmt * sign pipeline * specify dag * limit concurrency * update * separate pipelines * chain the pipelines, remove second fmt step * move sumo_env to the drone file * remove vars file * update pipeline * remove experiments
Tuning fluentd resource to fix Context canceled issue
3
replica rather than1
2Gi
to256Mi
(each container)200m
to100m
(each container)Monitoring for couple of hours and looks like no more
context canceled
error in Prometheus log nowFYI - definition of Kubernetes resource limitation
@frankreno @ggarg2906sumo