-
Notifications
You must be signed in to change notification settings - Fork 2.9k
-
Notifications
You must be signed in to change notification settings - Fork 2.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Logging: AWS Elasticsearch: Cannot get new connection from pool #21744
Comments
These settings should be built into our defaults. They are already the defaults in fluentd-kubernetes-daemonset: https://github.com/fluent/fluentd-kubernetes-daemonset/blob/85135da0b9ebc708f5ea5084ef13499f342063eb/templates/conf/fluent.conf.erb#L31-L33 |
It would be great if this change of the defaults would be included in the next Rancher 2.3 release. Is that planned? |
Reproduced in
Deployed an AWS ES Domain. (Domain should have enough EBS disk capacity for a long run)
|
Rancher I left running the log generator for 12 hours without no errors with the same steps I followed to reproduce in
|
Our team has been overriding these settings for the Elastic logging since Rancher 2.2. We have bare metal clusters. We found that |
Cannot reproduce the duplication logs issue in v2.4-dev Changed the NodePort of my ES svc to make it unreachable from fluentd. And change it back after 5 mins. In Kibana, I can see the logs after the fluentd retry. And there is no duplication.
Filed an issue to make it configurable in Rancher. For the duplication issues, I think we should leverage |
Our Elastic Stack is external to the cluster. I am not sure if that has something to do with it. There are probably a number of other factors. like I said we haven't played with this setting since Rancher 2.2. Maybe even the Elastic version could impact this as well. Regardless It does sound like the |
What kind of request is this (question/bug/enhancement/feature request):
Bug
Steps to reproduce (least amount of steps as possible):
Result:
After some period of time, new logs stop showing up in Elasticsearch. Checking the logs of the fluentd container shows the following error:
Other details that may be helpful:
Looking at atomita/fluent-plugin-aws-elasticsearch-service#15 (comment) this appears to be related to how connection reloading works with the ruby library for Elasticsearch and a specific incompatibility with AWS.
Using:
to restart the fluentd pods corrected the issue and backfilled the missing logs.
I think this can be worked around by editing the logging config file in the UI directly to contain:
Environment information
rancher/rancher
/rancher/server
image tag or shown bottom left in the UI): v2.2.5Cluster information
kubectl version
):The text was updated successfully, but these errors were encountered: