-
Notifications
You must be signed in to change notification settings - Fork 50
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Memory leak in ruby process? #368
Comments
Hey, thanks for reporting this! Can you provide some more detail about your environment and agent so that we can try to reproduce? |
This is a single Compute Engine VM running on GCP's |
Just had a failure of the whole cluster, because of Infrastructure is complelty logged with google-fluentd and prometheus, ubuntu/centos servers for docker hosts+nginx load balancers and some custom apps. Error recognized because the docker hosts where not able to run some containers (OOM), nginx still manged to work with 2% free memory. Don't know when it started, had a similiar issue a 2-3 months ago where I thought docker logs where the problem and during that time a reboot-fix was all i had time for. For most servers a Log of
On some servers, the Log on such a server only shows the latest
Commands used: sudo systemctl restart google-fluentd
top # ram was not free
sudo systemctl restart google-fluentd # took ages
sudo systemctl stop google-fluentd # took ages
sudo systemctl kill google-fluentd
sudo systemctl start google-fluentd
top # ram was not free
sudo kill -9 <pid-or-ruby>
top # ram was free
sudo reboot # just to be sure everything was/is running as it should What further infos would help and how could I get them easily? |
We are running into the same issue: It is installed on a VM instance running COS 89 with the metadata tag restarting the stackdriver-logging-agent mitigates the issue, but it will come back in a while. the specific image with the memory leak is: An older version of that image is not having the same memory issue: |
Is this the same as fluent/fluentd#3401? [EDIT] Actually, [ANOTHER EDIT] Mystery potentially solved: the |
I had one of these killed recently by the linux OOM killer – after 87 days running on the google cloud:
There seems to be a memory leak somewhere.
The text was updated successfully, but these errors were encountered: