-
Notifications
You must be signed in to change notification settings - Fork 63
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Config Reloader v0.6.x Crash Loop on "Too Many Open Files" #68
Comments
describe output:
|
Thanks @Kampe for the report, looking into this... |
Excellent, thank you! More interestingly, sometimes you can get one or two pods of a replica-set to stand up
|
Hi @Kampe would it be possible to try setting the following in your Helm chart to check if this fixes it for you?
|
We've had great preliminary success with this image in a few environments! |
I'm facing the very same issue in KinD cluster running on Linux. I already tried both your ad-hoc image The result is that the So (at least in my case) the root cause is not by missing
that actually calls EDIT: I checked my sysctl settings. I have Hope it helps. |
The issue still appeears on |
the NATS install that comes bundled in with TestKube also suffers from the same issue: pod/testkube-test-nats-0 2/3 CrashLoopBackOff 25 (3m8s ago) 54m $ kubectl logs -n testkube-test pod/testkube-test-nats-0 reloader This is installed by helm chart version: testkube-1.13.33 I tried increasing ulimit -n (max open file descriptors) from 1024 to 4096 but it made no difference. Anything I can do to further narrow it down? |
Same issue here. It seems random and I can get 1-3 pods running, but it was hard to do and I'm not sure why. The host has been configured with a large open file allowance. |
I ran into this issue as well running NATS 2.10.11 (NATS Server Reloader v0.14.1) on a k3s v1.28.3 cluster on Ubuntu 22.04 nodes. I had just one of my nodes having this issue as well. I noticed that I had a high number of pod restarts for a deployment on only that node having issues. After resolving the issue with my deployment and restarting my node that had issues everything has been working fine. I suspect that my excessive pod restarts caused too many files to remain open. I know that others have discussed ulimits. I just wanted to point out that my k3s processes are running as a systemd service because Ubuntu has switched to systemd from SysV. I know systemd runs each service with it's own open file limits and ignores ulimit. Here is a link to Red Hat documentation, but will likely pertain to whatever Linux distribution is using systemd. I hope someone else finds this helpful. |
I'll just add few notes from my case. For me, after setting all the open file limits (systemd, pam, etc) as permissive as possible, I was still running into exact same error. Turns out it was caused by |
Posting this here as the code seems to live here in this repo:
nats-io/k8s#488
The text was updated successfully, but these errors were encountered: