-
Notifications
You must be signed in to change notification settings - Fork 611
No disk space crashloop but pod healthy #3788
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Hey @mausch, I was able to replicate the issue you are seeing with a full disk. After the disk filled up, Postgres stopped working, but the pod was still ready. We use the Patroni As it stands now, if Patroni thinks that the cluster is ready then the pods will be ready. There is likely some work we could do to augment the Patroni readiness endpoint and I will be happy to get a story in our backlog. |
Hi, thanks for looking into this. Shouldn't the liveness probe (rather than readiness) apply here? The Patroni docs say:
I don't have a pgo cluster at hand to check but presumably Patroni's /liveness is already mapped to the pod liveness? |
Has this issue been resolved? My situation is that when restarting a physical machine, after restarting PostgreSQL, the database container fails to self-heal and continually shows a "no response" status. Additionally, it is not possible to restart the container using probes. Your situation seems very similar to mine. @mausch |
Overview
I'm running a trivial CrunchyData instance with 1 primary.
It ran out of disk space possibly due to #2531 but this is not relevant to this issue.
Because of this the postgres pod is stuck in a loop displaying this:
i.e. Postgres isn't running at all, I can't connect to it.
The problems are:
In short, Postgres is broken but the control plane or whatever you want to call it is not aware of it.
Environment
Please provide the following details:
ubi8-16.0-3.4-0
The text was updated successfully, but these errors were encountered: