Sai
Sai
@alram It is the OSD pod that fails readiness check. Here is the log for it. [osd-log.txt](https://github.com/ceph/ceph-helm/files/1537086/osd-log.txt) Based on the logs, do you think just disabling the readiness/liveness checks could...
@alram OSD actually keeps restarting. I suppose because of the readiness probe failing 3 times as you said. Here is the log: the pod gets terminate signal. ``` + log...
@alram Looks like osds are down. I get this when I do `ceph -s`. ``` cluster: id: 3d691e73-570c-477a-a3ac-a7a574faff6b health: HEALTH_WARN 2 osds down 2 hosts (2 osds) down 1 root...
Sure, will try it and get back soon.
@rootfs Sure. Attached the logs of osd, mgr, mon. [logs-from-ceph-mgr-in-ceph-mgr.txt](https://github.com/ceph/ceph-helm/files/1543914/logs-from-ceph-mgr-in-ceph-mgr-57c4d8c8c5-pj8jn.txt) [logs-from-osd-pod-in-ceph-osd.txt](https://github.com/ceph/ceph-helm/files/1543915/logs-from-osd-pod-in-ceph-osd-brhst.txt) [logs-from-ceph-mon-in-ceph-mon.txt](https://github.com/ceph/ceph-helm/files/1543927/logs-from-ceph-mon-in-ceph-mon-btm4w.txt)
@rootfs @alram Removing the readiness and liveness probe, osd's register with successfully. I guess something is wrong with readiness/liveness.
How can I get the log for live/ready checks? Is it part of the log of the pod?
In the describe, I just see this: `Readiness probe failed: dial tcp 10.211.55.186:6800: getsockopt: connection refused Back-off restarting failed container Error syncing pod`
So, you think it is not a timeout issue? Let me check if I am able to use ceph without ready/live probes.
Yes, I already did that, but I never actually checked mounting cephfs and writing to it.