0

One of our application service pods is getting restarted and the Last cause specified is "Error" due to 143 error code.

There are no errors shown in the last hundred lines of the previous container logs. The only error reference is the Readiness probe entry in the pod's description. However based in all the references I have seen in my searches state that failures in the readiness probe should only be preventing the routing of traffic to the pod instance. Is there a scenario wherein repeated readiness probe failures will cause the Kubernetes pod to be marked unhealthy and be restarted?

 Pod Description

ramfree17
  • 79
  • 2
  • 8
  • Do you have the `livenessProbe` also configured? – Bimal Sep 24 '20 at 03:43
  • yes, liveness probe is also configured. I have seen pods get restarted if they fail the liveness probe but those are explicitly stated in the events of the pod. In the screenshot above, there is no mention that the pod failed to answer the liveness probe. – ramfree17 Sep 24 '20 at 05:00
  • Hi, check if you are not running low on resources and check if storage is not overloaded while app is starting up. Have you tried to increase the `initialDelaySeconds` to check if its not too short? – Piotr Malec Sep 24 '20 at 15:49
  • The top nodes command is reporting that all nodes are within acceptable range (CPU averages at 60%, memory is at 40-45%). The initial delay is 30 seconds and I can ask the maintainers if they can increase it further. Is that something that can cause the pod to be marked unhealthy? – ramfree17 Sep 25 '20 at 09:18
  • Can you show the logs of the pod itself? It seams to me the application never starts successfully. – Mafor Sep 25 '20 at 15:17
  • I checked the last 100 lines of the previous container and there is no error posted. The lines are just showing it is processing requests and the occasional garbage collection printouts (the platform team is using these to monitor VM performance). – ramfree17 Oct 01 '20 at 03:31

0 Answers0