In our Kuberenetes cluster, we encounter sporadic situations where the cluster node runs out of memory and Linux calls the OOM killer. Looking at the logs, it seems that tabs planned on node require more memory than Node can allocate.
The problem is that when OOM killer is called, it prints a list of processes and their use in memory. However, since all of our Docker containers are Java services, the "process name" simply displays as "java", preventing us from tracking which specific Pod is causing the problems.
How can I get a history of what Pods were planned to run on a particular node and when?
source
share