-
Notifications
You must be signed in to change notification settings - Fork 84
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
OOM kill of the pelorus operator controller manager #777
Comments
To check the node resource usage/limits:
Describe the node (remove /IP from the Node: output):
|
After short debug session:
|
After discussing with @dmesser it may be required to adjust Maximum Concurrent Reconciles: |
@milles9393
Find the
Then save it and we will see if there was any improvement. |
Reading list for wes: |
Pelorus Operator that addresses dora-metrics#777 Signed-off-by: Michal Pryc <mpryc@redhat.com>
Pelorus Operator that addresses dora-metrics#777 Signed-off-by: Michal Pryc <mpryc@redhat.com>
Pelorus Operator that addresses dora-metrics#777 Signed-off-by: Michal Pryc <mpryc@redhat.com>
Pelorus Operator that addresses #777 Signed-off-by: Michal Pryc <mpryc@redhat.com> Signed-off-by: Michal Pryc <mpryc@redhat.com>
The bug should be closed now:
Note that our documentation on the https://pelorus.readthedocs.io/ is currently up to date and reviewed for the Pelrous Operator scenario. |
Another OOM Kill of the pelorus-operator-controller-manager was spotted.
One approach is to update documentation to ensure users do know how to adjust them dynamically.
It's unclear why this OOM Kill happened, from this diagram we can see it did not reach limit of 512MiB, however maybe node has not enough memory to allocate for this pod ?
The text was updated successfully, but these errors were encountered: