-
Notifications
You must be signed in to change notification settings - Fork 954
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Karpenter constantly scales-up for a single pod #2821
Comments
Can we add a integ test that enforces this? Runaway scaling test? Maybe a separate suite? |
This fits into the category of chaos/failure testing to me. Was planning to create a chaos testing Describe block in E2ETesting |
Able to repro a hypothesis for the runaway scaling issue by continually tainting the node after launch using Karpenter while true
do
kubectl get nodes -A --selector karpenter.sh/provisioner-name | cut -d " " -f 1 | xargs -I "{}" kubectl taint node {} special=true:NoExecute
sleep 1
done This issue was resolved with #2614 by removing the stabalization window, which would have prevented empty node removal during the infinite scale-up. This change will be released as part of the Provisioners that were using It's worth noting that without enabling either |
Tracked down the issue: The problem is that The solution here is to migrate the userData configuration with |
Version
Karpenter Version: v0.16.3
Expected Behavior
Karpenter should only provision a single node for the pod that needs to be scheduled on the cluster.
Actual Behavior
Karpenter schedules a node, the pod does not schedule to that node so Karpenter continually provisions nodes every 20s.
Steps to Reproduce the Problem
Unclear what the repro steps are at this time.
Resource Specs and Logs
Deployment Spec
Community Note
The text was updated successfully, but these errors were encountered: