-
Notifications
You must be signed in to change notification settings - Fork 999
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Cannot disrupt Node: state node is nominated for a pending pod #7521
Comments
Hi @vb-atelio, |
@jigisha620 A node is marked for deletion due to expiration, but it hosts a pod with the At the same time, Karpenter nominates the pod from the node marked for deletion to another node. For example, the nomination logic can be found here. The new node receiving the nominated pod might be empty or underutilized, but due to the presence of the nominated pod, Karpenter cannot disrupt it. Karpenter version: 1.1.1 |
@jigisha620 any updates here? |
I have the same issue. The node in question isn't running anything except normal daemonsets. So it should be ripe for deletion, yet for some reason karpenter is refusing to delete. I noticed this when one of my subnets ran out of IPs... I'm manually deleting the nodes now. |
same here
But I have a pending pod which should be scheduled on the node above, but can't
The only logs I have for the pod is
and for the node, no logs except the ones about the node creation |
Can confirm the same even wtih spotToSpotConsolidation set to true, for the following budget
or simply with default budget:
with pods arriving every 2-3minutes, still getting
On 0.37.6 this worked correctly with node being replaced, on 1.0.8 - broken |
@jigisha620 same issue on v1.1.1. 64-core node (spot) with even spotToSpot enabled not consolidated with replace over almost 2hrs: |
Some additional context here - in our case I think the reason was that Once I changed the period to 15s, I finally saw the following in logs:
So it seems that |
Description
Observed Behavior:
Karpenter refused to drain a node(instance type: m7i.12xlarge) when it's clearly underutilized(has 8 pods running) with reason: state node is nominated for a pending pod. When I run
kubectl get pods --all-namespaces --field-selector=status.phase=Pending
I see that there are no pending pods.Expected Behavior:
Karpenter should be disrupting this node and draining it and scheduling these pods on another node or atleast show the correct reason on why it's not able to drain the node
Reproduction Steps (Please include YAML):
nodepool.yaml
Versions:
kubectl version
):1.30The text was updated successfully, but these errors were encountered: