question

ShreyasArani-1167 avatar image
0 Votes"
ShreyasArani-1167 asked srbose-msft commented

Pods are going to pending state on particular node of AKS cluster

I have node in my AKS cluster with name aks-agentpool-50820296-vmss00000s.

All the pods deployed on this node are going to pending state.

I have attached the output of describe command of that node


Also attaching the describe output of the pod that is deployed on this particular node104536-pod.txt


I don't know why pods deployed in this node are going in pending state. But pods deployed in other nodes are getting deployed successfully. Please help to resolve this issue.

azure-kubernetes-service
describe.txt (5.6 KiB)
pod.txt (3.1 KiB)
· 3
5 |1600 characters needed characters left characters exceeded

Up to 10 attachments (including images) can be used with a maximum of 3.0 MiB each and 30.0 MiB total.

@ShreyasArani-1167 , based on the information shared it seems like this is a case of Inconsistent daemonset pod behaviour mentioned here. Please try deleting the pod with kubectl delete po promon-prometheus-node-exporter-fg6sp -n monitoring --force and check if the daemonset controller is spawning a new pod in its stead.

If not, then the controller manager on the managed cluster's control plane might not be bouncing. If that is the case please open a Microsoft Support ticket to investigate and troubleshoot further.


0 Votes 0 ·

I executed this command kubectl delete po promon-prometheus-node-exporter-fg6sp -n monitoring --force.
The pod got deleted and new pod was scheduled but still it is showing in pending state.
But when I describe the pod I am able to see it is successfully scheduled.

Events:


> Type Reason Age From Message
> ---- ------ ---- ---- -------
> Normal Scheduled 7m48s default-scheduler Successfully assigned monitoring/promon-prometheus-node-exporter-phsg4 to aks-agentpool-50820296-vmss00000s


So what should I do now?




0 Votes 0 ·

@ShreyasArani-1167 ,

Can you please restart the node virtual machine scale set instance and check again? You should find the corresponding resource ID in the Provider ID value in kubectl describe node

If that doesn't work, can you please upgrade the cluster to a currently supported version? How-to

If the issue still occurs, it will take a deeper investigation and troubleshooting. Please open a Microsoft Support ticket.





0 Votes 0 ·

0 Answers