I have multipe EKS clusters in separate accounts. These are managed by Rancher 2.6.1. One is kubernetes 1.19 and the another is 1.20. In the last two weeks, I have experienced nodes not getting added to the node group when a scaling event happens. They join the cluster but not the node group. It seems to stop working for no apparent reason and the only fix is to create a new node group and delete the old one. This has happened in two different EKS clusters. Two different rancher systems. Two different AWS accounts. Has anyone else experienced this?
Where can I find logs to troubleshoot this? I opened a ticket with AWS and they basically said to re-create the node group.