You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Describe the bug
I'm trying to deploy web application with pretty standard stack, that contains deployment, service and ingress. We are using 3 worker nodes on EKS, and when I scale replicas to at least 2, Load Balancer UI shows that targets randomly goes unhealthy. Also app responds with 504 Gateway timed out every few requests. It seems that traffic goes only to one pod/worker node, completely ignoring routing to other pods on other nodes.
Worth to notice is that in ALB UI shows that statuses of targets are some kind of random. In one moment all targets are healthy, after refresh only one node is healthy. After second refresh to of them are healthy, with one unhealthy. At least one node is always healthy, as this is the worker that responds for successful requests.
Steps to reproduce
set up v1.30 EKS cluster with 2 or 3 worker nodes
introduce AWS Load Balancer controller via chart:
resource "helm_release" "alb_controller" {
name = "aws-load-balancer-controller"
repository = "https://aws.github.io/eks-charts"
chart = "aws-load-balancer-controller"
namespace = "kube-system"
depends_on = [
kubernetes_service_account.load_balancer_service_account
]
set {
name = "serviceAccount.create"
value = "false"
}
set {
name = "serviceAccount.name"
value = "aws-load-balancer-controller"
}
set {
name = "clusterName"
value = module.eks.cluster_name
}
}
create deployment of app with at least 2 replicas on different nodes
create simplest NodePort service for app
create simplest Ingress for that service with declared spec.rules.0.host
Describe the bug
I'm trying to deploy web application with pretty standard stack, that contains deployment, service and ingress. We are using 3 worker nodes on EKS, and when I scale replicas to at least 2, Load Balancer UI shows that targets randomly goes unhealthy. Also app responds with
504 Gateway timed out
every few requests. It seems that traffic goes only to one pod/worker node, completely ignoring routing to other pods on other nodes.Worth to notice is that in ALB UI shows that statuses of targets are some kind of random. In one moment all targets are healthy, after refresh only one node is healthy. After second refresh to of them are healthy, with one unhealthy. At least one node is always healthy, as this is the worker that responds for successful requests.
Steps to reproduce
v1.30
EKS cluster with 2 or 3 worker nodesNodePort
service for appspec.rules.0.host
Expected outcome
Environment
AWS Load Balancer controller version:
v2.8.2
Kubernetes version:
v1.30
Using EKS:
eks.20
Additional Context:
Ingress annotations:
Potentially related issue: kubernetes/ingress-nginx#9990
The text was updated successfully, but these errors were encountered: