#OCPBUGS-45129 | issue | 35 hours ago | 4.16 EFS CSI driver fails on OCP 4.16 due to ports not available on nodes CLOSED |
Issue 16490620: 4.16 EFS CSI driver fails on OCP 4.16 due to ports not available on nodes Description: Description of problem: {code:none} Trying to install AWS EFS Driver 4.16 in 4.16 OCP. And driver pods get stuck with some pods in Pending or CLBO status: $oc get pods NAME READY STATUS RESTARTS AGE aws-efs-csi-driver-controller-76b86fc9c8-6xrz8 0/4 Pending 0 9d aws-efs-csi-driver-controller-97464c644-vm8bk 0/4 Pending 0 9d aws-efs-csi-driver-node-9w8d2 1/3 CrashLoopBackOff 5254 9d aws-efs-csi-driver-node-9zxnj 1/3 CrashLoopBackOff 4594 9d aws-efs-csi-driver-node-cw64z 1/3 CrashLoopBackOff 5158 9d aws-efs-csi-driver-node-qwzz7 1/3 CrashLoopBackOff 5279 9d aws-efs-csi-driver-node-sq4x5 1/3 CrashLoopBackOff 4620 9d aws-efs-csi-driver-node-x2wcs 1/3 CrashLoopBackOff 5217 9d aws-efs-csi-driver-operator-58c9bf657-96r7s 1/1 Running 0 9d $ oc get events LAST SEEN TYPE REASON OBJECT MESSAGE 4h32m Warning FailedScheduling pod/aws-efs-csi-driver-controller-76b86fc9c8-6xrz8 0/6 nodes are available: 3 node(s) didn't have free ports for the requested pod ports, 3 node(s) didn't match Pod's node affinity/selector. preemption: 0/6 nodes are available: 3 Preemption is not helpful for scheduling, 3 node(s) didn't have free ports for the requested pod ports. 4h32m Warning FailedScheduling pod/aws-efs-csi-driver-controller-97464c644-vm8bk 0/6 nodes are available: 3 node(s) didn't have free ports for the requested pod ports, 3 node(s) didn't match Pod's node affinity/selector. preemption: 0/6 nodes are available: 3 Preemption is not helpful for scheduling, 3 node(s) didn't have free ports for the requested pod ports. 30m Warning ProbeError pod/aws-efs-csi-driver-node-9w8d2 Liveness probe error: Get "http://10.204.107.72:10305/healthz": dial tcp 10.204.107.72:10305: connect: connection refused... 56s Warning BackOff pod/aws-efs-csi-driver-node-9w8d2 Back-off restarting failed container csi-driver in pod aws-efs-csi-driver-node-9w8d2_openshift-cluster-csi-drivers(638db9a7-2ad2-493e-bedb-7be54ba546d6) {code} Version-Release number of selected component (if applicable): {code:none} aws-efs-csi-driver-operator.v4.16.0-202410280205{code} How reproducible: {code:none} all the time{code} Steps to Reproduce: {code:none} 1. Install AWS EFS CSI driver 4.16 on OCP 4.16.12 2. 3. {code} Actual results: {code:none} EFS CSI driver pods in pending or CLBO status{code} Expected results: {code:none} {code} Additional info: {code:none} The sosreports from the nodes where the pods do not start shows the ports are being used by the vsphere CSI driver pods. For instance, deployment "openshift-cluster-csi-drivers/aws-efs-csi-driver-controller" has in this template: "name":"csi-driver","ports":[{"containerPort":10302,"name":"healthz","protocol":"TCP"} And the node reports that port is currently in use by the vsphere csi driver: root 4865 4823 4865 0 11 80 0 - 496699 futex_ Sep25 ? 00:01:41 /usr/bin/csi-node-driver-registrar --csi-address=/csi/csi.sock --kubelet-registration-path=/var/lib/kubelet/plugins/csi.vsphere.vmware.com/csi.sock --http-endpoint=:10302 --v=2{code} Status: CLOSED |
Found in 0.00% of runs (0.00% of failures) across 122692 total runs and 11847 jobs (23.97% failed) in 83ms - clear search | chart view - source code located on github