Job:
#OCPBUGS-45129issue35 hours ago4.16 EFS CSI driver fails on OCP 4.16 due to ports not available on nodes CLOSED
Issue 16490620: 4.16 EFS CSI driver fails on OCP 4.16 due to ports not available on nodes
Description: Description of problem:
 {code:none}
     Trying to install AWS EFS Driver 4.16 in 4.16 OCP. And driver pods get stuck with some pods in Pending or CLBO status:
 
 $oc get pods
 NAME                                             READY   STATUS             RESTARTS   AGE
 aws-efs-csi-driver-controller-76b86fc9c8-6xrz8   0/4     Pending            0          9d
 aws-efs-csi-driver-controller-97464c644-vm8bk    0/4     Pending            0          9d
 aws-efs-csi-driver-node-9w8d2                    1/3     CrashLoopBackOff   5254       9d
 aws-efs-csi-driver-node-9zxnj                    1/3     CrashLoopBackOff   4594       9d
 aws-efs-csi-driver-node-cw64z                    1/3     CrashLoopBackOff   5158       9d
 aws-efs-csi-driver-node-qwzz7                    1/3     CrashLoopBackOff   5279       9d
 aws-efs-csi-driver-node-sq4x5                    1/3     CrashLoopBackOff   4620       9d
 aws-efs-csi-driver-node-x2wcs                    1/3     CrashLoopBackOff   5217       9d
 aws-efs-csi-driver-operator-58c9bf657-96r7s      1/1     Running            0          9d
 
 $ oc get events
 LAST SEEN   TYPE      REASON             OBJECT                                               MESSAGE
 4h32m       Warning   FailedScheduling   pod/aws-efs-csi-driver-controller-76b86fc9c8-6xrz8   0/6 nodes are available: 3 node(s) didn't have free ports for the requested pod ports, 3 node(s) didn't match Pod's node affinity/selector. preemption: 0/6 nodes are available: 3 Preemption is not helpful for scheduling, 3 node(s) didn't have free ports for the requested pod ports.
 4h32m       Warning   FailedScheduling   pod/aws-efs-csi-driver-controller-97464c644-vm8bk    0/6 nodes are available: 3 node(s) didn't have free ports for the requested pod ports, 3 node(s) didn't match Pod's node affinity/selector. preemption: 0/6 nodes are available: 3 Preemption is not helpful for scheduling, 3 node(s) didn't have free ports for the requested pod ports. 30m         Warning   ProbeError         pod/aws-efs-csi-driver-node-9w8d2                    Liveness probe error: Get "http://10.204.107.72:10305/healthz": dial tcp 10.204.107.72:10305: connect: connection refused... 56s         Warning   BackOff            pod/aws-efs-csi-driver-node-9w8d2                    Back-off restarting failed container csi-driver in pod aws-efs-csi-driver-node-9w8d2_openshift-cluster-csi-drivers(638db9a7-2ad2-493e-bedb-7be54ba546d6)
 {code}
 Version-Release number of selected component (if applicable):
 {code:none}
 aws-efs-csi-driver-operator.v4.16.0-202410280205{code}
 How reproducible:
 {code:none}
     all the time{code}
 Steps to Reproduce:
 {code:none}
     1. Install AWS EFS CSI driver 4.16 on OCP 4.16.12
     2.
     3.
     {code}
 Actual results:
 {code:none}
     EFS CSI driver pods in pending or CLBO status{code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     The sosreports from the nodes where the pods do not start shows the ports are being used by the vsphere CSI driver pods. For instance, deployment "openshift-cluster-csi-drivers/aws-efs-csi-driver-controller" has in this template: "name":"csi-driver","ports":[{"containerPort":10302,"name":"healthz","protocol":"TCP"}
 
 And the node reports that port is currently in use by the vsphere csi driver:
 
 root        4865    4823    4865  0   11  80   0 - 496699 futex_ Sep25 ?       00:01:41 /usr/bin/csi-node-driver-registrar --csi-address=/csi/csi.sock --kubelet-registration-path=/var/lib/kubelet/plugins/csi.vsphere.vmware.com/csi.sock --http-endpoint=:10302 --v=2{code}
Status: CLOSED

Found in 0.00% of runs (0.00% of failures) across 122692 total runs and 11847 jobs (23.97% failed) in 83ms - clear search | chart view - source code located on github