Job:
#OCPBUGS-74135issue6 weeks ago[4.17] oauth-proxy e2e tests fail due to ambiguous container image short names on RHEL 9 POST
Issue 17781273: [4.17] oauth-proxy e2e tests fail due to ambiguous container image short names on RHEL 9
Description: This is a clone of issue OCPBUGS-73848. The following is the description of the original issue: 
 —
 The oauth-proxy e2e-component tests fail on OCP 4.17 because RHEL 9 enforces fully qualified container image names ("short name mode").
 
 The backend test image {{nginxdemos/nginx-hello:plain-text}} is rejected by Podman since it cannot determine which registry to pull from.
 h2. Version-Release number
 
 4.22
 h2. How reproducible
 
 Always
 h2. Steps to Reproduce
 
 Run the oauth-proxy e2e-component tests on OCP 4.22
 
 The test creates a pod with a backend container using image {{nginxdemos/nginx-hello:plain-text}}
 
 Observe the pod fails to start.
 h2. Actual results
 
 The backend container fails with:
 {code:java}
 ImageInspectError: short name mode is enforcing, but image name 
 nginxdemos/nginx-hello:plain-text returns ambiguous list
 {code}
 The pod never reaches Running state, and all e2e tests fail.
 h2. Expected results
 
 The e2e tests should pass. The backend container should start successfully using a fully qualified image reference.
 h2. Additional info
 
 _Root cause:_ RHEL 9 enforces "short name mode" for container images, requiring fully qualified registry paths.
 
 _Fix:_ Replace {{nginxdemos/nginx-hello:plain-text}} with {{{}registry.k8s.io/e2e-test-images/agnhost:2.53{}}}, which is:
  * Fully qualified (satisfies RHEL 9 short name policy)
  * Multi-arch (supports ARM64)
  * The standard test image adopted across OpenShift repositories ([openshift/origin PR #29672|https://github.com/openshift/origin/pull/29672])
  * Mirrored in OpenShift CI infrastructure
 
 _PR:_ [https://github.com/openshift/oauth-proxy/pull/343]
Status: POST
#OCPBUGS-29085issue6 days agoLoadBalancer ingressController not working with IPv6 and OVN-Octavia provider New
Issue 15784478: LoadBalancer ingressController not working with IPv6 and OVN-Octavia provider
Description: Description of problem:
 
 OCP4.15 running on top of OSP16.2 fails to configure an ipv6 ingress controller. The testcase that fails is: 
 
 {code}
 make openstack-tests && KUBECONFIG=~/.kube/config ./openstack-tests run-test "[sig-installer][Suite:openshift/openstack][lb][Serial] The Openstack platform should create a TCP OVN LoadBalancer when LoadBalancerService ingressController is created on Openshift"
 {code}
 
 and it should be run with below configuration applied:
 {code}
 $ oc get cm -n openshift-config cloud-provider-config -o json | jq -r .data.config 
 [Global]
 secret-name = openstack-credentials
 secret-namespace = kube-system
 region = regionOne
 ca-file = /etc/kubernetes/static-pod-resources/configmaps/cloud-config/ca-bundle.pem
 
 [LoadBalancer]
 floating-network-id = 663be8b6-09db-4be4-bb33-7fc87c0ec51a
 lb-provider = ovn
 lb-method = SOURCE_IP_PORT
 subnet-id = 648feab4-be61-4f69-a2a7-f54a7bbe5375
 create-monitor = False
 monitor-delay = 10s
 monitor-timeout = 10s
 monitor-max-retries = 1
 max-shared-lb = 2
 {code}
 
 It's observed an error in OVN-Octavia on the controller:
 
 {code}
 : 31592, 'pool_id': '1794b0ac-44c7-4450-abad-a0725bd8a84a', 'subnet_id': '648feab4-be61-4f69-a2a7-f54a7bbe5375', 'admin_state_up': True} request_handler /usr/lib/python3.6/site-packages/networking_ovn/octavia/ovn_
 driver.py:510
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver [-] Exception occurred during creation of member: netaddr.core.AddrFormatError: invalid IPNetwork [fd2e:6f44:5dd8:c956::3b4]
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver Traceback (most recent call last):
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver   File "/usr/lib/python3.6/site-packages/netaddr/strategy/ipv4.py", line 125, in str_to_int
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver     return _struct.unpack('>I', _inet_pton(AF_INET, addr))[0]
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver OSError: illegal IP address string passed to inet_pton
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver 
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver During handling of the above exception, another exception occurred:
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver 
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver Traceback (most recent call last):
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver   File "/usr/lib/python3.6/site-packages/netaddr/ip/__init__.py", line 311, in __init__
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver     self._value = self._module.str_to_int(addr, flags)
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver   File "/usr/lib/python3.6/site-packages/netaddr/strategy/ipv4.py", line 129, in str_to_int
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver     raise AddrFormatError('%r is not a valid IPv4 address string!' % addr)
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver netaddr.core.AddrFormatError: '[fd2e:6f44:5dd8:c956::3b4]' is not a valid IPv4 address string!
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver 
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver During handling of the above exception, another exception occurred:
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver 
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver Traceback (most recent call last):
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver   File "/usr/lib/python3.6/site-packages/netaddr/ip/__init__.py", line 791, in parse_ip_network
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver     ip = IPAddress(val1, module.version, flags=INET_PTON)
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver   File "/usr/lib/python3.6/site-packages/netaddr/ip/__init__.py", line 314, in __init__
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver     % (addr, self._module.version))
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver netaddr.core.AddrFormatError: base address '[fd2e:6f44:5dd8:c956::3b4]' is not IPv4
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver 
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver During handling of the above exception, another exception occurred:
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver 
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver Traceback (most recent call last):
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver   File "/usr/lib/python3.6/site-packages/netaddr/ip/__init__.py", line 928, in __init__
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver     implicit_prefix, flags)
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver   File "/usr/lib/python3.6/site-packages/netaddr/ip/__init__.py", line 795, in parse_ip_network
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver     expanded_addr = _ipv4.expand_partial_address(val1)
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver   File "/usr/lib/python3.6/site-packages/netaddr/strategy/ipv4.py", line 259, in expand_partial_address
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver     raise error
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver netaddr.core.AddrFormatError: invalid partial IPv4 address: '[fd2e:6f44:5dd8:c956::3b4]'!
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver 
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver During handling of the above exception, another exception occurred:
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver 
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver Traceback (most recent call last):
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver   File "/usr/lib/python3.6/site-packages/networking_ovn/octavia/ovn_driver.py", line 1883, in member_create
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver     self._add_member(member, ovn_lb, pool_key)
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver   File "/usr/lib/python3.6/site-packages/networking_ovn/octavia/ovn_driver.py", line 1841, in _add_member
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver     self._refresh_lb_vips(ovn_lb.uuid, external_ids)
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver   File "/usr/lib/python3.6/site-packages/networking_ovn/octavia/ovn_driver.py", line 1051, in _refresh_lb_vips
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver     vip_ips = self._frame_vip_ips(lb_external_ids)
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver   File "/usr/lib/python3.6/site-packages/networking_ovn/octavia/ovn_driver.py", line 1039, in _frame_vip_ips
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver     if netaddr.IPNetwork(lb_vip).version == 6:
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver   File "/usr/lib/python3.6/site-packages/netaddr/ip/__init__.py", line 938, in __init__
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver     raise AddrFormatError('invalid IPNetwork %s' % addr)
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver netaddr.core.AddrFormatError: invalid IPNetwork [fd2e:6f44:5dd8:c956::3b4]
 2024-02-06 15:28:23.072 13 ERROR networking_ovn.octavia.ovn_driver
 {code}
     
 However, this behaviour is only observed by the LB created through the openshift-ingress. Other tc that is creating a loadbalancer serviceType in ipv6 it's working fine, for example:
 
 {code}
 $ make openstack-tests && KUBECONFIG=~/.kube/config ./openstack-tests run-test "[sig-installer][Suite:openshift/openstack][lb][Serial] The Openstack platform should create a UDP OVN LoadBalancer when a UDP svc with type:LoadBalancer is created on Openshift"
 make: 'openstack-tests' is up to date.
   Feb  6 09:29:12.460: INFO: Enabling in-tree volume drivers
   Feb  6 09:29:12.954: INFO: configPath is now "/tmp/configfile48504311"
   Feb  6 09:29:12.954: INFO: The user is now "e2e-test-openstack-shwmj-user"
   Feb  6 09:29:12.954: INFO: Creating project "e2e-test-openstack-shwmj"
   Feb  6 09:29:13.275: INFO: Waiting on permissions in project "e2e-test-openstack-shwmj" ...
   Feb  6 09:29:13.504: INFO: Waiting for ServiceAccount "default" to be provisioned...
   Feb  6 09:29:13.615: INFO: Waiting for ServiceAccount "deployer" to be provisioned...
   Feb  6 09:29:13.736: INFO: Waiting for ServiceAccount "builder" to be provisioned...
   Feb  6 09:29:13.851: INFO: Waiting for RoleBinding "system:image-pullers" to be provisioned...
   Feb  6 09:29:13.864: INFO: Waiting for RoleBinding "system:image-builders" to be provisioned...
   Feb  6 09:29:13.945: INFO: Waiting for RoleBinding "system:deployers" to be provisioned...
   Feb  6 09:29:14.366: INFO: Project "e2e-test-openstack-shwmj" has been fully provisioned.
   Feb  6 09:29:15.054: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:0, Replicas:0, UpdatedReplicas:0, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition(nil), CollisionCount:(*int32)(nil)}
   Feb  6 09:29:17.064: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:2, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:2, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2024, time.February, 6, 9, 29, 15, 0, time.Local), LastTransitionTime:time.Date(2024, time.February, 6, 9, 29, 15, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2024, time.February, 6, 9, 29, 15, 0, time.Local), LastTransitionTime:time.Date(2024, time.February, 6,
 9, 29, 15, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"lb-default-dep-749c7dd84d\" is progressing."}}, CollisionCount:(*int32)(nil)}                                                                                   
   Feb  6 09:29:19.069: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:2, UpdatedReplicas:2, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:2, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2024, time.February, 6, 9, 29, 15, 0, time.Local), LastTransitionTime:time.Date(2024, time.February, 6, 9, 29, 15, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2024, time.February, 6, 9, 29, 15, 0, time.Local), LastTransitionTime:time.Date(2024, time.February, 6,
 9, 29, 15, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"lb-default-dep-749c7dd84d\" is progressing."}}, CollisionCount:(*int32)(nil)}                                                                                   
   Feb  6 09:29:21.245: INFO: Waiting up to 10m0s for service "lb-default-svc" to have a LoadBalancer
   Feb  6 09:29:31.554: INFO: Waiting for 6 members matching status NO_MONITOR|ONLINE in the pool pool_0_kube_service_kubernetes_e2e-test-openstack-shwmj_lb-default-svc                                                                      
   Feb  6 09:29:47.352: INFO: Found expected number of NO_MONITOR|ONLINE members, checking again 5 seconds later to confirm that it is stable                                                                                                 
   Feb  6 09:29:52.872: INFO: Wait completed, NO_MONITOR|ONLINE members found: ["ostest-qvxqg-master-2" "ostest-qvxqg-master-1" "ostest-qvxqg-worker-0-fnk9m" "ostest-qvxqg-worker-0-njvpg" "ostest-qvxqg-master-0" "ostest-qvxqg-worker-0-lmv7g"]
   Feb  6 09:29:53.240: INFO: Pods accessed after 100 requests:
   map[lb-default-dep-749c7dd84d-6j6k8:48 lb-default-dep-749c7dd84d-kjvnn:52]
 
   Feb  6 09:29:53.258: INFO: Checking if pod "lb-default-dep-749c7dd84d-6j6k8" has been accessed at least 30 number of times...                                                                                                              
   Feb  6 09:29:53.258: INFO: Checking if pod "lb-default-dep-749c7dd84d-kjvnn" has been accessed at least 30 number of times...                                                                                                              
   Feb  6 09:29:53.281: INFO: Deleted {user.openshift.io/v1, Resource=users  e2e-test-openstack-shwmj-user}, err: <nil>
   Feb  6 09:29:53.331: INFO: Deleted {oauth.openshift.io/v1, Resource=oauthclients  e2e-client-e2e-test-openstack-shwmj}, err: <nil>                                                                                                         
 {code}
Status: New
#OCPBUGS-57501issue4 days agoAfter the Cluster upgrade to 4.16.38 some of the iptables-alerter pod are in Create Container Error State - [runc]. ASSIGNED
Issue 17067136: After the Cluster upgrade to 4.16.38 some of the iptables-alerter pod are in Create Container Error State - [runc].
Description: Description of problem:
 {code:none}
  --> After the Cluster upgrade to 4.16.38 some of the iptables-alerter pod are in Create Container Error State as shown:
 pod/iptables-alerter-98mdm              0/1     CreateContainerError   0          23h
 pod/iptables-alerter-bgzl7              1/1     Running                1          11d
 pod/iptables-alerter-mpswn              0/1     CreateContainerError   0          5d
 
 --> On seeing the pod yaml, the pods are only showing ContextDeadlineExceeded as shown:
 containerStatuses:
   - image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3c3d0b212909bc8049aa6fb9006c0990b664dd0ca5089da224e08ba909ce0a0b
     imageID: ""
     lastState: {}
     name: iptables-alerter
     ready: false
     restartCount: 0
     started: false
     state:
       waiting:
         message: context deadline exceeded
         reason: CreateContainerError
 
 --> Similar error can be seen in the events as well:
 
 LAST SEEN   TYPE      REASON   OBJECT                       MESSAGE
 5m          Warning   Failed   pod/iptables-alerter-98mdm   Error: context deadline exceeded
 3m7s        Warning   Failed   pod/iptables-alerter-mpswn   Error: context deadline exceeded
 
 --> Manually deleted the iptables-alerter pods, but no luck.
 --> Manually restarted kubelet & crio service on the nodes, but no luck.
 --> Pruned the nodes of the exited containers, no luck.
 --> Hard & Soft reboots of the worker node from the VMware console, but no luck.
 --> Also after wiping the CRIO for the node the issue persists.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
 4.16.38    {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: ASSIGNED
#OCPBUGS-50994issue7 weeks agoGCP: Install failures due to inconsistent iam apply Verified
Issue 16677256: GCP: Install failures due to inconsistent iam apply
Description: *Description of problem*. In Advanced Cluster Security we rely for OCP creation in our CI and recently observed [an increase of cluster creation failures|https://docs.google.com/document/d/1vE-kLPWooO75BXxGub6YZpM4NZolKBcOrvJULZ5NBw0/edit]. While we've been [advised to retry the failures|https://redhat-internal.slack.com/archives/C68TNFWA2/p1722873956362609] (and we do so now, see [ROX-25416|https://issues.redhat.com/browse/ROX-25416]), I'm afraid our use case is not so unique and others are affected as well.
 
 We suggest upgrading terraform and provider to the latest version (possible before license changes) in openshift-installer for 4.12+. The underlying issue is [probably already fixed upstream|https://github.com/hashicorp/terraform-provider-google/issues/17332] and released in [v5.37.0|https://github.com/hashicorp/terraform-provider-google/releases/tag/v5.37.0].
 
 Version-Release number of selected component (if applicable): TBD
 
 How reproducible: TBD
 
 Steps to Reproduce: TBD
 
 Actual results: TBD
 
 Expected results: TBD
 
 h3.  Additional info.
 
 The most common error we see in our [JIRA issues|https://issues.redhat.com/browse/ROX-25724?jql=text%20~%20%22module.master.google_service_account.master-node-sa%22] is and that is something we could find similar issues with AWS provider too eg. OCPBUGS-4213.
 {code}
 level=error msg=Error: Provider produced inconsistent result after apply .... resource was present, but now absent
 {code}
 
 Summary of errors from:
 {code}
       3 failed to create cluster: failed to apply Terraform: error(GCPComputeBackendTimeout) from Infrastructure Provider: GCP is experiencing backend service interuptions, the compute instance failed to create in reasonable time.&#34;
       3 Provider produced inconsistent result after apply\n\nWhen applying changes to\nmodule.master.google_service_account.master-node-sa[0], provider\n\&#34;provider[\\\&#34;openshift/local/google\\\&#34;]\&#34; produced an unexpected new value: Root\nresource was present, but now absent.\n\n
       6 Error waiting to create Network: Error waiting for Creating Network: timeout while waiting for state to become &#39;DONE&#39; (last state: &#39;RUNNING&#39;, timeout: 4m0s)\n\n  with module.network.google_compute_network.cluster_network[0],\n  on network/network.tf line 1, in resource \&#34;google_compute_network\&#34; \&#34;cluster_network\&#34;:\n   1: resource \&#34;google_compute_network\&#34; \&#34;cluster_network\&#34; {\n\n&#34;
       9 error applying Terraform configs: failed to apply Terraform: error(GCPComputeBackendTimeout) from Infrastructure Provider: GCP is experiencing backend service interuptions, the compute instance failed to create in reasonable time.&#34;
      14 Provider produced inconsistent result after apply\n\nWhen applying changes to module.master.google_service_account.master-node-sa,\nprovider \&#34;provider[\\\&#34;openshift/local/google\\\&#34;]\&#34; produced an unexpected new\nvalue: Root resource was present, but now absent.
      16 Provider produced inconsistent result after apply\n\nWhen applying changes to google_service_account_key.bootstrap, provider\n\&#34;provider[\\\&#34;openshift/local/google\\\&#34;]\&#34; produced an unexpected new value: Root\nresource was present, but now absent.
      18 Provider produced inconsistent result after apply\n\nWhen applying changes to module.iam.google_service_account.worker-node-sa,\nprovider \&#34;provider[\\\&#34;openshift/local/google\\\&#34;]\&#34; produced an unexpected new\nvalue: Root resource was present, but now absent.
      34 Error creating service account key: googleapi: Error 404: Service account projects/acs-san-stackroxci/serviceAccounts/XXX@acs-san-stackroxci.iam.gserviceaccount.com does not exist., notFound\n\n  with google_service_account_key.bootstrap,\n  on main.tf line 38, in resource \&#34;google_service_account_key\&#34; \&#34;bootstrap\&#34;:\n  38: resource \&#34;google_service_account_key\&#34; \&#34;bootstrap\&#34; {\n\n&#34;
      45 error applying Terraform configs: failed to apply Terraform: exit status 1\n\nError: Provider produced inconsistent result after apply\n\nWhen applying changes to\nmodule.master.google_service_account.master-node-sa[0], provider\n\&#34;provider[\\\&#34;openshift/local/google\\\&#34;]\&#34; produced an unexpected new value: Root\nresource was present, but now absent.
      59 error applying Terraform configs: failed to apply Terraform: exit status 1\n\nError: Provider produced inconsistent result after apply\n\nWhen applying changes to module.iam.google_service_account.worker-node-sa,\nprovider \&#34;provider[\\\&#34;openshift/local/google\\\&#34;]\&#34; produced an unexpected new\nvalue: Root resource was present, but now absent.
     100 Provider produced inconsistent result after apply\n\nWhen applying changes to google_service_account.bootstrap-node-sa, provider\n\&#34;provider[\\\&#34;openshift/local/google\\\&#34;]\&#34; produced an unexpected new value: Root\nresource was present, but now absent.
     103 Provider produced inconsistent result after apply\n\nWhen applying changes to module.iam.google_service_account.worker-node-sa[0],\nprovider \&#34;provider[\\\&#34;openshift/local/google\\\&#34;]\&#34; produced an unexpected new\nvalue: Root resource was present, but now absent.
     116 Provider produced inconsistent result after apply\n\nWhen applying changes to\nmodule.master.google_service_account.master-node-sa[0], provider\n\&#34;provider[\\\&#34;openshift/local/google\\\&#34;]\&#34; produced an unexpected new value: Root\nresource was present, but now absent.
 {code}
 
 The openshift installer contains a bundled terraform and google-provider
 * 4.12 
 ** Terraform 1.0.11 (November 10, 2021)
 ** https://github.com/openshift/installer/blob/release-4.12/terraform/terraform/go.mod#L5
 ** Provider: v4.5.0 (Dec 20, 2021)
 ** https://github.com/openshift/installer/blob/release-4.12/terraform/providers/google/go.mod#L5
 * 4.14+
 ** Terraform: 1.3.7 (January 04, 2023)
 ** https://github.com/openshift/installer/blob/release-4.14/terraform/terraform/go.mod#L5
 ** Provider: v4.74.0 (Jul 18, 2023)
 ** https://github.com/openshift/installer/blob/release-4.14/terraform/providers/google/go.mod#L5
 * 4.18+
 ** Does not contain google provider at all: https://github.com/openshift/installer/pull/8723
Status: Verified
#OCPBUGS-72360issue7 weeks agoGinkgo test failure: [Apiserver] MicroShift enable SCC admission for pods Verified
Issue 17741919: Ginkgo test failure: [Apiserver] MicroShift enable SCC admission for pods
Description: Description of problem:
 {code:none}
 "[sig-api-machinery] API_Server on Microshift [It] Author:rgangwar-MicroShiftBoth-ConnectedOnly-Medium-55394-[Apiserver] MicroShift enable SCC admission for pods test" failed because the pod is created with a missing fsGroup
 {code}
 Version-Release number of selected component (if applicable):
 {code:none}
     4.21.0-rc.1 on x86. This is not reproducible on ARM architecture.{code}
 How reproducible:
 {code:none}
     always on 4.21.0-rc.1{code}
 Steps to Reproduce:
 {code:none}
     1. Run Ginkgo test case: "[sig-api-machinery] API_Server on Microshift [It] Author:rgangwar-MicroShiftBoth-ConnectedOnly-Medium-55394-[Apiserver] MicroShift enable SCC admission for pods test"
     2.
     3.
     {code}
 Actual results:
 {code:none}
  • [PANICKED] [222.518 seconds]
   [sig-api-machinery] API_Server on Microshift [It] Author:rgangwar-MicroShiftBoth-ConnectedOnly-Medium-55394-[Apiserver] MicroShift enable SCC admission for pods
   /tmp/tmp.I4ekJytyWF/openshift-tests-private/test/extended/apiserverauth/apiserver_microshift.go:480
 
     Timeline >>
     STEP: Creating a kubernetes client @ 01/08/26 04:07:05.353
     I0108 04:07:05.368507 507110 client.go:486] Creating namespace "e2e-test-default-wb9lb"
     I0108 04:07:05.380389 507110 client.go:491] Waiting for ServiceAccount "default" to be provisioned...
     I0108 04:07:05.485060 507110 client.go:495] Configuring kubeconfig with user "e2e-test-default-wb9lb-user" certificates...
     I0108 04:07:07.078970 507110 client.go:293] configPath is now "/tmp/configfile4051781145"
     I0108 04:07:07.079012 507110 client.go:498] Waiting for RoleBinding "e2e-test-default-wb9lb-user" to be provisioned...
     I0108 04:07:07.084506 507110 client.go:502] Waiting on permissions in namespace "e2e-test-default-wb9lb" ...
     I0108 04:07:07.364467 507110 client.go:521] Namespace "e2e-test-default-wb9lb" has been fully provisioned.
     I0108 04:07:07.364710 507110 apiserver_microshift.go:37] The cluster should be healthy before running case.
     namespace/r6c171tq created
     namespace "r6c171tq" deleted
     I0108 04:07:13.076170 507110 apiserver_util.go:1892] Cluster sanity check passed
     STEP: 1.Create temporary namespace @ 01/08/26 04:07:13.076
     STEP: 2. Create one pod security-context-demo-ocp55394 with the specified security context. @ 01/08/26 04:07:13.183
     I0108 04:07:13.183100 507110 apiserver_util.go:460] apiserver fixture dir is not initialized, start to create
     I0108 04:07:13.187939 507110 apiserver_util.go:463] apiserver fixture dir is initialized: /tmp/fixture-testdata-dir2892104936/test/extended/testdata/apiserverauth
     pod/security-context-demo-ocp55394 created
     STEP: 3. Verify that the Pod's security-context-demo-ocp55394 Container is running @ 01/08/26 04:07:13.306
     I0108 04:07:23.409925 507110 pods.go:176] Pod security-context-demo-ocp55394 is ready!
     STEP: 4.1 Verify if the processes are running with the specified user ID 1000 in the pod. @ 01/08/26 04:07:23.41
     I0108 04:07:38.587390 507110 apiserver_util.go:801] Attempting to execute command on pod security-context-demo-ocp55394. Output: PID   USER     TIME  COMMAND
         1 1000      0:00 sleep 1h
         2 1000      0:00 ps, Error: <nil>
     I0108 04:07:38.587444 507110 apiserver_util.go:832] Successfully retrieved non-empty output from pod security-context-demo-ocp55394: PID   USER     TIME  COMMAND
         1 1000      0:00 sleep 1h
         2 1000      0:00 ps
     I0108 04:07:38.587500 507110 apiserver_microshift.go:506] Processes are running on pod security-context-demo-ocp55394 with user id 1000 :: PID   USER     TIME  COMMAND
         1 1000      0:00 sleep 1h
         2 1000      0:00 ps
     STEP: 4.2 Verify that user is running with specified uid=1000 gid=3000 groups=2000 @ 01/08/26 04:07:38.587
     I0108 04:07:53.775700 507110 apiserver_util.go:801] Attempting to execute command on pod security-context-demo-ocp55394. Output: uid=1000(1000) gid=3000 groups=2000,3000, Error: <nil>
     I0108 04:07:53.775738 507110 apiserver_util.go:832] Successfully retrieved non-empty output from pod security-context-demo-ocp55394: uid=1000(1000) gid=3000 groups=2000,3000
     I0108 04:07:53.775826 507110 apiserver_microshift.go:515] On pod security-context-demo-ocp55394 User running with :: uid=1000(1000) gid=3000 groups=2000,3000
     STEP: 5. Create one pod security-context-demo-2-ocp55394 with the specified security context. @ 01/08/26 04:07:53.775
     I0108 04:07:53.775904 507110 apiserver_util.go:466] apiserver fixture dir found in cache: /tmp/fixture-testdata-dir2892104936/test/extended/testdata/apiserverauth
     pod/security-context-demo-2-ocp55394 created
     STEP: 6. Verify that the Pod's security-context-demo-2-ocp55394 Container is running @ 01/08/26 04:07:53.967
     I0108 04:08:04.065851 507110 pods.go:176] Pod security-context-demo-2-ocp55394 is ready!
     STEP: 7. Verify that processes are running with the specified user ID 2000 in the pod. @ 01/08/26 04:08:04.066
     I0108 04:08:19.255042 507110 apiserver_util.go:801] Attempting to execute command on pod security-context-demo-2-ocp55394. Output: USER         PID %CPU %MEM    VSZ   RSS TTY      STAT START   TIME COMMAND
     2000           1  0.0  0.0 705960   768 ?        Ssl  04:07   0:00 /hello_openshift
     2000           6  0.0  0.0   1704   768 ?        Rs   04:08   0:00 ps aux, Error: <nil>
     I0108 04:08:19.255086 507110 apiserver_util.go:832] Successfully retrieved non-empty output from pod security-context-demo-2-ocp55394: USER         PID %CPU %MEM    VSZ   RSS TTY      STAT START   TIME COMMAND
     2000           1  0.0  0.0 705960   768 ?        Ssl  04:07   0:00 /hello_openshift
     2000           6  0.0  0.0   1704   768 ?        Rs   04:08   0:00 ps aux
     I0108 04:08:19.255128 507110 apiserver_microshift.go:534] Processes are running on pod security-context-demo-2-ocp55394 with user id 2000 :: USER         PID %CPU %MEM    VSZ   RSS TTY      STAT START   TIME COMMAND
     2000           1  0.0  0.0 705960   768 ?        Ssl  04:07   0:00 /hello_openshift
     2000           6  0.0  0.0   1704   768 ?        Rs   04:08   0:00 ps aux
     STEP: 8. Ensure that pods in different namespaces are launched with different UIDs. @ 01/08/26 04:08:19.255
     STEP: 8.1 Create two different namespaces. @ 01/08/26 04:08:19.255
     I0108 04:08:19.291544 507110 apiserver_util.go:466] apiserver fixture dir found in cache: /tmp/fixture-testdata-dir2892104936/test/extended/testdata/apiserverauth
     STEP: 8.2 Create pods in both namespaces @ 01/08/26 04:08:19.291
     STEP: 8.3 Verify pods should have different UID's and desc in both namespaces. @ 01/08/26 04:08:19.291
     pod/pod-for-ping created
     I0108 04:08:25.585028 507110 apiserver_util.go:1035] The resource returned:
     pod-for-ping
     I0108 04:08:25.585071 507110 apiserver_util.go:1263] Namespace testpod-namespace-1 pods are: pod-for-ping
     I0108 04:08:35.699575 507110 pods.go:176] Pod pod-for-ping is ready!
     I0108 04:08:41.978615 507110 apiserver_util.go:1035] The resource returned:
     'pod-for-ping runAsUser: 1000580000 fsGroup:  seLinuxOptions: s0:c24,c14
     '
     pod/pod-for-ping created
     I0108 04:08:48.301631 507110 apiserver_util.go:1035] The resource returned:
     pod-for-ping
     I0108 04:08:48.301670 507110 apiserver_util.go:1263] Namespace testpod-namespace-2 pods are: pod-for-ping
     I0108 04:08:58.471873 507110 pods.go:176] Pod pod-for-ping is ready!
     I0108 04:09:06.403000 507110 apiserver_util.go:1035] The resource returned:
     'pod-for-ping runAsUser: 1000590000 fsGroup:  seLinuxOptions: s0:c24,c19
     '
     pod "security-context-demo-2-ocp55394" deleted from test-scc-ocp55394 namespace
     pod "security-context-demo-ocp55394" deleted from test-scc-ocp55394 namespace
     namespace "test-scc-ocp55394" deleted
     [PANICKED] in [It] - /home/ec2-user/go/pkg/mod/golang.org/toolchain@v0.0.1-go1.24.11.linux-amd64/src/runtime/panic.go:115 @ 01/08/26 04:10:47.819
     Jan  8 04:10:47.819: INFO: test dir /tmp/-OCP-microshift-apiseerver-cases-7zer0254/ is cleaned up
     STEP: Collecting events from namespace "e2e-test-default-wb9lb". @ 01/08/26 04:10:47.819
     STEP: Found 0 events. @ 01/08/26 04:10:47.83
     I0108 04:10:47.833385 507110 resource.go:151] POD  NODE  PHASE  GRACE  CONDITIONS
     I0108 04:10:47.833421 507110 resource.go:161] 
     I0108 04:10:47.837817 507110 dump.go:81] skipping dumping cluster info - cluster too large
     I0108 04:10:47.857338 507110 client.go:681] Deleted {certificates.k8s.io/v1, Resource=certificatesigningrequests  e2e-test-default-wb9lb-user}, err: <nil>
     STEP: Destroying namespace "e2e-test-default-wb9lb" for this suite. @ 01/08/26 04:10:47.857
     << Timeline
 
     [PANICKED] Test Panicked
     In [It] at: /home/ec2-user/go/pkg/mod/golang.org/toolchain@v0.0.1-go1.24.11.linux-amd64/src/runtime/panic.go:115 @ 01/08/26 04:10:47.819
 
     runtime error: index out of range [6] with length 6
 
     Full Stack Trace
       github.com/openshift/openshift-tests-private/test/extended/apiserverauth.init.func3.8()
       	/tmp/tmp.I4ekJytyWF/openshift-tests-private/test/extended/apiserverauth/apiserver_microshift.go:590 +0x1bcf
       github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?})
       	/home/ec2-user/go/pkg/mod/github.com/openshift/onsi-ginkgo/v2@v2.6.1-0.20240806135314-3946b2b7b2a8/internal/node.go:472 +0x13
       github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3()
       	/home/ec2-user/go/pkg/mod/github.com/openshift/onsi-ginkgo/v2@v2.6.1-0.20240806135314-3946b2b7b2a8/internal/suite.go:901 +0x7b
       created by github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 1
       	/home/ec2-user/go/pkg/mod/github.com/openshift/onsi-ginkgo/v2@v2.6.1-0.20240806135314-3946b2b7b2a8/internal/suite.go:888 +0xd7b
   ------------------------------
 
   Summarizing 1 Failure:
     [PANICKED!] [sig-api-machinery] API_Server on Microshift [It] Author:rgangwar-MicroShiftBoth-ConnectedOnly-Medium-55394-[Apiserver] MicroShift enable SCC admission for pods
     /home/ec2-user/go/pkg/mod/golang.org/toolchain@v0.0.1-go1.24.11.linux-amd64/src/runtime/panic.go:115
 
   Ran 1 of 1 Specs in 222.518 seconds
   FAIL! -- 0 Passed | 1 Failed | 0 Pending | 0 Skipped{code}
 Expected results:
 {code:none}
     test pass{code}
 Additional info: 
 
 [     link to ginkgo log|https://gcsweb-ci.apps.ci.l2s4.p1.openshiftapps.com/gcs/test-platform-results/logs/periodic-ci-openshift-microshift-release-4.21-periodics-e2e-aws-tests-release-periodic/2009098095907835904/artifacts/e2e-aws-tests-release-periodic/openshift-microshift-e2e-metal-tests/artifacts/scenario-info/el96-lrel@ginkgo-tests/ginkgo-results/test-output.log]
 {code:none}
     {code}
Status: Verified
#OCPBUGS-77196issue2 weeks ago[openshift-4.16] ose-libvirt-machine-controllers can't be build hermetically New
Issue 17925861: [openshift-4.16] ose-libvirt-machine-controllers can't be build hermetically
Description: Description of problem:
 
 ART can't build ose-libvirt-machine-controllers hermetically.
 The error is the following in the prefetch-dependencies konflux task:
 {code:none}
     Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.  Please try running `go mod vendor` and committing the changes.  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
  
 
  
 
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-18387issue8 weeks ago[CORS-2550] Installer should have pre-check for vm type, DES encryption type when install with Confidential VM ASSIGNED
Issue 15452967: [CORS-2550] Installer should have pre-check for vm type, DES encryption type when install with Confidential VM
Description: Description of problem:
 {code:none}
 Install IPI cluster with confidential VM, installer should have pre-check for vm type, disk encryption type etc to avoid installation failed during infrastructure creation
 
 1. vm type
 Different security type support on different vm type
 for example, set platfrom.azure.defaultMachinePlatform.type to Standard_DC8ads_v5 and platform.azure.defaultMachinePlatform.settings.securityType to TrustedLaunch, installation will be failed as Standard_DC8ads_v5 only support security type ConfidentialVM
 
 ERROR Error: creating Linux Virtual Machine: (Name "jimaconf1-89qmp-bootstrap" / Resource Group "jimaconf1-89qmp-rg"): compute.VirtualMachinesClient#CreateOrUpdate: Failure sending request: StatusCode=400 -- Original Error: Code="BadRequest" Message="The VM size 'Standard_DC16ads_v5' is not supported for creation of VMs and Virtual Machine Scale Set with 'TrustedLaunch' security type." 
 
 2. Disk encryption Set
 When install cluster with ConfidentialVM +securityEncryptionType:DiskWithVMGuestState, then using customer-managed key, it requires that DES encryption type is ConfidentialVmEncryptedWithCustomerKey, else installer throw error as below:
 
 08-31 10:12:54.443  level=error msg=Error: creating Linux Virtual Machine: (Name "jima30confa-vtrm2-bootstrap" / Resource Group "jima30confa-vtrm2-rg"): compute.VirtualMachinesClient#CreateOrUpdate: Failure sending request: StatusCode=400 -- Original Error: Code="BadRequest" Message="The type of the Disk Encryption Set in the request is 'ConfidentialVmEncryptedWithCustomerKey', but this Disk Encryption Set was created with type 'EncryptionAtRestWithCustomerKey'." Target="/subscriptions/53b8f551-f0fc-4bea-8cba-6d1fefd54c8a/resourceGroups/jima30confa-vtrm2-rg/providers/Microsoft.Compute/disks/jima30confa-vtrm2-bootstrap_OSDisk"
 
 Installer should check vm type and DES's encryption type to make sure that expected DES is set.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
 4.14 nightly build{code}
 How reproducible:
 {code:none}
 Always{code}
 Steps to Reproduce:
 {code:none}
 1. Prepare install-config, 
    1) enable confidentialVM but use vm type which does not support Confidential VM
    2) enable TrustedLaunch but use vm type which support confidentialVM
    3) enable confidentialVM + securityEncryptionType: DiskWithVMGuestState, use customer-managed  key to encrypt managed key, but customer-managed key's encryption type is the default one "EncryptionAtRestWithPlatformKey"
 2. Create cluster
 3.
 {code}
 Actual results:
 {code:none}
 Installation failed when creating infrastructure{code}
 Expected results:
 {code:none}
 Installer should have pre-check for those scenarios, and exit with expected error message.{code}
 Additional info:
 {code:none}
  {code}
Status: ASSIGNED
#OCPBUGS-60159issue5 weeks agoopenshift-cluster-csi-drivers vmware-vsphere-csi-driver-operator-metrics endpoint 8445 port should return 401 instead 500 Verified
Issue 17201621: openshift-cluster-csi-drivers vmware-vsphere-csi-driver-operator-metrics endpoint 8445 port should return 401 instead 500
Description: Description of problem:
 
 see similar bug OCPBUGS-57585
 
 Stole the testing result from [https://github.com/openshift/origin/pull/30014#issuecomment-3153130188]
 
 8445 port is for openshift-cluster-csi-drivers vmware-vsphere-csi-driver-operator-metrics endpoint, did not see this port on other IAAS(example AWS,OSP)checked on vsphere, curl 8445 without auth returns 500(actually it's forbidden),
 {noformat}
 $ oc get infrastructure/cluster -o jsonpath={.spec.platformSpec.type}
 VSphere
 
 $ oc -n openshift-cluster-csi-drivers get ep
 Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice
 NAME                                           ENDPOINTS                                                        AGE
 vmware-vsphere-csi-driver-controller-metrics   10.129.0.17:9203,10.130.0.21:9203,10.129.0.17:9202 + 9 more...   28m
 vmware-vsphere-csi-driver-operator-metrics     10.130.0.17:8445                                                 28m
 vmware-vsphere-csi-driver-webhook-svc          10.128.0.45:8443,10.130.0.23:8443                                28m
 
 $ oc -n openshift-monitoring exec -c prometheus prometheus-k8s-0 -- curl -k -s -o /dev/null -w '%{http_code}' https://10.130.0.17:8445/metrics
 500
 {noformat}
 return code is 500, expecting 401 forbidden instead of 500.
 {code:java}
 $ oc -n openshift-monitoring exec -c prometheus prometheus-k8s-0 -- curl -k https://10.130.0.17:8445/metrics
 
 Internal Server Error: "/metrics": subjectaccessreviews.authorization.k8s.io is forbidden: User "system:serviceaccount:openshift-cluster-csi-drivers:vmware-vsphere-csi-driver-operator" cannot create resource "subjectaccessreviews" in API group "authorization.k8s.io" at the cluster scope  {code}
 curl with auth, no issue
 {noformat}
 $ token=`oc create token prometheus-k8s -n openshift-monitoring`
 $ oc -n openshift-monitoring exec -c prometheus prometheus-k8s-0 -- curl -k -H "Authorization: Bearer $token" 'https://10.130.0.17:8445/metrics' | head -n3
 # HELP aggregator_discovery_aggregation_count_total [ALPHA] Counter of number of times discovery was aggregated
 # TYPE aggregator_discovery_aggregation_count_total counter
 aggregator_discovery_aggregation_count_total 0 {noformat}
 [https://github.com/openshift/origin/pull/30014#issuecomment-3154900379]
 
  
 
 When the bug is fixed, please remove it from this code snippet to avoid regression.
 
 [https://github.com/openshift/origin/blob/4f183dd3427cffd8d97b44557caa782d65726416/test/extended/prometheus/prometheus.go#L64-L71]
Status: Verified
#OCPBUGS-77198issue2 weeks ago[openshift-4.14] ose-libvirt-machine-controllers can't be build hermetically New
Issue 17925864: [openshift-4.14] ose-libvirt-machine-controllers can't be build hermetically
Description: Description of problem:
 
 ART can't build ose-libvirt-machine-controllers hermetically.
 The error is the following in the prefetch-dependencies konflux task:
 {code:none}
     Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.  Please try running `go mod vendor` and committing the changes.  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
  
 
  
 
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-77786issue6 days ago[openshift-4.17] ose-sriov-rdma-cni can't be built hermetically New
Issue 17955337: [openshift-4.17] ose-sriov-rdma-cni can't be built hermetically
Description: This is a clone of issue OCPBUGS-77775. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-sriov-rdma-cni hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-77211issue4 days ago[openshift-4.15] hermetic build failure - ose-ibm-vpc-block-csi-driver Verified
Issue 17926179: [openshift-4.15] hermetic build failure - ose-ibm-vpc-block-csi-driver
Description: This is a clone of issue OCPBUGS-77210. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77209. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77207. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77206. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77205. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77203. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77167. The following is the description of the original issue:  --- Description of problem:
 {code:none}
 IBM VPC Block CSI Driver build fails in hermetic mode due to non-deterministic GOPATH handling in Makefile. The build system rewrites $GOPATH causing the binary output path -o ${GOPATH}/bin/${EXE_DRIVER_NAME} to become invalid, resulting in build failures.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
 OpenShift 4.19+ (affects hermetic build conversion) {code}
 How reproducible:
 {code:none}
 Always - when building ibm-vpc-block-csi-driver in hermetic mode{code}
 Steps to Reproduce:
 {code:none}
   1. Build ibm-vpc-block-csi-driver container in hermetic mode
   2. Makefile uses absolute GOPATH-dependent output path: -o ${GOPATH}/bin/${EXE_DRIVER_NAME}
   3. Build system rewrites $GOPATH during hermetic build process     {code}
 Actual results:
 {code:none}
 Build fails because binary cannot be found at expected GOPATH location after build system modifies $GOPATH
 
 Error: building at STEP "COPY --from=builder /cachi2/output/deps/gomod/bin/ibm-vpc-block-csi-driver /bin/ibm-vpc-block-csi-driver": checking on sources under [...]: copier: stat: "/cachi2/output/deps/gomod/bin/ibm-vpc-block-csi-driver": no such file or directory{code}
 Expected results:
 {code:none}
 Build succeeds using relative path bin/${EXE_DRIVER_NAME} that doesn't depend on GOPATH manipulation, consistent with other CSI drivers like aws-ebs-csi-driver  {code}
 Additional info:
 {code:none}
 Fix implemented: Change Makefile to use relative path -o bin/${EXE_DRIVER_NAME} instead of absolute GOPATH-dependent path. This aligns with other CSI drivers and makes builds hermetic-compatible. Comment added to indicate this is a permanent carry patch for OpenShift builds.{code}
Status: Verified
#OCPBUGS-77854issue5 days ago[openshift-4.17] multus-cni can't be built hermetically New
Issue 17959379: [openshift-4.17] multus-cni can't be built hermetically
Description: This is a clone of issue OCPBUGS-77853. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build multus-cni hermetically. Same applies for multus-cni-container-microshift
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.    {code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-70155issue6 weeks agoiDRAC 10 BIOS setting succeeded but with servicing error New
Issue 17714675: iDRAC 10 BIOS setting succeeded but with servicing error
Description: Description of problem:
 
 Though iDRAC 10 bios settings update using HFS succeeded, oc describe shows servicing error. in addition, when node returned to Ready state, servicing error disappeared
 
 Normal PowerOff 11m metal3-baremetal-controller Host soft powered off Normal ServicingError 3m18s metal3-baremetal-controller Node 08387329-dc15-4cef-932b-cc2cb323738d failed step {'args': {'settings': [
 
 {'name': 'XptPrefetch', 'value': 'Disabled'}
 
 ]}, 'interface': 'bios', 'step': 'apply_configuration', 'abortable': False, 'priority': 0}: Redfish exception occurred. Error: In system 4c4c4544-0053-3410-804e-c8c04f4b3434 for node 08387329-dc15-4cef-932b-cc2cb323738d all managers failed: set virtual boot device. Errors: ['Manager 34344b4f-c0c8-4e80-3410-00534c4c4544: Resource https://10.26.16.132/redfish/v1/TaskService/TaskMonitors/JID_663271577404 not found'] Normal ServicingError 108s metal3-baremetal-controller Node 08387329-dc15-4cef-932b-cc2cb323738d failed step {'args': {'settings': [
 
 {'name': 'XptPrefetch', 'value': 'Disabled'}
 
 ]}, 'interface': 'bios', 'step': 'apply_configuration', 'abortable': False, 'priority': 0}: Redfish exception occurred. Error: In system 4c4c4544-0053-3410-804e-c8c04f4b3434 for node 08387329-dc15-4cef-932b-cc2cb323738d all managers failed: set virtual boot device. Errors: ['Manager 34344b4f-c0c8-4e80-3410-00534c4c4544: Resource https://10.26.16.132/redfish/v1/TaskService/TaskMonitors/JID_663271577404 not found']  
 
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1. Perform bios settings change on IDRAC 10 server using HFS
     2. Check "oc describe bmh  -n openshift-machine-api <host>" command output
     3.
     {code}
 Actual results:
 {code:none}
     passed with servicing errors{code}
 Expected results:
 {code:none}
     To pass without servicing errors{code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-73850issue7 weeks agooauth-proxy e2e tests fail due to ambiguous container image short names on RHEL 9 New
Issue 17769399: oauth-proxy e2e tests fail due to ambiguous container image short names on RHEL 9
Description: The oauth-proxy e2e-component tests fail on OCP 4.20 because RHEL 9 enforces fully qualified container image names ("short name mode"). The backend test image {{nginxdemos/nginx-hello:plain-text}} is rejected by Podman since it cannot determine which registry to pull from.
 h2. Version-Release number
 
 4.20
 h2. How reproducible
 
 Always
 h2. Steps to Reproduce
 
 Run the oauth-proxy e2e-component tests on OCP 4.20
 
 The test creates a pod with a backend container using image {{nginxdemos/nginx-hello:plain-text}}
 
 Observe the pod fails to start
 h2. Actual results
 
 The backend container fails with:
 {code:java}
 ImageInspectError: short name mode is enforcing, but image name 
 nginxdemos/nginx-hello:plain-text returns ambiguous list
 {code}
 The pod never reaches Running state, and all e2e tests fail.
 h2. Expected results
 
 The e2e tests should pass. The backend container should start successfully using a fully qualified image reference.
 h2. Additional info
 
 _Root cause:_ RHEL 9 enforces "short name mode" for container images, requiring fully qualified registry paths.
 
 _Fix:_ Replace {{nginxdemos/nginx-hello:plain-text}} with {{{}registry.k8s.io/e2e-test-images/agnhost:2.53{}}}.
 
 _Backport of:_ [OCPBUGS-73848|https://issues.redhat.com/browse/OCPBUGS-73848]
Status: New
#OCPBUGS-78180issue9 hours ago[openshift-4.12] ose-azure-cluster-api-controllers can't be built hermetically New
Issue 17974852: [openshift-4.12] ose-azure-cluster-api-controllers can't be built hermetically
Description: This is a clone of issue OCPBUGS-78179. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78178. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78177. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78176. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77983. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77905. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77904. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-azure-cluster-api-controllers
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-72546issue3 weeks agocgroupsv2: failed to write on cpuset.cpus.exclusive ASSIGNED
Issue 17747711: cgroupsv2: failed to write on cpuset.cpus.exclusive
Description: Description of problem:
 {code:none}
 When deploying a workload that makes use of cpuset.cpus.exclusive, it fails    {code}
 Version-Release number of selected component (if applicable):
 {code:none}
 4.18.11{code}
 How reproducible:
 {code:none}
 Happens always with this particular workload using helm chart    {code}
 Steps to Reproduce:
 {code:none}
 1.Deploy the workload by using the customer helm chart
      {code}
 Actual results:
 {code:none}
 Error: failed to run pre-start hook for container "ctr-du-vppxx": set CPU load balancing: failed to write "7-16": write /sys/fs/cgroup/kubepods.slice/kubepods-podbf0bb53f_790a_4f26_ada7_52axxxxf43c.slice/crio-ctr-du-vppxx.scope/cpuset.cpus.exclusive: invalid argument
 {code}
 Expected results:
 {code:none}
  Pods can be deployed successfully{code}
 Additional info:
 {code:none}
     {code}
Status: ASSIGNED
#OCPBUGS-78183issue9 hours ago[openshift-4.12] ose-machine-api-provider-gcp can't be built hermetically New
Issue 17974914: [openshift-4.12] ose-machine-api-provider-gcp can't be built hermetically
Description: This is a clone of issue OCPBUGS-78182. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78181. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-machine-api-provider-gcp hermetically
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-78175issue9 hours ago[openshift-4.18] ose-aws-cluster-api-controllers can't be built hermetically New
Issue 17974775: [openshift-4.18] ose-aws-cluster-api-controllers can't be built hermetically
Description: This is a clone of issue OCPBUGS-77903. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77902. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-aws-cluster-api-controllers
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.
      {code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-78110issue16 hours ago[openshift-4.12] openshift-enterprise-haproxy-router can't be built hermetically New
Issue 17972087: [openshift-4.12] openshift-enterprise-haproxy-router can't be built hermetically
Description: This is a clone of issue OCPBUGS-78109. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78108. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build openshift-enterprise-haproxy-router & ose-haproxy-router-base hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-74487issue2 weeks agoAgent service is failing with input/output error Verified
Issue 17801998: Agent service is failing with input/output error
Description: *Description of the problem:*
 
 - Unable to discover hosts using BMC because the agent service is failing to start with this error:
 
 Error: netavark: failed to start process: fork/exec /usr/libexec/podman/netavark: input/output error
 
 
 *How reproducible:* Always
 
 *Steps to reproduce:*
 
 1. Boot a system using BMC/BMH
 
 2. ssh the node and check the agent service logs
 
 *Actual results:*
 The agent service fails 
  
 
 *Expected results:*
 The Agent service should start and the hosts should get discovered.
 
 *Additional Info:*
 
 This issue is not faced when:
 
 1. Hosts are discovered using a discovery iso
 2. When we disable converged_flow
Status: Verified
#OCPBUGS-76557issue34 hours agoRHCOS10: toolbox failed to get current user's default shell New
Issue 17865189: RHCOS10: toolbox failed to get current user's default shell
Description: When running toolbox on an RHCOS 10.1 host it fails to launch with the following.
 
 
 {code:none}
 oc debug node/ip-10-0-10-196.us-east-2.compute.internal
 Starting pod/ip-10-0-10-196us-east-2computeinternal-debug-nm4qh ...
 To use host binaries, run `chroot /host`. Instead, if you need to access host namespaces, run `nsenter -a -t 1`.
 Pod IP: 10.0.10.196
 If you don't see a command prompt, try pressing enter.
 sh-5.1# chroot /host
 sh-5.2# toolbox
 Error: missing command
 
 create    Create a new Toolbx container
 enter     Enter an existing Toolbx container
 list      List all existing Toolbx containers and images
 
 Run 'toolbox --help' for usage.
 sh-5.2# toolbox enter
 Error: failed to get the current user's default shell
 sh-5.2# toolbox create
 Image required to create Toolbx container.
 Download registry.access.redhat.com/ubi10/toolbox:10.1 (336.9MB)? [y/N]: y
 Error: failed to get the current user's default shell
 {code}
Status: New
#OCPBUGS-77210issue12 hours ago[openshift-4.16] hermetic build failure - ose-ibm-vpc-block-csi-driver Verified
Issue 17926437: [openshift-4.16] hermetic build failure - ose-ibm-vpc-block-csi-driver
Description: This is a clone of issue OCPBUGS-77209. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77207. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77206. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77205. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77203. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77167. The following is the description of the original issue:  --- Description of problem:
 {code:none}
 IBM VPC Block CSI Driver build fails in hermetic mode due to non-deterministic GOPATH handling in Makefile. The build system rewrites $GOPATH causing the binary output path -o ${GOPATH}/bin/${EXE_DRIVER_NAME} to become invalid, resulting in build failures.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
 OpenShift 4.19+ (affects hermetic build conversion) {code}
 How reproducible:
 {code:none}
 Always - when building ibm-vpc-block-csi-driver in hermetic mode{code}
 Steps to Reproduce:
 {code:none}
   1. Build ibm-vpc-block-csi-driver container in hermetic mode
   2. Makefile uses absolute GOPATH-dependent output path: -o ${GOPATH}/bin/${EXE_DRIVER_NAME}
   3. Build system rewrites $GOPATH during hermetic build process     {code}
 Actual results:
 {code:none}
 Build fails because binary cannot be found at expected GOPATH location after build system modifies $GOPATH
 
 Error: building at STEP "COPY --from=builder /cachi2/output/deps/gomod/bin/ibm-vpc-block-csi-driver /bin/ibm-vpc-block-csi-driver": checking on sources under [...]: copier: stat: "/cachi2/output/deps/gomod/bin/ibm-vpc-block-csi-driver": no such file or directory{code}
 Expected results:
 {code:none}
 Build succeeds using relative path bin/${EXE_DRIVER_NAME} that doesn't depend on GOPATH manipulation, consistent with other CSI drivers like aws-ebs-csi-driver  {code}
 Additional info:
 {code:none}
 Fix implemented: Change Makefile to use relative path -o bin/${EXE_DRIVER_NAME} instead of absolute GOPATH-dependent path. This aligns with other CSI drivers and makes builds hermetic-compatible. Comment added to indicate this is a permanent carry patch for OpenShift builds.{code}
Status: Verified
#OCPBUGS-78170issue9 hours ago[openshift-4.16] ose-machine-api-provider-openstack can't be built hermetically New
Issue 17974767: [openshift-4.16] ose-machine-api-provider-openstack can't be built hermetically
Description: This is a clone of issue OCPBUGS-78169. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78168. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-machine-api-provider-openstack hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-77194issue2 weeks ago[openshift-4.18] ose-libvirt-machine-controllers can't be build hermetically New
Issue 17925858: [openshift-4.18] ose-libvirt-machine-controllers can't be build hermetically
Description: Description of problem:
 
 ART can't build ose-libvirt-machine-controllers hermetically.
 The error is the following in the prefetch-dependencies konflux task:
 {code:none}
     Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.  Please try running `go mod vendor` and committing the changes.  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
  
 
  
 
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-77488issue5 days agoOCP 4.20.z: metal3-baremetal-operator pod not honoring NO_PROXY and routing metal3-state service traffic via proxy (works in 4.19.z) Verified
Issue 17941975: OCP 4.20.z: metal3-baremetal-operator pod not honoring NO_PROXY and routing metal3-state service traffic via proxy (works in 4.19.z)
Description: *Summary:*
 In OpenShift Container Platform (OCP) *4.20.z* versions, the *{{metal3-baremetal-operator}}* pod is not honoring the configured *{{NO_PROXY}}* environment variable and is attempting to connect to the internal *{{metal3-state}} service* via the configured HTTP/HTTPS proxy. {*}This behavior is not observed in OCP 4.19.z{*}, where the operator correctly bypasses the proxy for internal service communication.
 
 *Environment:*
  * OCP Version: 4.20.z (multiple z-streams observed)
 
  * Working Version: 4.19.z
 
  * Platform: Baremetal (IPI)
 
  * Proxy configuration enabled (cluster-wide proxy configured)
 
  * Affected Pod: {{metal3-baremetal-operator}}
 
  * Target Service: {{metal3-state}} (ClusterIP service)
 
  
 
 *Please find the reproducer details below. All cluster information is from my test environment and does not contain any sensitive data. I will also share the cluster must-gather shortly.*
 
  
 
  
 {code:java}
 > oc get clusterversion                            
 NAME      VERSION   AVAILABLE   PROGRESSING   SINCE   STATUS
 version   4.20.14   True        False         6h8m    Cluster version is 4.20.14
 
 > oc get proxy -oyaml  
 apiVersion: v1
 items:
 - apiVersion: config.openshift.io/v1
   kind: Proxy
   metadata:
     creationTimestamp: "2026-02-27T07:44:48Z"
     generation: 2
     name: cluster
     resourceVersion: "154487"
     uid: dc9dbe2b-645a-4b2b-98d1-41ab0ef7eebb
   spec:
     httpProxy: http://10.74.234.38:3128
     httpsProxy: http://10.74.234.38:3128
     noProxy: example.com
     trustedCA:
       name: ""
   status:
     httpProxy: http://10.74.234.38:3128
     httpsProxy: http://10.74.234.38:3128
     noProxy: .cluster.local,.svc,10.128.0.0/14,10.74.232.0/21,127.0.0.1,172.30.0.0/16,api-int.ayush.example.com,example.com,localhost
 kind: List
 metadata:
   resourceVersion: ""
 
 > oc get infrastructure/cluster -oyaml 
 apiVersion: config.openshift.io/v1
 kind: Infrastructure
 metadata:
   creationTimestamp: "2026-02-27T07:44:46Z"
   generation: 1
   name: cluster
   resourceVersion: "558"
   uid: 446c09a0-2b33-4ef1-bdf9-683d2cb00203
 spec:
   cloudConfig:
     name: ""
   platformSpec:
     baremetal:
       apiServerInternalIPs:
       - 10.74.236.126
       ingressIPs:
       - 10.74.236.127
       machineNetworks:
       - 10.74.232.0/21
     type: BareMetal
 status:
   apiServerInternalURI: https://api-int.ayush.example.com:6443
   apiServerURL: https://api.ayush.example.com:6443
   controlPlaneTopology: HighlyAvailable
   cpuPartitioning: None
   etcdDiscoveryDomain: ""
   infrastructureName: ayush-hvsdf
   infrastructureTopology: HighlyAvailable
   platform: BareMetal
   platformStatus:
     baremetal:
       apiServerInternalIP: 10.74.236.126
       apiServerInternalIPs:
       - 10.74.236.126
       ingressIP: 10.74.236.127
       ingressIPs:
       - 10.74.236.127
       loadBalancer:
         type: OpenShiftManagedDefault
       machineNetworks:
       - 10.74.232.0/21
     type: BareMetal
 
 > oc get pod                           
 NAME                                                  READY   STATUS    RESTARTS   AGE
 cluster-autoscaler-operator-794cbb6fb-qfqmm           2/2     Running   0          14m
 cluster-baremetal-operator-5d44678794-bvdk6           2/2     Running   0          14m
 control-plane-machine-set-operator-675b5f7b45-xzxt8   1/1     Running   0          14m
 ironic-proxy-b8km9                                    1/1     Running   0          14m
 ironic-proxy-flvtt                                    1/1     Running   1          14m
 ironic-proxy-vmbjm                                    1/1     Running   1          14m
 machine-api-controllers-6687f5fd9f-5mj5t              7/7     Running   0          10m
 machine-api-operator-bc77b9d7b-trm8j                  2/2     Running   0          14m
 metal3-6b6ddc8cf9-nl6hn                               3/3     Running   0          14m
 metal3-baremetal-operator-84f9846555-cqpfx            1/1     Running   0          10m
 metal3-image-customization-b4847b9d6-7hlfz            1/1     Running   0          14m
 
 > oc logs metal3-baremetal-operator-84f9846555-cqpfx | tail -n5                                
 {"level":"info","ts":1772211615.0165863,"logger":"controllers.BareMetalHost","msg":"provisioner is not ready","baremetalhost":{"name":"master2.ayush.example.com","namespace":"openshift-machine-api"},"Error":"Not ready","RequeueAfter":30}
 {"level":"info","ts":1772211615.0166247,"logger":"provisioner.ironic","msg":"error caught while checking endpoint, will retry","host":"openshift-machine-api~compute1.ayush.example.com","endpoint":"https://metal3-state.openshift-machine-api.svc.cluster.local.:6388/v1/","error":"Get \"https://metal3-state.openshift-machine-api.svc.cluster.local.:6388/v1/\": Internal Server Error"}
 {"level":"info","ts":1772211615.0166388,"logger":"controllers.BareMetalHost","msg":"provisioner is not ready","baremetalhost":{"name":"compute1.ayush.example.com","namespace":"openshift-machine-api"},"Error":"Not ready","RequeueAfter":30}
 {"level":"info","ts":1772211615.0168898,"logger":"provisioner.ironic","msg":"error caught while checking endpoint, will retry","host":"openshift-machine-api~master0.ayush.example.com","endpoint":"https://metal3-state.openshift-machine-api.svc.cluster.local.:6388/v1/","error":"Get \"https://metal3-state.openshift-machine-api.svc.cluster.local.:6388/v1/\": Internal Server Error"}
 {"level":"info","ts":1772211615.0169315,"logger":"controllers.BareMetalHost","msg":"provisioner is not ready","baremetalhost":{"name":"master0.ayush.example.com","namespace":"openshift-machine-api"},"Error":"Not ready","RequeueAfter":30}
 
 > oc rsh metal3-baremetal-operator-84f9846555-cqpfx            
 sh-5.1$ 
 sh-5.1$ curl -kv https://metal3-state.openshift-machine-api.svc.cluster.local.:6388/v1/
 * Uses proxy env variable NO_PROXY == '.cluster.local,.svc,10.128.0.0/14,10.74.232.0/21,127.0.0.1,172.30.0.0/16,api-int.ayush.example.com,example.com,localhost,'
 * Uses proxy env variable HTTPS_PROXY == 'http://10.74.234.38:3128'
 *   Trying 10.74.234.38:3128...
 * Connected to 10.74.234.38 (10.74.234.38) port 3128 (#0)
 * allocate connect buffer!
 * Establish HTTP proxy tunnel to metal3-state.openshift-machine-api.svc.cluster.local:6388
 > CONNECT metal3-state.openshift-machine-api.svc.cluster.local:6388 HTTP/1.1
 > Host: metal3-state.openshift-machine-api.svc.cluster.local:6388
 > User-Agent: curl/7.76.1
 > Proxy-Connection: Keep-Alive
 > 
 < HTTP/1.1 500 Internal Server Error
 < Server: squid/5.5
 < Mime-Version: 1.0
 < Date: Fri, 27 Feb 2026 17:00:39 GMT
 < Content-Type: text/html;charset=utf-8
 < Content-Length: 3450
 < X-Squid-Error: ERR_CANNOT_FORWARD 0
 < Vary: Accept-Language
 < Content-Language: en
 < 
 * Received HTTP code 500 from proxy after CONNECT
 * CONNECT phase completed!
 * Closing connection 0
 curl: (56) Received HTTP code 500 from proxy after CONNECT
 sh-5.1$ 
 sh-5.1$ 
 sh-5.1$  
 sh-5.1$ 
 sh-5.1$ unset HTTPS_PROXY
 sh-5.1$ 
 sh-5.1$ curl -kv https://metal3-state.openshift-machine-api.svc.cluster.local.:6388/v1/
 * Uses proxy env variable NO_PROXY == '.cluster.local,.svc,10.128.0.0/14,10.74.232.0/21,127.0.0.1,172.30.0.0/16,api-int.ayush.example.com,example.com,localhost,'
 *   Trying 172.30.114.184:6388...
 * Connected to metal3-state.openshift-machine-api.svc.cluster.local (172.30.114.184) port 6388 (#0)
 * ALPN, offering h2
 * ALPN, offering http/1.1
 *  CAfile: /etc/pki/tls/certs/ca-bundle.crt
 * TLSv1.0 (OUT), TLS header, Certificate Status (22):
 * TLSv1.3 (OUT), TLS handshake, Client hello (1):
 * TLSv1.2 (IN), TLS header, Certificate Status (22):
 * TLSv1.3 (IN), TLS handshake, Server hello (2):
 * TLSv1.2 (IN), TLS header, Finished (20):
 * TLSv1.2 (IN), TLS header, Unknown (23):
 * TLSv1.3 (IN), TLS handshake, Encrypted Extensions (8):
 * TLSv1.2 (IN), TLS header, Unknown (23):
 * TLSv1.3 (IN), TLS handshake, Certificate (11):
 * TLSv1.2 (IN), TLS header, Unknown (23):
 * TLSv1.3 (IN), TLS handshake, CERT verify (15):
 * TLSv1.2 (IN), TLS header, Unknown (23):
 * TLSv1.3 (IN), TLS handshake, Finished (20):
 * TLSv1.2 (OUT), TLS header, Finished (20):
 * TLSv1.3 (OUT), TLS change cipher, Change cipher spec (1):
 * TLSv1.2 (OUT), TLS header, Unknown (23):
 * TLSv1.3 (OUT), TLS handshake, Finished (20):
 * SSL connection using TLSv1.3 / TLS_AES_256_GCM_SHA384
 * ALPN, server accepted to use http/1.1
 * Server certificate:
 *  subject: CN=localhost
 *  start date: Feb 27 07:51:49 2026 GMT
 *  expire date: Feb 27 07:51:50 2028 GMT
 *  issuer: CN=metal3-ironic
 *  SSL certificate verify result: self-signed certificate in certificate chain (19), continuing anyway.
 * TLSv1.2 (OUT), TLS header, Unknown (23):
 > GET /v1/ HTTP/1.1
 > Host: metal3-state.openshift-machine-api.svc.cluster.local:6388
 > User-Agent: curl/7.76.1
 > Accept: */*
 > 
 * TLSv1.2 (IN), TLS header, Unknown (23):
 * TLSv1.3 (IN), TLS handshake, Newsession Ticket (4):
 * TLSv1.2 (IN), TLS header, Unknown (23):
 * TLSv1.3 (IN), TLS handshake, Newsession Ticket (4):
 * old SSL session ID is stale, removing
 * TLSv1.2 (IN), TLS header, Unknown (23):
 * Mark bundle as not supporting multiuse
 < HTTP/1.1 200 OK
 < Date: Fri, 27 Feb 2026 17:00:51 GMT
 < Server: ironic_api
 < X-OpenStack-Ironic-API-Minimum-Version: 1.1
 < X-OpenStack-Ironic-API-Maximum-Version: 1.99
 < X-OpenStack-Ironic-API-Version: 1.1
 < Content-Length: 1029
 < Content-Type: application/json
 < Openstack-Request-Id: req-1d437637-879b-4af3-a8dd-9d7f9d11b172
 < 
 {"id": "v1", "links": [{"href": "https://10.74.236.31:6385/v1/", "rel": "self"}, {"href": "https://docs.openstack.org//ironic/latest/contributor//webapi.html", "rel": "describedby", "type": "text/html"}], "media_types": {"base": "application/json", "type": "application/vnd.openstack.ironic.v1+json"}, "chassis": [{"href": "https://10.74.236.31:6385/v1/chassis/", "rel": "self"}, {"href": "https://10.74.236.31:6385/chassis/", "rel": "bookmark"}], "nodes": [{"href": "https://10.74.236.31:6385/v1/nodes/", "rel": "self"}, {"href": "https://10.74.236.31:6385/nodes/", "rel": "bookmark"}], "ports": [{"href": "https://10.74.236.31:6385/v1/ports/", "rel": "self"}, {"href": "https://10.74.236.31:6385/ports/", "rel": "bookmark"}], "drivers": [{"href": "https://10.74.236.31:6385/v1/drivers/", "rel": "self"}, {"href": "https://10.74.236.31:6385/drivers/", "rel": "bookmark"}], "version": {"id": "v1", "links": [{"href": "https://10.74.236.31:6385/v1/", "rel": "self"}], "status": "CURRENT", "min_version": "1.1", "version": "1.* Connection #0 to host metal3-state.openshift-machine-api.svc.cluster.local left intact
 {code}
 We can ignore the curl behavior for now, as this is expected. Curl does not consistently honor the {{no_proxy}} variables in certain scenarios. The output is shared solely to illustrate that the same response is observed when accessing the service via the proxy as when the pod process connects to the service.
Status: Verified
#OCPBUGS-78039issue34 hours ago[openshift-4.14] ose-powervs-machine-controllers can't be built hermetically POST
Issue 17969611: [openshift-4.14] ose-powervs-machine-controllers can't be built hermetically
Description: This is a clone of issue OCPBUGS-78038. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-powervs-machine-controllers hermetically.
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: POST
#OCPBUGS-77952issue2 days ago[CI watcher] cannot import golang devfile - 40% impact New
Issue 17966207: [CI watcher] cannot import golang devfile - 40% impact
Description: Due to some sort of upstream change in devfile and/or some bug in devfile golang library and/or some bug in our backend
 
 [https://registry.devfile.io/viewer/devfiles/community/go-basic] 
 
  
 
 Create Sample Application from Add page: GS-03-TC05 (example #2): Create the different workloads from Add page Create Sample Application from Add page: GS-03-TC05 (example #2) expand_less     1m43s
 
 {Timed out waiting for [data-test-id="submit-button"] to become enabled Error Error: Timed out waiting for [data-test-id="submit-button"] to become enabled     at checkEnabled (webpack://@dev-console/integration-tests/./support/commands/app.ts:146:17)}
 
  
 
  
 
 [pull-ci-openshift-console-main-e2e-gcp-console|https://prow.ci.openshift.org/job-history/test-platform-results/pr-logs/directory/pull-ci-openshift-console-main-e2e-gcp-console] [(all)|https://search.dptools.openshift.org/?context=1&excludeName=&groupByJob=job&maxAge=336h0m0s&maxBytes=20971520&maxMatches=5&mode=text&name=%5Epull-ci-openshift-console-main-e2e-gcp-console%24&search=%22submit-button%22&searchType=build-log&wrapLines=false] - _39 runs, 100% failed, 46% of failures match = 46% impact_
 
  
 [https://search.dptools.openshift.org/?search=%22submit-button%22&maxAge=24h&context=1&type=build-log&name=pull-ci-openshift-.*&excludeName=&maxMatches=5&maxBytes=20971520&groupBy=job]
Status: New
#OCPBUGS-78172issue9 hours ago[openshift-4.14] ose-machine-api-provider-openstack can't be built hermetically New
Issue 17974701: [openshift-4.14] ose-machine-api-provider-openstack can't be built hermetically
Description: This is a clone of issue OCPBUGS-78171. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78170. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78169. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78168. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-machine-api-provider-openstack hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-35478issue3 months agoconsole never force user log out when accessTokenInactivityTimeoutSeconds reached CLOSED
The error is
app.jsx:652 unhandled error: Uncaught TypeError: B.b.logoutOpenShift is not a function TypeError: B.b.logoutOpenShift is not a function at https://console-openshift-console.apps.qe-uidaily-1015.qe.azure.devcluster.openshift.com/static/main-chunk-2447484….min.js:1:1309163 TypeError: B.b.logoutOpenShift is not a function  at masthead-toolbar.jsx:614:17
masthead-toolbar.jsx:614 Uncaught TypeError: B.b.logoutOpenShift is not a function  at masthead-toolbar.jsx:614:17
Comment 25791250 by UNKNOWN at 2024-10-16T07:55:12.044+0000
#OCPBUGS-66046issue5 weeks agoThe gather job restart loop when username/password field is missing in the Secret Verified
Issue 17613506: The gather job restart loop when username/password field is missing in the Secret
Description: Description of problem:
 {code:none}
     If the password field is missing in the Secret, the pod will restart repeatedly.{code}
 {code:none}
 mustgather-83956-figl-b5v2s            0/2     Completed   0          10m
 mustgather-83956-figl-hl4rh            0/2     Completed   0          18m
 mustgather-83956-figl-rq7m8            2/2     Running     0          81s{code}
 {code:none}
 jitli@jitlideMacBook-Pro:~/work/src/test/must-gather$ oc logs -f pod/mustgather-83956-figl-hl4rh -c upload
 waiting for gathers to complete ...
 waiting for gathers to complete ...
 waiting for gathers to complete ...
 no gather is running (0 / 4)
 no gather is running (1 / 4)
 no gather is running (2 / 4)
 no gather is running (3 / 4)
 no gather is running (4 / 4)
 Error: Required Parameters have not been provided. Exiting...{code}
 {code:none}
 build/bin/upload:63-67
 if [ -z "${caseid}" ] || [ -z "${username}" ] || [ -z "${password}" ];
 then
 echo "Error: Required Parameters have not been provided. Exiting..."
 exit 1
 fi{code}
 {code:none}
 The script checks three variables: caseid, username, and password. If any one of them is empty, it will exit and trigger a restart loop.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     Create a MustGather CR with a no-password Secret {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
 mustgather-83956-figl-b5v2s 0/2 Completed 0 10m 
 mustgather-83956-figl-hl4rh 0/2 Completed 0 18m 
 mustgather-83956-figl-rq7m8 2/2 Running 0 81s{code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: Verified
#OCPBUGS-76501issue3 weeks agocoreos-installer iso kargs show broken on OVE Agent ISO ASSIGNED
Issue 17858359: coreos-installer iso kargs show broken on OVE Agent ISO
Description: This issue is a manually created clone of OCPBUGS-64929 to backport the PR [https://github.com/openshift/appliance/pull/656] manually to 4.21
 
 ----------------
 
 In the 4.20.2 build, attempts to inspect or modify kernel arguments (kargs) in the generated OVE ISO fail.
 {code:none}
 $ coreos-installer iso kargs show agent-ove.x86_64.iso
 Error: No karg embed areas found; old or corrupted CoreOS ISO image.{code}
 Previously, the appliance build did not use the ISO directly as built, it was unpacked and repacked using osirrox / xorriso. The recent change now uses the ISO produced by the appliance directly, which appears to expose an existing issue in how karg embed areas are handled.
 
 [https://github.com/openshift/agent-installer-utils/commit/6d64a0891291701aa0b38b57c8ac9cfa81a69fe4] 
 
 Version-Release number of selected component (if applicable):
 {code:none}
 4.20.2{code}
 How reproducible:
 {code:none}
 Always  {code}
Status: ASSIGNED
#OCPBUGS-77209issue5 days ago[openshift-4.17] hermetic build failure - ose-ibm-vpc-block-csi-driver Verified
Issue 17926347: [openshift-4.17] hermetic build failure - ose-ibm-vpc-block-csi-driver
Description: This is a clone of issue OCPBUGS-77207. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77206. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77205. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77203. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77167. The following is the description of the original issue:  --- Description of problem:
 {code:none}
 IBM VPC Block CSI Driver build fails in hermetic mode due to non-deterministic GOPATH handling in Makefile. The build system rewrites $GOPATH causing the binary output path -o ${GOPATH}/bin/${EXE_DRIVER_NAME} to become invalid, resulting in build failures.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
 OpenShift 4.19+ (affects hermetic build conversion) {code}
 How reproducible:
 {code:none}
 Always - when building ibm-vpc-block-csi-driver in hermetic mode{code}
 Steps to Reproduce:
 {code:none}
   1. Build ibm-vpc-block-csi-driver container in hermetic mode
   2. Makefile uses absolute GOPATH-dependent output path: -o ${GOPATH}/bin/${EXE_DRIVER_NAME}
   3. Build system rewrites $GOPATH during hermetic build process     {code}
 Actual results:
 {code:none}
 Build fails because binary cannot be found at expected GOPATH location after build system modifies $GOPATH
 
 Error: building at STEP "COPY --from=builder /cachi2/output/deps/gomod/bin/ibm-vpc-block-csi-driver /bin/ibm-vpc-block-csi-driver": checking on sources under [...]: copier: stat: "/cachi2/output/deps/gomod/bin/ibm-vpc-block-csi-driver": no such file or directory{code}
 Expected results:
 {code:none}
 Build succeeds using relative path bin/${EXE_DRIVER_NAME} that doesn't depend on GOPATH manipulation, consistent with other CSI drivers like aws-ebs-csi-driver  {code}
 Additional info:
 {code:none}
 Fix implemented: Change Makefile to use relative path -o bin/${EXE_DRIVER_NAME} instead of absolute GOPATH-dependent path. This aligns with other CSI drivers and makes builds hermetic-compatible. Comment added to indicate this is a permanent carry patch for OpenShift builds.{code}
Status: Verified
#OCPBUGS-78043issue33 hours ago[openshift-4.12] ose-powervs-machine-controllers can't be built hermetically New
Issue 17969720: [openshift-4.12] ose-powervs-machine-controllers can't be built hermetically
Description: This is a clone of issue OCPBUGS-78041. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78039. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78038. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-powervs-machine-controllers hermetically.
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-78102issue17 hours ago[openshift-4.13] baremetal-machine-controller can't be built hermetically New
Issue 17972066: [openshift-4.13] baremetal-machine-controller can't be built hermetically
Description: This is a clone of issue OCPBUGS-78101. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78100. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78099. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78098. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build baremetal-machine-controller hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-78163issue9 hours ago[openshift-4.13] ose-metallb-operator can't be built hermetically New
Issue 17974597: [openshift-4.13] ose-metallb-operator can't be built hermetically
Description: This is a clone of issue OCPBUGS-78162. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78161. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-metallb-operator hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-78009issue35 hours ago[openshift-4.13] sriov-network-* images can't be built hermetically POST
Issue 17969326: [openshift-4.13] sriov-network-* images can't be built hermetically
Description: This is a clone of issue OCPBUGS-78007. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build sriov-network-* images hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: POST
#OCPBUGS-77905issue38 hours ago[openshift-4.18] ose-azure-cluster-api-controllers can't be built hermetically Verified
Issue 17962500: [openshift-4.18] ose-azure-cluster-api-controllers can't be built hermetically
Description: This is a clone of issue OCPBUGS-77904. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-azure-cluster-api-controllers
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: Verified
#OCPBUGS-76539issue5 days ago[4.20] Agent service is failing with input/output error Verified
Issue 17862601: [4.20] Agent service is failing with input/output error
Description: *Description of the problem:*
 
 - Unable to discover hosts using BMC because the agent service is failing to start with this error:
 
 Error: netavark: failed to start process: fork/exec /usr/libexec/podman/netavark: input/output error
 
 
 *How reproducible:* Always
 
 *Steps to reproduce:*
 
 1. Boot a system using BMC/BMH
 
 2. ssh the node and check the agent service logs
 
 *Actual results:*
 The agent service fails 
  
 
 *Expected results:*
 The Agent service should start and the hosts should get discovered.
 
 *Additional Info:*
 
 This issue is not faced when:
 
 1. Hosts are discovered using a discovery iso
 2. When we disable converged_flow
Status: Verified
#OCPBUGS-77651issue6 days ago[openshift-4.22] hermetic build failure - node-feature-discovery Verified
Issue 17948666: [openshift-4.22] hermetic build failure - node-feature-discovery
Description: Description of problem:
 {code:none}
   Node Feature Discovery build fails in hermetic mode due to non-deterministic binary path handling in Dockerfile. The build uses "go install" command which places binaries in unpredictable GOPATH locations, and the Dockerfile copies from multiple inconsistent paths causing binary path conflicts.
   {code}
 Version-Release number of selected component (if applicable):
 {code:none}
   OpenShift Container Platform 4.19+
   node-feature-discovery component across all affected versions
   {code}
 How reproducible:
 {code:none}
   Always - when building node-feature-discovery in hermetic mode
   {code}
 Steps to Reproduce:
 {code:none}
   1. Build node-feature-discovery container in hermetic mode
   2. Dockerfile copies binaries from multiple paths: /go/bin/* and /go/node-feature-discovery/bin/*
   3. "go install" command places binaries in non-deterministic GOPATH locations during hermetic build process
   {code}
 Actual results:
 {code:none}
   Build fails because binaries are placed in unpredictable locations by "go install" and Dockerfile attempts to copy from multiple conflicting paths
 
   Error: Inconsistent binary placement causing copy failures during Docker image build process when switching from "go install" to "go build" in
   hermetic mode
   {code}
 Expected results:
 {code:none}
   Build succeeds using deterministic binary paths. Should use "go build" with explicit output paths instead of "go install" and copy binaries from single consistent location in Dockerfile.
   {code}
 Additional info:
 {code:none}
   Root cause: "go install" command places binaries in GOPATH-dependent locations that are not deterministic in hermetic builds, unlike "go build"
    which allows explicit output path control.
 
   Fix implemented: Remove redundant /go/bin/* copy command from Dockerfile and keep only the specific /go/node-feature-discovery/bin/* path. This
    ensures consistent binary placement and eliminates path conflicts during hermetic builds.
 
   Reference upstream work: https://github.com/openshift/node-feature-discovery/pull/163/changes
   Related hermetic conversion tracking: ART-14586
   {code}
Status: Verified
#OCPBUGS-77205issue5 days ago[openshift-4.20] hermetic build failure - ose-ibm-vpc-block-csi-driver Verified
Issue 17926331: [openshift-4.20] hermetic build failure - ose-ibm-vpc-block-csi-driver
Description: This is a clone of issue OCPBUGS-77203. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77167. The following is the description of the original issue:  --- Description of problem:
 {code:none}
 IBM VPC Block CSI Driver build fails in hermetic mode due to non-deterministic GOPATH handling in Makefile. The build system rewrites $GOPATH causing the binary output path -o ${GOPATH}/bin/${EXE_DRIVER_NAME} to become invalid, resulting in build failures.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
 OpenShift 4.19+ (affects hermetic build conversion) {code}
 How reproducible:
 {code:none}
 Always - when building ibm-vpc-block-csi-driver in hermetic mode{code}
 Steps to Reproduce:
 {code:none}
   1. Build ibm-vpc-block-csi-driver container in hermetic mode
   2. Makefile uses absolute GOPATH-dependent output path: -o ${GOPATH}/bin/${EXE_DRIVER_NAME}
   3. Build system rewrites $GOPATH during hermetic build process     {code}
 Actual results:
 {code:none}
 Build fails because binary cannot be found at expected GOPATH location after build system modifies $GOPATH
 
 Error: building at STEP "COPY --from=builder /cachi2/output/deps/gomod/bin/ibm-vpc-block-csi-driver /bin/ibm-vpc-block-csi-driver": checking on sources under [...]: copier: stat: "/cachi2/output/deps/gomod/bin/ibm-vpc-block-csi-driver": no such file or directory{code}
 Expected results:
 {code:none}
 Build succeeds using relative path bin/${EXE_DRIVER_NAME} that doesn't depend on GOPATH manipulation, consistent with other CSI drivers like aws-ebs-csi-driver  {code}
 Additional info:
 {code:none}
 Fix implemented: Change Makefile to use relative path -o bin/${EXE_DRIVER_NAME} instead of absolute GOPATH-dependent path. This aligns with other CSI drivers and makes builds hermetic-compatible. Comment added to indicate this is a permanent carry patch for OpenShift builds.{code}
Status: Verified
#OCPBUGS-77203issue8 hours ago[openshift-4.21] hermetic build failure - ose-ibm-vpc-block-csi-driver CLOSED
Issue 17926162: [openshift-4.21] hermetic build failure - ose-ibm-vpc-block-csi-driver
Description: This is a clone of issue OCPBUGS-77167. The following is the description of the original issue:  --- Description of problem:
 {code:none}
 IBM VPC Block CSI Driver build fails in hermetic mode due to non-deterministic GOPATH handling in Makefile. The build system rewrites $GOPATH causing the binary output path -o ${GOPATH}/bin/${EXE_DRIVER_NAME} to become invalid, resulting in build failures.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
 OpenShift 4.19+ (affects hermetic build conversion) {code}
 How reproducible:
 {code:none}
 Always - when building ibm-vpc-block-csi-driver in hermetic mode{code}
 Steps to Reproduce:
 {code:none}
   1. Build ibm-vpc-block-csi-driver container in hermetic mode
   2. Makefile uses absolute GOPATH-dependent output path: -o ${GOPATH}/bin/${EXE_DRIVER_NAME}
   3. Build system rewrites $GOPATH during hermetic build process     {code}
 Actual results:
 {code:none}
 Build fails because binary cannot be found at expected GOPATH location after build system modifies $GOPATH
 
 Error: building at STEP "COPY --from=builder /cachi2/output/deps/gomod/bin/ibm-vpc-block-csi-driver /bin/ibm-vpc-block-csi-driver": checking on sources under [...]: copier: stat: "/cachi2/output/deps/gomod/bin/ibm-vpc-block-csi-driver": no such file or directory{code}
 Expected results:
 {code:none}
 Build succeeds using relative path bin/${EXE_DRIVER_NAME} that doesn't depend on GOPATH manipulation, consistent with other CSI drivers like aws-ebs-csi-driver  {code}
 Additional info:
 {code:none}
 Fix implemented: Change Makefile to use relative path -o bin/${EXE_DRIVER_NAME} instead of absolute GOPATH-dependent path. This aligns with other CSI drivers and makes builds hermetic-compatible. Comment added to indicate this is a permanent carry patch for OpenShift builds.{code}
Status: CLOSED
#OCPBUGS-71238issue6 weeks agoComponent Readiness: [Installer / openshift-installer] [Other] metal IPI ipv6 Bootstrap failed. Verified
{noformat}
time="2026-01-08T05:22:05Z" level=warning msg="Failed, retrying in 1s ... (1/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp [2600:1f18:483:cf02:c1fc:f41a:9516:ad63]:443: connect: network is unreachable"
time="2026-01-08T05:22:09Z" level=warning msg="Failed, retrying in 1s ... (1/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp [2600:1f18:483:cf02:c1fc:f41a:9516:ad63]:443: connect: network is unreachable"{noformat}
IPv6 jobs run with mirroring because of this issue with reaching Quay through RH networks, but it seems that MCO ignores ImageContentSourcePolicy? In another job, I see references to OSImageStream:
#OCPBUGS-77994issue35 hours ago[openshift-4.14] openshift-enterprise-console-operator can't be built hermetically New
Issue 17968945: [openshift-4.14] openshift-enterprise-console-operator can't be built hermetically
Description: This is a clone of issue OCPBUGS-77993. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77992. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77862. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77861. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77860. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build openshift-enterprise-console-operator hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-78158issue9 hours ago[openshift-4.12] ose-nutanix-machine-controllers can't be built hermetically New
Issue 17974724: [openshift-4.12] ose-nutanix-machine-controllers can't be built hermetically
Description: This is a clone of issue OCPBUGS-78157. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78156. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-nutanix-machine-controllers hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-54795issue13 days agohcp cli "--render" option should work without cluster connectivity New
Issue 15662484: hcp cli "--render" option should work without cluster connectivity
Description: Running {{hcp create cluster agent --render}} ... connects to a remote cluster to validate the existence of the agent namespace:
 
 
 {code:sh}
 $ hcp create cluster agent --render --agent-namespace example
 2023-12-06T14:30:15-05:00       INFO    detected "10.30.8.82" from node "wrk-72" as external-api-server-address
 2023-12-06T14:30:15-05:00       ERROR   Failed to create cluster        {"error": "failed to get agent namespace: namespaces \"example\" not found"}
 github.com/openshift/hypershift/product-cli/cmd/cluster/agent.NewCreateCommand.func1
         /remote-source/app/product-cli/cmd/cluster/agent/create.go:37
 github.com/spf13/cobra.(*Command).execute
         /remote-source/app/vendor/github.com/spf13/cobra/command.go:916
 github.com/spf13/cobra.(*Command).ExecuteC
         /remote-source/app/vendor/github.com/spf13/cobra/command.go:1044
 github.com/spf13/cobra.(*Command).Execute
         /remote-source/app/vendor/github.com/spf13/cobra/command.go:968
 github.com/spf13/cobra.(*Command).ExecuteContext
         /remote-source/app/vendor/github.com/spf13/cobra/command.go:961
 main.main
         /remote-source/app/product-cli/main.go:60
 runtime.main
         /usr/lib/golang/src/runtime/proc.go:250
 Error: failed to get agent namespace: namespaces "example" not found
 failed to get agent namespace: namespaces "example" not found
 
 {code}
 
 {{hcp create cluster --render}} should run without requiring remote connectivity.
 
 Note that this behavior is *different* from that implemented by {{hcp create cluster kubevirt}}, which when run with {{--render}} does not make a connection to the remote cluster.
Status: New
#OCPBUGS-62819issue8 days agoRunning OpenShift 4.19.10, in Azure, with IPI and ACM 2.14 on a Hub-A cluster. Azure appends '-internal' to the load balance ASSIGNED
Issue 17457429: Running OpenShift 4.19.10, in Azure, with IPI and ACM  2.14 on a Hub-A cluster.  Azure appends '-internal' to the load balance
Description: Description of problem:{code:none}
 
 Running OpenShift 4.19.10, in Azure, with IPI and ACM on a Hub-A cluster.
 Start with the following MachinePool yaml
 ``` yaml
 apiVersion: hive.openshift.io/v1
 kind: MachinePool
 metadata:
   annotations:
     argocd.argoproj.io/sync-wave: '2'
   labels:
     app.kubernetes.io/instance: cluster-azncusnprd
   name: azncusnprd-infra
   namespace: azncusnprd
 spec:
   clusterDeploymentRef:
     name: azncusnprd
   labels:
     node-role.kubernetes.io/infra: ''
   name: infra
   platform:
     azure:
       computeSubnet: snet-openshift-dev-ncus-compute
       networkResourceGroupName: az-openshift-dev-001
       osDisk:
         diskSizeGB: 256
       type: Standard_D4s_v3
       virtualNetwork: vnet-openshift-dev-ncus-001
       zones:
         - '1'
         - '2'
         - '3'
   replicas: 3
   taints:
     - effect: NoSchedule
       key: node-role.kubernetes.io/infra
       value: reserved
 ```
 
 The MachinePool is created on Hub-A cluster and successfully generates the corresponding MachineSets on the spoke cluster.
 ```yaml
     - lastProbeTime: '2025-09-30T17:54:44Z'
       lastTransitionTime: '2025-09-30T17:54:44Z'
       message: MachineSets generated successfully
       reason: MachineSetGenerationSucceeded
       status: 'True'
       type: MachineSetsGenerated
 ```
 
 Issue 1 pops up in the Machine status conditions:
 ```yaml
 status:
   phase: Provisioning
   providerStatus:
     conditions:
       - lastTransitionTime: '2025-09-30T17:57:07Z'
         message: 'failed to create nic azncusnprd-zcwrb-infra-northcentralus1-b8fwr-nic for machine azncusnprd-zcwrb-infra-northcentralus1-b8fwr: unable to create VM network interface: load balancer azncusnprd-zcwrb not found: network.LoadBalancersClient#Get: Failure responding to request: StatusCode=404 -- Original Error: autorest/azure: Service returned an error. Status=404 Code="ResourceNotFound" Message="The Resource ''Microsoft.Network/loadBalancers/azncusnprd-zcwrb'' under resource group ''azncusnprd-zcwrb-rg'' was not found. For more details please go to https://aka.ms/ARMResourceNotFoundFix"'
         reason: MachineCreationFailed
         status: 'False'
         type: MachineCreated
 ```
 
 The balancer in Azure has appended `-internal` to the name, so the Machine controller is looking for the wrong name. I do not see an option to pass a custom name into the MachinePool to match the balancer name. This may even have to be an object created in the default Azure resource group before cluster creation.
 
 If I manually create a balancer without the `-internal` suffix the Machine gets further and hits Issue 2:
 ```yaml
   errorMessage: |
     failed to reconcile machine "azncusnprd-zcwrb-infra-northcentralus1-b8fwr": failed to create vm azncusnprd-zcwrb-infra-northcentralus1-b8fwr: failed to create VM: cannot create vm: PUT https://management.azure.com/subscriptions/56a334d6-620c-49b9-a0b7-08fd25ebf304/resourceGroups/azncusnprd-zcwrb-rg/providers/Microsoft.Compute/virtualMachines/azncusnprd-zcwrb-infra-northcentralus1-b8fwr
     --------------------------------------------------------------------------------
     RESPONSE 400: 400 Bad Request
     ERROR CODE: LocationNotSupportAvailabilityZones
     --------------------------------------------------------------------------------
     {
       "error": {
         "code": "LocationNotSupportAvailabilityZones",
         "message": "The resource 'Microsoft.Compute/virtualMachines/azncusnprd-zcwrb-infra-northcentralus1-b8fwr' does not support availability zones at location 'northcentralus'."
       }
     }
     --------------------------------------------------------------------------------
   errorReason: InvalidConfiguration
   lastUpdated: '2025-09-30T18:11:49Z'
   phase: Failed
 ```
 Azure documentation states that the VM does support zones in `northcentralus`.
 If I go into one of the MachineSets and manually remove the spec.template.spec.providerSpec.value.zone: 1 from azncusnprd-zcwrb-infra-northcentralus1 then the machine Provisions successfully and eventually starts up as a node.
 ```yaml
 status:
   providerStatus:
     conditions:
       - lastTransitionTime: '2025-09-30T18:18:31Z'
         message: machine successfully created
         reason: MachineCreationSucceeded
         status: 'True'
         type: MachineCreated
 ```
 
 If I delete the MachinePool, remove the zones section from the yaml, and recreate the MachinePool then OpenShift on Hub-A that creates the MachinePool errors out because of requiring zones.
 ```yaml
 status:
   conditions:
     - lastProbeTime: '2025-09-30T18:34:09Z'
       lastTransitionTime: '2025-09-30T18:34:09Z'
       message: 'could not generate machinesets: zero zones returned for region northcentralus'
       reason: MachineSetGenerationFailed
       status: 'False'
       type: MachineSetsGenerated
 ```
 1. unsure about the load balancer naming issue.
 2. Openshift MachinePool requires a zone for northcentralus. 
 Azure requires no zone for northcentralus.
 
 
 Describe the impact to you or the business
 This is testing for future use cases.
 
     {code}
     Version-Release number of selected component (if applicable):{code:none}
 
     {code}
     How reproducible:{code:none}
 
     {code}
     Steps to Reproduce:{code:none}
 See above 
     {code}
     Actual results:{code:none}
 
 Requires manual intervention to fix the install issues
 
     {code}
     Expected results:{code:none}
 
 No manual intervention
 
     {code}
     Additional info:{code:none}
 
     {code}
Status: ASSIGNED
#OCPBUGS-78020issue34 hours ago[openshift-4.16] ose-vsphere-cluster-api-controllers can't be built hermetically New
Issue 17969352: [openshift-4.16] ose-vsphere-cluster-api-controllers can't be built hermetically
Description: This is a clone of issue OCPBUGS-78019. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78018. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78017. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-vsphere-cluster-api-controllers hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-78019issue34 hours ago[openshift-4.17] ose-vsphere-cluster-api-controllers can't be built hermetically New
Issue 17969423: [openshift-4.17] ose-vsphere-cluster-api-controllers can't be built hermetically
Description: This is a clone of issue OCPBUGS-78018. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78017. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-vsphere-cluster-api-controllers hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-72371issue8 weeks agoInstallation behind proxy failed when enable TP in 4.22 New
Issue 17742056: Installation behind proxy failed when enable TP in 4.22
Description: Description of problem:
 {code:none}
  https://reportportal-openshift.apps.dno.ocp-hub.prod.psi.redhat.com/ui/#prow/launches/1578/823195 
 this prow job failed and i try to reproduce. finally i found the installation behind proxy will failed when enable TP in 4.22. the bootstrap was failed, the master machines was created but can't fetch the ignition.
 The machine-config-controller has been continuously rolling back.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     4.22.0-0.nightly-2026-01-06-164201{code}
 How reproducible:
 {code:none}
     install cluster enable TP and proxy.{code}
 Steps to Reproduce:
 {code:none}
     1.install a cluster behind proxy and enable tp.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     installation failed.{code}
 Expected results:
 {code:none}
     installation successful.    {code}
 Additional info:
 {code:none}
 time="2026-01-08T06:33:46Z" level=warning msg="Failed, retrying in 1s ... (1/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 52.203.142.72:443: i/o timeout"
 E0108 06:34:16.850064    1 imagestream_source.go:62] error inspecting ImageStream Image for OSImageStream generation. quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:53f09f0ee42099faddae15e2a3364f6467396c55f34660e6cb8492e6a9b9f3ae image will is ignored. context deadline exceeded
 E0108 06:34:16.850074    1 imagestream_source.go:62] error inspecting ImageStream Image for OSImageStream generation. quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:cec23b7e42ea89169a7f5c0f42bcc78cfb64d7297833ea989db33d9ead97e322 image will is ignored. error getting image source for quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:cec23b7e42ea89169a7f5c0f42bcc78cfb64d7297833ea989db33d9ead97e322: error getting image source for //quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:cec23b7e42ea89169a7f5c0f42bcc78cfb64d7297833ea989db33d9ead97e322: pinging container registry quay.io: Get "https://quay.io/v2/": context deadline exceeded 
 F0108 06:34:16.850456    1 bootstrap.go:47] error running MCC[BOOTSTRAP]: error inspecting available OSImageStreams: unable to retrieve any OSImageStream from the configured sources{code}
Status: New
#OCPBUGS-78112issue16 hours ago[openshift-4.14] ose-agent-installer-api-server can't be built hermetically POST
Issue 17972532: [openshift-4.14] ose-agent-installer-api-server can't be built hermetically
Description: This is a clone of issue OCPBUGS-78111. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-agent-installer-api-server hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: POST
#OCPBUGS-77195issue2 weeks ago[openshift-4.17] ose-libvirt-machine-controllers can't be build hermetically New
Issue 17925859: [openshift-4.17] ose-libvirt-machine-controllers can't be build hermetically
Description: Description of problem:
 
 ART can't build ose-libvirt-machine-controllers hermetically.
 The error is the following in the prefetch-dependencies konflux task:
 {code:none}
     Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.  Please try running `go mod vendor` and committing the changes.  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
  
 
  
 
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-78177issue9 hours ago[openshift-4.15] ose-azure-cluster-api-controllers can't be built hermetically New
Issue 17974782: [openshift-4.15] ose-azure-cluster-api-controllers can't be built hermetically
Description: This is a clone of issue OCPBUGS-78176. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77983. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77905. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77904. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-azure-cluster-api-controllers
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-76642issue6 days agoTestImageRegistryPods test failing in CMO end-to-end test suite Verified
Issue 17871754: TestImageRegistryPods test failing in CMO end-to-end test suite
Description: This is a clone of issue OCPBUGS-76488. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-66047. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-63235. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-63123. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-63017. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-62453. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-62308. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-62109. The following is the description of the original issue:  --- We can't merge pull requests in github.com/openshift/cluster-monitoring because of the test failure. Example of failed job run:
 https://prow.ci.openshift.org/view/gs/test-platform-results/pr-logs/pull/openshift_cluster-monitoring-operator/2649/pull-ci-openshift-cluster-monitoring-operator-main-e2e-agnostic-operator/1970148446123855872
 
 {code:none}
  === RUN   TestUserWorkloadMonitoringAlerting/assert_containers_images_registry_is_the_same_than_CMO
     image_registry_test.go:52: 
         	Error Trace:	/go/src/github.com/openshift/cluster-monitoring-operator/test/e2e/image_registry_test.go:52
         	            				/go/src/github.com/openshift/cluster-monitoring-operator/test/e2e/image_registry_test.go:18
         	Error:      	Not equal: 
         	            	expected: "registry.build05.ci.openshift.org"
         	            	actual  : "quay-proxy.ci.openshift.org"
         	            	
         	            	Diff:
         	            	--- Expected
         	            	+++ Actual
         	            	@@ -1 +1 @@
         	            	-registry.build05.ci.openshift.org
         	            	+quay-proxy.ci.openshift.org
         	Test:       	TestUserWorkloadMonitoringAlerting/assert_containers_images_registry_is_the_same_than_CMO
 === RUN   TestUserWorkloadMonitoringAlerting/assert_cross-namespace_rules_are_not_allowed_from_CMO_config
 --- FAIL: TestUserWorkloadMonitoringAlerting (219.63s)
     --- PASS: TestUserWorkloadMonitoringAlerting/assert_user_workload_rules (0.07s)
     --- PASS: TestUserWorkloadMonitoringAlerting/assert_tenancy_model_is_enforced_for_rules_and_alerts (10.29s)
     --- PASS: TestUserWorkloadMonitoringAlerting/assert_rules_without_namespace_enforcement (50.01s)
     --- PASS: TestUserWorkloadMonitoringAlerting/assert_prometheus_is_not_deployed_in_user_namespace (0.00s)
     --- PASS: TestUserWorkloadMonitoringAlerting/assert_alertmanager_is_not_deployed_in_user_namespace (0.01s)
     --- FAIL: TestUserWorkloadMonitoringAlerting/assert_containers_images_registry_is_the_same_than_CMO (2.01s)
     --- PASS: TestUserWorkloadMonitoringAlerting/assert_cross-namespace_rules_are_not_allowed_from_CMO_config (85.02s)
 
 {code}
Status: Verified
#OCPBUGS-77191issue2 weeks ago[openshift-4.21] ose-libvirt-machine-controllers can't be build hermetically New
Issue 17925851: [openshift-4.21] ose-libvirt-machine-controllers can't be build hermetically
Description: Description of problem:
 
 ART can't build ose-libvirt-machine-controllers hermetically.
 The error is the following in the prefetch-dependencies konflux task:
 {code:none}
     Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.  Please try running `go mod vendor` and committing the changes.  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
  
 
  
 
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-77152issue5 days ago[openshift-4.20] ose-gcp-cloud-controller-manager can't be build hermetically Verified
Issue 17918009: [openshift-4.20] ose-gcp-cloud-controller-manager can't be build hermetically
Description: Description of problem:
 
 ART can't build ose-gcp-cloud-controller-manager hermetically.
 The error is the following in the prefetch-dependencies konflux task:
 {code:none}
    Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.  Please try running `go mod vendor` and committing the changes.  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 This happens only for openshift-4.20
 
  
 
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: Verified
#OCPBUGS-76535issue2 weeks ago[openshift-4.22] ose-libvirt-machine-controllers can't be build hermetically POST
Issue 17861873: [openshift-4.22] ose-libvirt-machine-controllers can't be build hermetically
Description: Description of problem:
 
 ART can't build ose-libvirt-machine-controllers hermetically.
 The error is the following in the prefetch-dependencies konflux task:
 {code:none}
     Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.  Please try running `go mod vendor` and committing the changes.  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
  
 
  
 
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: POST
#OCPBUGS-77993issue35 hours ago[openshift-4.15] openshift-enterprise-console-operator can't be built hermetically New
Issue 17968941: [openshift-4.15] openshift-enterprise-console-operator can't be built hermetically
Description: This is a clone of issue OCPBUGS-77992. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77862. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77861. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77860. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build openshift-enterprise-console-operator hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-77991issue33 hours agoComponent Readiness: nested container should pass podman localsystem test - suite failing POST
{code:java}
Error: building at STEP "ADD https://github.com/containers/podman/blob/main/README.md /tmp/": reading "https://github.com/containers/podman/blob/main/README.md": invalid response status 429{code}
The error indicates that [https://github.com/containers/podman/blob/main/README.md] returns 429 TooManyRequests.
#OCPBUGS-77784issue31 hours ago[openshift-4.18] ose-sriov-rdma-cni can't be built hermetically New
Issue 17955267: [openshift-4.18] ose-sriov-rdma-cni can't be built hermetically
Description: This is a clone of issue OCPBUGS-77775. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-sriov-rdma-cni hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: POST
#OCPBUGS-662issue2 months agoGitLab runner operator fails to install. New
Issue 14818715: GitLab runner operator fails to install.
Description: Description of problem: The Certified GitLab Operator fails to install and results in below error:
 ~~~
 10s         Warning   Failed                         pod/gitlab-runner-controller-manager-86668cd5c8-tvmnr        Failed to pull image "registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:ead726da6f75d80ff7b389dac853096d8680b9910cd5079d3aaf77a31846bf7b": rpc error: code = Unknown desc = reading manifest sha256:ead726da6f75d80ff7b389dac853096d8680b9910cd5079d3aaf77a31846bf7b in registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator: manifest unknown: manifest unknown
 10s         Warning   Failed                         pod/gitlab-runner-controller-manager-86668cd5c8-tvmnr        Error: ErrImagePull
 24s         Normal    BackOff                        pod/gitlab-runner-controller-manager-86668cd5c8-tvmnr        Back-off pulling image "registry.gitlab.com/gitlab-org/gl-openshift/gitlab-runner-operator/gitlab-runner-operator@sha256:ead726da6f75d80ff7b389dac853096d8680b9910cd5079d3aaf77a31846bf7b"
 24s         Warning   Failed                         pod/gitlab-runner-controller-manager-86668cd5c8-tvmnr        Error: ImagePullBackOff
 
 ~~~
 
 The pod is in ImagePullBackOff state:
 ~~~
 $ oc get pods
 NAME                                                READY   STATUS             RESTARTS   AGE
 gitlab-runner-controller-manager-86668cd5c8-tvmnr   1/2     ImagePullBackOff   0          3m53s
 ~~~
 
 Version-Release number of selected component (if applicable): OCP 4.10
 
 How reproducible: Always
 
 
 Steps to Reproduce:
 1. Go to Operator Hub
 2. Search for GitLab Runner
 3. Install
 4. Monitor the Events and pod.
 
 Actual results: Pod in ImagePullBackOff
 
 Expected results: Operator should work as expected.
 
 Additional info: We tried installing the Community GitLab runner and it works as expected.
Status: New
#OCPBUGS-78178issue9 hours ago[openshift-4.14] ose-azure-cluster-api-controllers can't be built hermetically New
Issue 17974849: [openshift-4.14] ose-azure-cluster-api-controllers can't be built hermetically
Description: This is a clone of issue OCPBUGS-78177. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78176. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77983. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77905. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77904. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-azure-cluster-api-controllers
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-49422issue5 weeks agoUnable to setup AMD GPU operator in an air-gapped env New
Issue 16626817: Unable to setup AMD GPU operator in an air-gapped env
Description: h3.Description
 Installing AMD GPU operator (amd-gpu-operator.v1.1.1) from certified-operators (channel 'alpha') in an air-gapped env (AWS).
 
 h3.Problem
 The problem seems to be with the {{DeviceConfig}} which triggers a "build" pod (pls correct me if I'm wrong), in my case
 {code}
 dc-internal-registry-build-6rvp6-build
 {code}
 Unfortunately, one of the steps in that build is
 {code}
 [1/2] STEP 5/9: RUN dnf install https://dl.fedoraproject.org/pub/epel/epel-release-latest-9.noarch.rpm -y &&     crb enable &&     sed -i "s/\$releasever/9/g" /etc/yum.repos.d/epel*.repo &&     dnf install dnf-plugin-config-manager -y &&     dnf clean all
 {code}
 which fails on
 {code}
 Curl error (28): Timeout was reached for https://cdn-ubi.redhat.com/content/public/ubi/dist/ubi9/9/x86_64/baseos/os/repodata/repomd.xml [Failed to connect to cdn-ubi.redhat.com port 443: Connection timed out]
 Error: Failed to download metadata for repo 'ubi-9-baseos': Cannot download repomd.xml: Cannot download repodata/repomd.xml: All mirrors were tried
 error: build error: building at STEP "RUN dnf install https://dl.fedoraproject.org/pub/epel/epel-release-latest-9.noarch.rpm -y &&     crb enable &&     sed -i "s/\$releasever/9/g" /etc/yum.repos.d/epel*.repo &&     dnf install dnf-plugin-config-manager -y &&     dnf clean all": while running runtime: exit status 1
 {code}
 Obviously, we cannot do such steps in a disconnected env.
 
 EDIT:
 To avoid the building step, we could use a pre-compiled image as described at
 https://dcgpu.docs.amd.com/projects/gpu-operator/en/main/drivers/precompiled-driver.html#using-pre-compiled-images
Status: New
#OCPBUGS-58473issue7 months ago[CI][Suite:openshift/openstack][lb][Serial] The Openstack platform LoadBalancer tests are failing CLOSED
Issue 17124490: [CI][Suite:openshift/openstack][lb][Serial] The Openstack platform LoadBalancer tests are failing
Description: *Description of problem:*
 
 The Openstack platform LoadBalancer tests are failing only on [periodic-ci-shiftstack-ci-release-4.20-e2e-openstack-ccpmso|https://prow.ci.openshift.org/job-history/gs/test-platform-results/logs/periodic-ci-shiftstack-ci-release-4.20-e2e-openstack-ccpmso] jobs by authentication issues.
 
 The failed tests are:
 {code:none}
 [sig-installer][Suite:openshift/openstack] Bugfix ocpbug_1765: [Serial] noAllowedAddressPairs on one port should not affect other ports
 [sig-installer][Suite:openshift/openstack] The OpenShift cluster should allow the manual setting of enable_topology to false [Serial]
 [sig-installer][Suite:openshift/openstack][lb][Serial] The Openstack platform should apply lb-method on TCP Amphora LoadBalancer when a TCP svc with monitors and ETP:Local is created on Openshift
 [sig-installer][Suite:openshift/openstack][lb][Serial] The Openstack platform should apply lb-method on TCP OVN LoadBalancer when a TCP svc with monitors and ETP:Local is created on Openshift
 [sig-installer][Suite:openshift/openstack][lb][Serial] The Openstack platform should apply lb-method on UDP Amphora LoadBalancer when a UDP svc with monitors and ETP:Local is created on Openshift
 [sig-installer][Suite:openshift/openstack][lb][Serial] The Openstack platform should apply lb-method on UDP OVN LoadBalancer when a UDP svc with monitors and ETP:Local is created on Openshift
 [sig-installer][Suite:openshift/openstack][lb][Serial] The Openstack platform should create a TCP Amphora LoadBalancer when LoadBalancerService ingressController is created on Openshift
 [sig-installer][Suite:openshift/openstack][lb][Serial] The Openstack platform should create a TCP OVN LoadBalancer when LoadBalancerService ingressController is created on Openshift
 [sig-installer][Suite:openshift/openstack][lb][Serial] The Openstack platform should create a TCP OVN LoadBalancer when a TCP svc with type:LoadBalancer is created on Openshift
 [sig-installer][Suite:openshift/openstack][lb][Serial] The Openstack platform should create a UDP Amphora LoadBalancer when a UDP svc with type:LoadBalancer is created on Openshift
 [sig-installer][Suite:openshift/openstack][lb][Serial] The Openstack platform should create a UDP OVN LoadBalancer when a UDP svc with type:LoadBalancer is created on Openshift
 [sig-installer][Suite:openshift/openstack][lb][Serial] The Openstack platform should create an UDP Amphora LoadBalancer using a pre-created FIP when an UDP LoadBalancer svc setting the LoadBalancerIP spec is created on Openshift
 [sig-installer][Suite:openshift/openstack][lb][Serial] The Openstack platform should create an UDP OVN LoadBalancer using a pre-created FIP when an UDP LoadBalancer svc setting the LoadBalancerIP spec is created on Openshift
 [sig-installer][Suite:openshift/openstack][lb][Serial] The Openstack platform should limit service access on an UDP OVN LoadBalancer when an UDP LoadBalancer svc setting the loadBalancerSourceRanges spec is created on Openshift
 [sig-installer][Suite:openshift/openstack][lb][Serial] The Openstack platform should not re-use an existing UDP Amphora LoadBalancer if shared internal svc is created
 [sig-installer][Suite:openshift/openstack][lb][Serial] The Openstack platform should re-use an existing UDP Amphora LoadBalancer when new svc is created on Openshift with the proper annotation
 [sig-installer][Suite:openshift/openstack][lb][Serial] The Openstack platform should re-use an existing UDP OVN LoadBalancer when new svc is created on Openshift with the proper annotation    {code}
 *Version-Release number of selected component (if applicable):*
 {code:none}
 4.20    {code}
 *How reproducible:*
 
 [https://steps.ci.openshift.org/job?org=shiftstack&repo=ci&branch=release-4.20&test=e2e-openstack-ccpmso]
 
 *Steps to Reproduce:*
 {code:none}
     1.
     2.
     3.
     {code}
 *Actual results:*
 {code:none}
 Failing    {code}
 Expected results:
 {code:none}
 Pass    {code}
 *Additional info:*
 
 From this log [https://gcsweb-ci.apps.ci.l2s4.p1.openshiftapps.com/gcs/test-platform-results/logs/periodic-ci-shiftstack-ci-release-4.20-e2e-openstack-ccpmso/1942220378185142272/artifacts/e2e-openstack-ccpmso/openstack-test-openstack/build-log.txt]:
 {code:none}
 fail [github.com/openshift/openstack-test/test/extended/openstack/loadbalancers.go:76]: Failed to build the OpenStack client
 Unexpected error:
     <gophercloud.ErrUnexpectedResponseCode>: 
     Expected HTTP response code [201 202] when accessing [POST https://identity.rdo.mtl2.vexxhost.net/v3/auth/tokens], but got 401 instead: {"error":{"code":401,"message":"The request you have made requires authentication.","title":"Unauthorized"}}
     {
         BaseError: {DefaultErrString: "", Info: ""},
         URL: "https://identity.rdo.mtl2.vexxhost.net/v3/auth/tokens",
         Method: "POST",
         Expected: [201, 202],
         Actual: 401,    {code}
Status: CLOSED
#OCPBUGS-78099issue5 hours ago[openshift-4.16] baremetal-machine-controller can't be built hermetically MODIFIED
Issue 17972332: [openshift-4.16] baremetal-machine-controller can't be built hermetically
Description: This is a clone of issue OCPBUGS-78098. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build baremetal-machine-controller hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: ON_QA
#OCPBUGS-66942issue2 months agosos report hit error during collect datas for openshift_ovn plugin New
Issue 17648864: sos report hit error during collect datas for openshift_ovn plugin
Description: Description of problem:
 {code:none}
 sos report hit error during collect datas for openshift_ovn plugin {code}
 Version-Release number of selected component (if applicable):
 {code:none}
 ocp version: 4.21
 
 sos version:
 sh-5.1# rpm -qa | grep sos     
 sos-4.10.0-4.el9_4.noarch{code}
 How reproducible:
 {code:none}
 always{code}
 Steps to Reproduce:
 {code:none}
 1. setup 4.21 cluster
 2. debug into one of the node
 3. run sos report to collect datas for openshift_ovn plugin
 
 
 
 sh-5.1# sos report --only-plugins=openshift_ovn
 
 
 sos report (version 4.10.0)
 
 
 WARNING: unable to set option for disabled or non-existing plugin (crio).
 WARNING: unable to set option for disabled or non-existing plugin (podman).
 
 
 This command will collect diagnostic and configuration information from
 this Red Hat Enterprise Linux system and installed applications.
 
 
 An archive containing the collected information will be generated in
 /host/var/tmp/sos.k_j0wvlj and may be provided to a Red Hat support
 representative.
 
 
 Any information provided to Red Hat will be treated in accordance with
 the published support policies at:
 
 
         Distribution Website : https://www.redhat.com/
         Commercial Support   : https://access.redhat.com/
 
 
 The generated archive may contain data considered sensitive and its
 content should be reviewed by the originating organization before being
 passed to any third party.
 
 
 No changes will be made to system configuration.
 
 
 Press ENTER to continue, or CTRL-C to quit.
 
 
 Optionally, please enter the case id that you are generating this report for: 
  Setting up archive ...
  Setting up plugins ...
 caught exception in plugin method "openshift_ovn.setup()"
 writing traceback to sos_logs/openshift_ovn-plugin-errors.txt
  Running plugins. Please wait ...
 
 
   Starting 1/1   openshift_ovn   [Running: openshift_ovn]                                 
   Finished running plugins                                                               
 Creating compressed archive...
 
 
 Your sos report has been generated and saved in:
 	/host/var/tmp/sosreport-ip-10-0-33-119-2025-12-08-tjievgx.tar.xz
 
 
  Size	345.62KiB
  Owner	root
  sha256	b009e4bcb2cf4d183e8c795d1329f5ad7979b4e53153f60adaeabab4a62c2e0c
 
 
 Please send this file to your support representative.
 
 
 sh-5.1# tar -xvf /host/var/tmp/sosreport-ip-10-0-33-119-2025-12-08-tjievgx.tar.xz
 sosreport-ip-10-0-33-119-2025-12-08-tjievgx/version.txt
 sosreport-ip-10-0-33-119-2025-12-08-tjievgx/sos_reports/
 sosreport-ip-10-0-33-119-2025-12-08-tjievgx/sos_reports/manifest.json
 sosreport-ip-10-0-33-119-2025-12-08-tjievgx/sos_reports/sos.html
 sosreport-ip-10-0-33-119-2025-12-08-tjievgx/sos_reports/sos.json
 sosreport-ip-10-0-33-119-2025-12-08-tjievgx/sos_reports/sos.txt
 sosreport-ip-10-0-33-119-2025-12-08-tjievgx/sos_logs/
 sosreport-ip-10-0-33-119-2025-12-08-tjievgx/sos_logs/openshift_ovn-plugin-errors.txt
 sosreport-ip-10-0-33-119-2025-12-08-tjievgx/sos_logs/sos.log
 sosreport-ip-10-0-33-119-2025-12-08-tjievgx/sos_logs/ui.log
 sosreport-ip-10-0-33-119-2025-12-08-tjievgx/
 sosreport-ip-10-0-33-119-2025-12-08-tjievgx/environment
 sosreport-ip-10-0-33-119-2025-12-08-tjievgx/sos_commands/
 sosreport-ip-10-0-33-119-2025-12-08-tjievgx/var/
 sosreport-ip-10-0-33-119-2025-12-08-tjievgx/var/lib/
 sosreport-ip-10-0-33-119-2025-12-08-tjievgx/var/lib/openvswitch/
 sosreport-ip-10-0-33-119-2025-12-08-tjievgx/var/lib/openvswitch/etc/
 sosreport-ip-10-0-33-119-2025-12-08-tjievgx/var/lib/openvswitch/etc/keys/
 sosreport-ip-10-0-33-119-2025-12-08-tjievgx/var/lib/openvswitch/etc/keys/ipsec-cacert.pem
 sosreport-ip-10-0-33-119-2025-12-08-tjievgx/var/lib/openvswitch/etc/keys/ipsec-cert.pem
 sosreport-ip-10-0-33-119-2025-12-08-tjievgx/var/lib/openvswitch/etc/keys/ipsec-privkey.pem
 sosreport-ip-10-0-33-119-2025-12-08-tjievgx/var/lib/openvswitch/etc/keys/ipsec-req.pem
 sosreport-ip-10-0-33-119-2025-12-08-tjievgx/var/lib/ovn-ic/
 sosreport-ip-10-0-33-119-2025-12-08-tjievgx/var/lib/ovn-ic/etc/
 sosreport-ip-10-0-33-119-2025-12-08-tjievgx/var/lib/ovn-ic/etc/libovsdb.log
 sosreport-ip-10-0-33-119-2025-12-08-tjievgx/var/lib/ovn-ic/etc/ovnnb_db.db
 sosreport-ip-10-0-33-119-2025-12-08-tjievgx/var/lib/ovn-ic/etc/ovnsb_db.db     {code}
 Actual results:
 {code:none}
 There is no directory openshift_ovn under sos_commands directory:
 sh-5.1# ls -l sosreport-ip-10-0-33-119-2025-12-08-tjievgx/sos_commands/
 total 0
 
 
 And there is errors for collecting openshift_ovn plugin:
 sh-5.1# cat sosreport-ip-10-0-33-119-2025-12-08-tjievgx/sos_logs/openshift_ovn-plugin-errors.txt Traceback (most recent call last):   File "/usr/lib/python3.9/site-packages/sos/report/__init__.py", line 1260, in setup     plug.setup()   File "/usr/lib/python3.9/site-packages/sos/report/plugins/openshift_ovn.py", line 74, in setup     self.add_cmd_output([   File "/usr/lib/python3.9/site-packages/sos/report/plugins/__init__.py", line 2212, in add_cmd_output     cmd = self.fmt_container_cmd(container, cmd, runtime=runtime,   File "/usr/lib/python3.9/site-packages/sos/report/plugins/__init__.py", line 2970, in fmt_container_cmd     return _runtime.fmt_container_cmd(container, cmd, quotecmd) UnboundLocalError: local variable '_runtime' referenced before assignment{code}
 Expected results:
 {code:none}
 sos report can collect datas for openshift_ovn plugins{code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-77087issue6 days agomultus-networkpolicy not building successfully in Konflux CLOSED
Issue 17910000: multus-networkpolicy not building successfully in Konflux
Description: multus-networkpolicy hasn't built successfully in Konflux since [https://github.com/openshift/multus-networkpolicy/pull/77] went in adding nftables package dependency
 
 Example here [https://art-build-history-art-build-history.apps.artc2023.pc3z.p1.openshiftapps.com/logs?nvr=ose-multus-networkpolicy-container-v4.22.0-202602191351.p2.g25d18cf.assembly.stream.el9&record_id=ca61f00c-c694-859d-a8c9-08afbcc2e55c&group=openshift-4.22]
 
 Logs end at
 [2/2] STEP 1/19: FROM quay.io/redhat-user-workloads/ocp-art-tenant/art-images@sha256:955b4479354cd94013ca7b6707a3affafb0806c8d99f70cd76e4a0d2f929c9b6
 [2/2] STEP 2/19: ENV ART_BUILD_ENGINE=konflux
 [2/2] STEP 3/19: ENV ART_BUILD_DEPS_METHOD=cachi2
 [2/2] STEP 4/19: ENV ART_BUILD_NETWORK=hermetic
 [2/2] STEP 5/19: RUN . /cachi2/cachi2.env &&     go clean -cache || true
 /bin/sh: line 1: go: command not found
 [2/2] STEP 6/19: ENV ART_BUILD_DEPS_MODE=default
 [2/2] STEP 7/19: ENV __doozer=update BUILD_RELEASE=202602191351.p2.g25d18cf.assembly.stream.el9 BUILD_VERSION=v4.22.0 OS_GIT_MAJOR=4 OS_GIT_MINOR=22 OS_GIT_PATCH=0 OS_GIT_TREE_STATE=clean OS_GIT_VERSION=4.22.0-202602191351.p2.g25d18cf.assembly.stream.el9 SOURCE_GIT_TREE_STATE=clean __doozer_group=openshift-4.22 __doozer_key=multus-networkpolicy __doozer_uuid_tag=ose-multus-networkpolicy-rhel9-v4.22.0-20260219.142513 __doozer_version=v4.22.0 
 [2/2] STEP 8/19: ENV __doozer=merge OS_GIT_COMMIT=25d18cf OS_GIT_VERSION=4.22.0-202602191351.p2.g25d18cf.assembly.stream.el9-25d18cf SOURCE_DATE_EPOCH=1771275974 SOURCE_GIT_COMMIT=25d18cf8247ad6ba1a6cc8e37cc63f236eaee689 SOURCE_GIT_TAG=25d18cf8 SOURCE_GIT_URL=https://github.com/openshift/multus-networkpolicy 
 [2/2] STEP 9/19: RUN . /cachi2/cachi2.env &&     dnf install -y nftables && dnf clean all
 error: No package matches 'nftables'
 subprocess exited with status 1
 subprocess exited with status 1
 Error: building at STEP "RUN . /cachi2/cachi2.env &&     dnf install -y nftables && dnf clean all": exit status 1
Status: CLOSED
#OCPBUGS-77983issue38 hours ago[openshift-4.17] ose-azure-cluster-api-controllers can't be built hermetically POST
Issue 17968065: [openshift-4.17] ose-azure-cluster-api-controllers can't be built hermetically
Description: This is a clone of issue OCPBUGS-77905. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77904. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-azure-cluster-api-controllers
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: POST
#OCPBUGS-30345issue2 weeks agoOperatorHub.yaml is missing from the ztp source-crs Verified
Issue 15860647: OperatorHub.yaml is missing from the ztp source-crs
Description: This is a clone of issue OCPBUGS-27893. The following is the description of the original issue:  --- Description of problem:
 {code:none}
 gitops deployment fails - Policies are not syncing because OperatorHub.yaml is missing from the ztp source-crs.{code}
 [kni@registry.kni-qe-27 ~]$ oc get apps -A
 NAMESPACE NAME SYNC STATUS HEALTH STATUS
 openshift-gitops clusters Synced Healthy
 openshift-gitops policies Unknown Healthy
 
 [kni@registry.kni-qe-27 ~]$ oc get apps -n openshift-gitops policies -o yaml 
 [...]
 status:
 conditions:
  - lastTransitionTime: "2024-01-24T15:08:10Z"
 message: |-
 Failed to load target state: failed to generate manifest for source 1 of 1: rpc error: code = Unknown desc = `kustomize build <path to cached source>/policygentemplates --enable-alpha-plugins` failed exit status 1: 2024/01/24 15:08:10 Could not build the entire policy defined by /tmp/kust-plugin-config-673779590: OperatorHub.yaml is not found both in Git path: <path to cached source>/policygentemplates/source-crs/OperatorHub.yaml and in the ztp container path: /.config/kustomize/plugin/ran.openshift.io/v1/policygentemplate/source-crs/OperatorHub.yaml
 Error: failure in plugin configured via /tmp/kust-plugin-config-673779590; exit status 1: exit status 1
 type: ComparisonError
  
 
 Version-Release number of selected component (if applicable):
 {code:none}
  OCP version: 4.15
 
 ztp-site-generator:v4.15.0-30{code}
 How reproducible:
 {code:none}
     100%{code}
 Steps to Reproduce:
 {code:none}
     1. Deploy 4.15 SNO 
     2. Deploy gitops-ztp
     3. Check apps
 
 [kni@registry.kni-qe-27 ~]$ oc get apps -A
 NAMESPACE          NAME       SYNC STATUS   HEALTH STATUS
 openshift-gitops   clusters   Synced        Healthy
 openshift-gitops   policies   Unknown       Healthy
 
 [kni@registry.kni-qe-27 ~]$ oc get apps -n openshift-gitops policies -o yaml 
 [...]
 status:
   conditions:
   - lastTransitionTime: "2024-01-24T15:08:10Z"
     message: |-
       Failed to load target state: failed to generate manifest for source 1 of 1: rpc error: code = Unknown desc = `kustomize build <path to cached source>/policygentemplates --enable-alpha-plugins` failed exit status 1: 2024/01/24 15:08:10 Could not build the entire policy defined by /tmp/kust-plugin-config-673779590: OperatorHub.yaml is not found both in Git path: <path to cached source>/policygentemplates/source-crs/OperatorHub.yaml and in the ztp container path: /.config/kustomize/plugin/ran.openshift.io/v1/policygentemplate/source-crs/OperatorHub.yaml
       Error: failure in plugin configured via /tmp/kust-plugin-config-673779590; exit status 1: exit status 1
     type: ComparisonError    {code}
 Actual results:
 {code:none}
  openshift-gitops policies Unknown Healthy  message: |- Failed to load target state: failed to generate manifest for source 1 of 1: rpc error: code = Unknown desc = `kustomize build <path to cached source>/policygentemplates --enable-alpha-plugins` failed exit status 1: 2024/01/24 15:08:10 Could not build the entire policy defined by /tmp/kust-plugin-config-673779590: OperatorHub.yaml is not found both in Git path: <path to cached source>/policygentemplates/source-crs/OperatorHub.yaml and in the ztp container path: /.config/kustomize/plugin/ran.openshift.io/v1/policygentemplate/source-crs/OperatorHub.yaml{code}
 Expected results:
 {code:none}
      openshift-gitops policies sync{code}
 Additional info:
 {code:none}
     {code}
Status: Verified
#OCPBUGS-78171issue9 hours ago[openshift-4.15] ose-machine-api-provider-openstack can't be built hermetically New
Issue 17974839: [openshift-4.15] ose-machine-api-provider-openstack can't be built hermetically
Description: This is a clone of issue OCPBUGS-78170. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78169. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78168. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-machine-api-provider-openstack hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-67013issue4 days agoUnexpected + character in apirequestcounts CRD ASSIGNED
Issue 17653802: Unexpected + character in apirequestcounts CRD
Description: Description of problem:
 {code:none}
 $ oc get --raw /apis/apiserver.openshift.io/v1/apirequestcounts
 Error from server: invalid character '+' looking for beginning of value
 
 $ oc logs kube-apiserver-master-1  kube-apiserver | less
 2025-11-21T13:35:07.009092150Z E1121 13:35:07.009007      18 status.go:71] apiserver received an error that is not an metav1.Status: &json.SyntaxError{msg:"invalid character '+' looking for beginning of value", Offset:3190}: invalid character '+' looking for beginning of value
 
 We used a python script to detect the json syntax stored in etcd for apirequestcount and we found:
 
 Key: /kubernetes.io/apiserver.openshift.io/apirequestcounts/consolesamples.v1.console.openshift.io
 Error: Expecting value: line 1 column 3190 (char 3189)
 Bad Content Snippet: ","requestCount":8},+"byUser":[{"byVerb"
 
 There is an unexpected + character in front of "byUser". After deleting the bad apirequestcount from etcd, the api works{code}
 Version-Release number of selected component (if applicable):
 {code:none}
 4.16{code}
 How reproducible:
 {code:none}
 Unknown{code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: ASSIGNED
#OCPBUGS-77673issue7 days ago[CI Failure] ocp-spoke-assisted-operator-cim-ui-deploy - Build #4427 - Product Code - MCE Backend 500 Errors CLOSED
Issue 17950639: [CI Failure] ocp-spoke-assisted-operator-cim-ui-deploy - Build #4427 - Product Code - MCE Backend 500 Errors
Description: {panel:title=CI Failure Analysis|borderStyle=solid|borderColor=#ccc|titleBGColor=#F7D6C1|bgColor=#FFFFCE} _Job:_ ocp-spoke-assisted-operator-cim-ui-deploy _Build:_ #4427 _Classification:_ Product Code _Environment:_ OCP 4.21.0 Nightly / MCE 2.11.0 / IPv6 Disconnected  h3. Root Cause UI Test timed out waiting for networking fields because the MCE backend returned 500 Internal Server Errors.  h3. Error Details {code} AssertionError: Timed out retrying after 30000ms: Expected to find element: #form-input-clusterNetworks-0-cidr-field {code} The backend returned: {code} Status: 500 {code} {panel}
Status: CLOSED
#OCPBUGS-77976issue40 hours ago[openshift-4.12] ose-cluster-api can't be built hermetically New
Issue 17967571: [openshift-4.12] ose-cluster-api can't be built hermetically
Description: This is a clone of issue OCPBUGS-77975. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77974. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77973. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-cluster-api hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-72595issue4 weeks agoGinkgo test failure: [Apiserver] MicroShift enable SCC admission for pods Verified
Issue 17751602: Ginkgo test failure: [Apiserver] MicroShift enable SCC admission for pods
Description: This is a clone of issue OCPBUGS-72360. The following is the description of the original issue:  --- Description of problem:
 {code:none}
 "[sig-api-machinery] API_Server on Microshift [It] Author:rgangwar-MicroShiftBoth-ConnectedOnly-Medium-55394-[Apiserver] MicroShift enable SCC admission for pods test" failed because the pod is created with a missing fsGroup
 {code}
 Version-Release number of selected component (if applicable):
 {code:none}
     4.21.0-rc.1 on x86. This is not reproducible on ARM architecture.{code}
 How reproducible:
 {code:none}
     always on 4.21.0-rc.1{code}
 Steps to Reproduce:
 {code:none}
     1. Run Ginkgo test case: "[sig-api-machinery] API_Server on Microshift [It] Author:rgangwar-MicroShiftBoth-ConnectedOnly-Medium-55394-[Apiserver] MicroShift enable SCC admission for pods test"
     2.
     3.
     {code}
 Actual results:
 {code:none}
  • [PANICKED] [222.518 seconds]
   [sig-api-machinery] API_Server on Microshift [It] Author:rgangwar-MicroShiftBoth-ConnectedOnly-Medium-55394-[Apiserver] MicroShift enable SCC admission for pods
   /tmp/tmp.I4ekJytyWF/openshift-tests-private/test/extended/apiserverauth/apiserver_microshift.go:480
 
     Timeline >>
     STEP: Creating a kubernetes client @ 01/08/26 04:07:05.353
     I0108 04:07:05.368507 507110 client.go:486] Creating namespace "e2e-test-default-wb9lb"
     I0108 04:07:05.380389 507110 client.go:491] Waiting for ServiceAccount "default" to be provisioned...
     I0108 04:07:05.485060 507110 client.go:495] Configuring kubeconfig with user "e2e-test-default-wb9lb-user" certificates...
     I0108 04:07:07.078970 507110 client.go:293] configPath is now "/tmp/configfile4051781145"
     I0108 04:07:07.079012 507110 client.go:498] Waiting for RoleBinding "e2e-test-default-wb9lb-user" to be provisioned...
     I0108 04:07:07.084506 507110 client.go:502] Waiting on permissions in namespace "e2e-test-default-wb9lb" ...
     I0108 04:07:07.364467 507110 client.go:521] Namespace "e2e-test-default-wb9lb" has been fully provisioned.
     I0108 04:07:07.364710 507110 apiserver_microshift.go:37] The cluster should be healthy before running case.
     namespace/r6c171tq created
     namespace "r6c171tq" deleted
     I0108 04:07:13.076170 507110 apiserver_util.go:1892] Cluster sanity check passed
     STEP: 1.Create temporary namespace @ 01/08/26 04:07:13.076
     STEP: 2. Create one pod security-context-demo-ocp55394 with the specified security context. @ 01/08/26 04:07:13.183
     I0108 04:07:13.183100 507110 apiserver_util.go:460] apiserver fixture dir is not initialized, start to create
     I0108 04:07:13.187939 507110 apiserver_util.go:463] apiserver fixture dir is initialized: /tmp/fixture-testdata-dir2892104936/test/extended/testdata/apiserverauth
     pod/security-context-demo-ocp55394 created
     STEP: 3. Verify that the Pod's security-context-demo-ocp55394 Container is running @ 01/08/26 04:07:13.306
     I0108 04:07:23.409925 507110 pods.go:176] Pod security-context-demo-ocp55394 is ready!
     STEP: 4.1 Verify if the processes are running with the specified user ID 1000 in the pod. @ 01/08/26 04:07:23.41
     I0108 04:07:38.587390 507110 apiserver_util.go:801] Attempting to execute command on pod security-context-demo-ocp55394. Output: PID   USER     TIME  COMMAND
         1 1000      0:00 sleep 1h
         2 1000      0:00 ps, Error: <nil>
     I0108 04:07:38.587444 507110 apiserver_util.go:832] Successfully retrieved non-empty output from pod security-context-demo-ocp55394: PID   USER     TIME  COMMAND
         1 1000      0:00 sleep 1h
         2 1000      0:00 ps
     I0108 04:07:38.587500 507110 apiserver_microshift.go:506] Processes are running on pod security-context-demo-ocp55394 with user id 1000 :: PID   USER     TIME  COMMAND
         1 1000      0:00 sleep 1h
         2 1000      0:00 ps
     STEP: 4.2 Verify that user is running with specified uid=1000 gid=3000 groups=2000 @ 01/08/26 04:07:38.587
     I0108 04:07:53.775700 507110 apiserver_util.go:801] Attempting to execute command on pod security-context-demo-ocp55394. Output: uid=1000(1000) gid=3000 groups=2000,3000, Error: <nil>
     I0108 04:07:53.775738 507110 apiserver_util.go:832] Successfully retrieved non-empty output from pod security-context-demo-ocp55394: uid=1000(1000) gid=3000 groups=2000,3000
     I0108 04:07:53.775826 507110 apiserver_microshift.go:515] On pod security-context-demo-ocp55394 User running with :: uid=1000(1000) gid=3000 groups=2000,3000
     STEP: 5. Create one pod security-context-demo-2-ocp55394 with the specified security context. @ 01/08/26 04:07:53.775
     I0108 04:07:53.775904 507110 apiserver_util.go:466] apiserver fixture dir found in cache: /tmp/fixture-testdata-dir2892104936/test/extended/testdata/apiserverauth
     pod/security-context-demo-2-ocp55394 created
     STEP: 6. Verify that the Pod's security-context-demo-2-ocp55394 Container is running @ 01/08/26 04:07:53.967
     I0108 04:08:04.065851 507110 pods.go:176] Pod security-context-demo-2-ocp55394 is ready!
     STEP: 7. Verify that processes are running with the specified user ID 2000 in the pod. @ 01/08/26 04:08:04.066
     I0108 04:08:19.255042 507110 apiserver_util.go:801] Attempting to execute command on pod security-context-demo-2-ocp55394. Output: USER         PID %CPU %MEM    VSZ   RSS TTY      STAT START   TIME COMMAND
     2000           1  0.0  0.0 705960   768 ?        Ssl  04:07   0:00 /hello_openshift
     2000           6  0.0  0.0   1704   768 ?        Rs   04:08   0:00 ps aux, Error: <nil>
     I0108 04:08:19.255086 507110 apiserver_util.go:832] Successfully retrieved non-empty output from pod security-context-demo-2-ocp55394: USER         PID %CPU %MEM    VSZ   RSS TTY      STAT START   TIME COMMAND
     2000           1  0.0  0.0 705960   768 ?        Ssl  04:07   0:00 /hello_openshift
     2000           6  0.0  0.0   1704   768 ?        Rs   04:08   0:00 ps aux
     I0108 04:08:19.255128 507110 apiserver_microshift.go:534] Processes are running on pod security-context-demo-2-ocp55394 with user id 2000 :: USER         PID %CPU %MEM    VSZ   RSS TTY      STAT START   TIME COMMAND
     2000           1  0.0  0.0 705960   768 ?        Ssl  04:07   0:00 /hello_openshift
     2000           6  0.0  0.0   1704   768 ?        Rs   04:08   0:00 ps aux
     STEP: 8. Ensure that pods in different namespaces are launched with different UIDs. @ 01/08/26 04:08:19.255
     STEP: 8.1 Create two different namespaces. @ 01/08/26 04:08:19.255
     I0108 04:08:19.291544 507110 apiserver_util.go:466] apiserver fixture dir found in cache: /tmp/fixture-testdata-dir2892104936/test/extended/testdata/apiserverauth
     STEP: 8.2 Create pods in both namespaces @ 01/08/26 04:08:19.291
     STEP: 8.3 Verify pods should have different UID's and desc in both namespaces. @ 01/08/26 04:08:19.291
     pod/pod-for-ping created
     I0108 04:08:25.585028 507110 apiserver_util.go:1035] The resource returned:
     pod-for-ping
     I0108 04:08:25.585071 507110 apiserver_util.go:1263] Namespace testpod-namespace-1 pods are: pod-for-ping
     I0108 04:08:35.699575 507110 pods.go:176] Pod pod-for-ping is ready!
     I0108 04:08:41.978615 507110 apiserver_util.go:1035] The resource returned:
     'pod-for-ping runAsUser: 1000580000 fsGroup:  seLinuxOptions: s0:c24,c14
     '
     pod/pod-for-ping created
     I0108 04:08:48.301631 507110 apiserver_util.go:1035] The resource returned:
     pod-for-ping
     I0108 04:08:48.301670 507110 apiserver_util.go:1263] Namespace testpod-namespace-2 pods are: pod-for-ping
     I0108 04:08:58.471873 507110 pods.go:176] Pod pod-for-ping is ready!
     I0108 04:09:06.403000 507110 apiserver_util.go:1035] The resource returned:
     'pod-for-ping runAsUser: 1000590000 fsGroup:  seLinuxOptions: s0:c24,c19
     '
     pod "security-context-demo-2-ocp55394" deleted from test-scc-ocp55394 namespace
     pod "security-context-demo-ocp55394" deleted from test-scc-ocp55394 namespace
     namespace "test-scc-ocp55394" deleted
     [PANICKED] in [It] - /home/ec2-user/go/pkg/mod/golang.org/toolchain@v0.0.1-go1.24.11.linux-amd64/src/runtime/panic.go:115 @ 01/08/26 04:10:47.819
     Jan  8 04:10:47.819: INFO: test dir /tmp/-OCP-microshift-apiseerver-cases-7zer0254/ is cleaned up
     STEP: Collecting events from namespace "e2e-test-default-wb9lb". @ 01/08/26 04:10:47.819
     STEP: Found 0 events. @ 01/08/26 04:10:47.83
     I0108 04:10:47.833385 507110 resource.go:151] POD  NODE  PHASE  GRACE  CONDITIONS
     I0108 04:10:47.833421 507110 resource.go:161] 
     I0108 04:10:47.837817 507110 dump.go:81] skipping dumping cluster info - cluster too large
     I0108 04:10:47.857338 507110 client.go:681] Deleted {certificates.k8s.io/v1, Resource=certificatesigningrequests  e2e-test-default-wb9lb-user}, err: <nil>
     STEP: Destroying namespace "e2e-test-default-wb9lb" for this suite. @ 01/08/26 04:10:47.857
     << Timeline
 
     [PANICKED] Test Panicked
     In [It] at: /home/ec2-user/go/pkg/mod/golang.org/toolchain@v0.0.1-go1.24.11.linux-amd64/src/runtime/panic.go:115 @ 01/08/26 04:10:47.819
 
     runtime error: index out of range [6] with length 6
 
     Full Stack Trace
       github.com/openshift/openshift-tests-private/test/extended/apiserverauth.init.func3.8()
       	/tmp/tmp.I4ekJytyWF/openshift-tests-private/test/extended/apiserverauth/apiserver_microshift.go:590 +0x1bcf
       github.com/onsi/ginkgo/v2/internal.extractBodyFunction.func3({0x0?, 0x0?})
       	/home/ec2-user/go/pkg/mod/github.com/openshift/onsi-ginkgo/v2@v2.6.1-0.20240806135314-3946b2b7b2a8/internal/node.go:472 +0x13
       github.com/onsi/ginkgo/v2/internal.(*Suite).runNode.func3()
       	/home/ec2-user/go/pkg/mod/github.com/openshift/onsi-ginkgo/v2@v2.6.1-0.20240806135314-3946b2b7b2a8/internal/suite.go:901 +0x7b
       created by github.com/onsi/ginkgo/v2/internal.(*Suite).runNode in goroutine 1
       	/home/ec2-user/go/pkg/mod/github.com/openshift/onsi-ginkgo/v2@v2.6.1-0.20240806135314-3946b2b7b2a8/internal/suite.go:888 +0xd7b
   ------------------------------
 
   Summarizing 1 Failure:
     [PANICKED!] [sig-api-machinery] API_Server on Microshift [It] Author:rgangwar-MicroShiftBoth-ConnectedOnly-Medium-55394-[Apiserver] MicroShift enable SCC admission for pods
     /home/ec2-user/go/pkg/mod/golang.org/toolchain@v0.0.1-go1.24.11.linux-amd64/src/runtime/panic.go:115
 
   Ran 1 of 1 Specs in 222.518 seconds
   FAIL! -- 0 Passed | 1 Failed | 0 Pending | 0 Skipped{code}
 Expected results:
 {code:none}
     test pass{code}
 Additional info: 
 
 [     link to ginkgo log|https://gcsweb-ci.apps.ci.l2s4.p1.openshiftapps.com/gcs/test-platform-results/logs/periodic-ci-openshift-microshift-release-4.21-periodics-e2e-aws-tests-release-periodic/2009098095907835904/artifacts/e2e-aws-tests-release-periodic/openshift-microshift-e2e-metal-tests/artifacts/scenario-info/el96-lrel@ginkgo-tests/ginkgo-results/test-output.log]
 {code:none}
     {code}
Status: Verified
#OCPBUGS-77299issue11 days ago[openshift-4.16] openshift-enterprise-hyperkube can't be built hermetically New
Issue 17932658: [openshift-4.16] openshift-enterprise-hyperkube can't be built hermetically
Description: Description of problem:
 ART can't build openshift-enterprise-hyperkube hermetically. Same applies for ose-installer-kube-apiserver-artifacts
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.
      {code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: POST
#OCPBUGS-59807issue8 days agoInspection error on Dell XR8620t ASSIGNED
Issue 17173358: Inspection error on Dell XR8620t
Description: Description of problem:
 {code:none}
     After deprovisioning cluster and deleting BMH for my xr8620t, I recreated the BMH and hit an inspection failure.
 
   errorMessage: 'Failed to inspect hardware. Reason: unable to start inspection: HTTP
     POST https://10.x.x.x/redfish/v1/Systems/System.Embedded.1/VirtualMedia/1/Actions/VirtualMedia.InsertMedia
     returned code 500. Base.1.12.GeneralError: A general error has occurred. See ExtendedInfo
     for more information Extended information: [{''Message'': ''Virtual Media is detached
     or Virtual Media devices are already in use.'', ''MessageArgs'': [], ''MessageArgs@odata.count'':
     0, ''MessageId'': ''IDRAC.2.9.VRM0021'', ''RelatedProperties'': [], ''RelatedProperties@odata.count'':
     0, ''Resolution'': ''Change the Virtual Media Attach Mode to Attach or Auto-Attach.Stop
     existing Virtual Media sessions and retry the operation.'', ''Severity'': ''Informational''}]'
   errorType: inspection error
 
 I've confirmed that there is no virtual media attached. Setting online=true has no effect, and manually powering on the node from idrac ends up with the node quickly being powered off again.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
 OCP: 4.20.0-0.nightly-2025-07-20-021531
 ACM: 2.14.0-131
 MCE: 2.9.0-102{code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: ASSIGNED
{code:java}
{"level":"info","ts":1765324228.08247,"logger":"provisioner.ironic","msg":"starting provisioning","host":"lab2-bmh~node-4","lastError":"Failed to prepare to deploy. Exception: HTTP POST https://10.237.132.17/redfish/v1/Systems/System.Embedded.1/VirtualMedia/1/Actions/VirtualMedia.InsertMedia returned code 500. Base.1.12.GeneralError: A general error has occurred. See ExtendedInfo for more information Extended information: [{'Message': 'Virtual Media is detached or Virtual Media devices are already in use.', 'MessageArgs': [], 'MessageArgs@odata.count': 0, 'MessageId': 'IDRAC.2.9.VRM0021', 'RelatedProperties': [], 'RelatedProperties@odata.count': 0, 'Resolution': 'Change the Virtual Media Attach Mode to Attach or Auto-Attach.Stop existing Virtual Media sessions and retry the operation.', 'Severity': 'Informational'}]","current":"deploy failed","target":"active","deploy step":{}} {code}
 
#OCPBUGS-76930issue12 days agoGCP: Provider produced inconsistent result after apply - google_service_account ASSIGNED
Issue 17884795: GCP: Provider produced inconsistent result after apply - google_service_account
Description: Description of problem:
 {code:none}
     The fix given for issue https://issues.redhat.com/browse/OCPBUGS-42014(pointing to same problem) was at service account key creation. but the actual problem lies at service account creation itself.  
 
 We suggest upgrading terraform and provider to the latest version in openshift-installer for 4.12+. The underlying issue is already fixed upstream and released in v6.45.0.
 
 Note: we see this problem more often in lower openshift versions 4.12, 4.13 than in 4.16.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
 Intermittent{code}
 Steps to Reproduce:
 {code:none}
     1.Install the UPI openshift cluster in GCP platform
     2.
     3.
     {code}
 Actual results:
 {code:none}
     level=error msg=Error: Provider produced inconsistent result after apply
 level=error
 level=error msg=When applying changes to
 level=error msg=module.master.google_service_account.master-node-sa[0], provider
 level=error msg="provider[\"openshift/local/google\"]" produced an unexpected new value: Root
 level=error msg=resource was present, but now absent.
 level=error
 level=error msg=This is a bug in the provider, which should be reported in the provider's own{code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: POST
#OCPBUGS-77308issue7 days ago[openshift-4.22] hermetic build failure - nmstate-console-plugin ON_QA
Issue 17933419: [openshift-4.22] hermetic build failure - nmstate-console-plugin
Description: Description of problem:
 {code:none}
   nmstate-console-plugin fails to build in hermetic mode due to yarn package manager incompatibility and Cypress test framework attempting network downloads during build process.
   {code}
 Version-Release number of selected component (if applicable):
 {code:none}
   OpenShift Container Platform 4.14+
   nmstate-console-plugin component across all affected versions
   {code}
 How reproducible:
 {code:none}
   Always - occurs when converting any nmstate-console-plugin build from network_mode: open to hermetic mode
   {code}
 Steps to Reproduce:
 {code:none}
   1. Remove network_mode: open from nmstate-console-plugin.yml configuration
   2. Attempt hermetic build via Konflux build system
   3. Build fails during npm install phase when Cypress tries to download binaries
   {code}
 Actual results:
 {code:none}
   npm error Installing Cypress (version: 12.17.4)
   npm error The Cypress App could not be downloaded.
   npm error URL: https://download.cypress.io/desktop/12.17.4?platform=linux&arch=x64
   npm error Error: getaddrinfo ENOTFOUND download.cypress.io
 
   Build fails because:
   - Yarn package manager not supported by cachi2 for hermetic dependency management
   - Cypress testing framework requires network access to download.cypress.io
   - Custom yarn artifacts and dockerfile modifications conflict with hermetic requirements
   {code}
 Expected results:
 {code:none}
   nmstate-console-plugin builds successfully in hermetic mode without requiring network access during build process. Component should use npm package manager and disable Cypress binary
   downloads via CYPRESS_INSTALL_BINARY=0 environment variable.
   {code}
 Additional info:
 {code:none}
   Root cause: Component uses yarn package manager which is incompatible with cachi2 hermetic builds, plus Cypress test framework downloads binaries during build.
 
   Solution requires:
   1. Upstream yarn→npm migration (similar to networking-console-plugin)
   2. Set CYPRESS_INSTALL_BINARY=0 to disable test framework downloads
   3. Remove yarn-specific artifacts and dockerfile modifications
   4. Update ocp-build-data package manager configuration
 
   Reference upstream work: https://github.com/openshift/nmstate-console-plugin/pull/169
   Similar successful conversion: networking-console-plugin PRs #347 and #331
 
   Tracking: ART-14616 (parent task with technical analysis)
   {code}
Status: ON_QA
#OCPBUGS-77995issue35 hours ago[openshift-4.13] openshift-enterprise-console-operator can't be built hermetically New
Issue 17968948: [openshift-4.13] openshift-enterprise-console-operator can't be built hermetically
Description: This is a clone of issue OCPBUGS-77994. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77993. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77992. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77862. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77861. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77860. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build openshift-enterprise-console-operator hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-76479issue13 hours agoOpenshift pods stay in a "CreateContainerError" status and show the executable file not found in $PATH" error. New
Issue 17856179: Openshift pods stay in a "CreateContainerError" status and show the executable file not found in $PATH" error.
Description: $ oc get pods
 NAME                                                         READY   STATUS                      RESTARTS   AGE
 cluster-59fd4ffbfd-pvws6                                     1/1     Running                     0          1d
 gitops-plugin-895cc759-wth87                                 1/1     Running                     0          1d
 openshift-gitops-application-controller-0                    0/1     CrashLoopBackOff            121        1d
 openshift-gitops-applicationset-controller-8bb69d85d-xdqzs   0/1     CrashLoopBackOff            311        1d
 openshift-gitops-dex-server-5bf4bf4f8c-vd298                 1/1     Running                     0          20d
 openshift-gitops-dex-server-6876d69597-kl8xp                 0/1     Init:CreateContainerError   0          1d
 openshift-gitops-redis-7848d4b859-t8k2m                      1/1     Running                     0          20d
 openshift-gitops-repo-server-56fdcdd764-w6866                1/1     Running                     0          20d
 openshift-gitops-repo-server-6f4b8564bd-znhd5                0/1     Init:CreateContainerError   0          1d
 openshift-gitops-server-5f87669c76-cfg9w                     0/1     CrashLoopBackOff            120        1d
 
 $ oc get events
 LAST SEEN   TYPE      REASON    OBJECT                                                           MESSAGE
 37m         Normal    Pulled    pod/openshift-gitops-application-controller-0                    Container image "registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:69b28d34a1a6ae55a6fcca5980bf4f6bdf5f901240b411c8a56205027671e4af" already present on machine
 2m51s       Warning   BackOff   pod/openshift-gitops-application-controller-0                    Back-off restarting failed container argocd-application-controller in pod openshift-gitops-application-controller-0_openshift-gitops(c1a7c45c-2772-4010-a2c5-164a18dde3b1)
 42m         Normal    Pulled    pod/openshift-gitops-applicationset-controller-8bb69d85d-xdqzs   Container image "registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:69b28d34a1a6ae55a6fcca5980bf4f6bdf5f901240b411c8a56205027671e4af" already present on machine
 2m48s       Warning   BackOff   pod/openshift-gitops-applicationset-controller-8bb69d85d-xdqzs   Back-off restarting failed container argocd-applicationset-controller in pod openshift-gitops-applicationset-controller-8bb69d85d-xdqzs_openshift-gitops(8105b135-261e-4f08-8e01-864b0ea0ae5d)
 7m          Normal    Pulled    pod/openshift-gitops-dex-server-6876d69597-kl8xp                 Container image "registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:69b28d34a1a6ae55a6fcca5980bf4f6bdf5f901240b411c8a56205027671e4af" already present on machine
 2m30s       Warning   Failed    pod/openshift-gitops-dex-server-6876d69597-kl8xp                 Error: container create failed: executable file `cp` not found in $PATH: No such file or directory
 7m          Normal    Pulled    pod/openshift-gitops-repo-server-6f4b8564bd-znhd5                Container image "registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:69b28d34a1a6ae55a6fcca5980bf4f6bdf5f901240b411c8a56205027671e4af" already present on machine
 2m56s       Warning   Failed    pod/openshift-gitops-repo-server-6f4b8564bd-znhd5                Error: container create failed: executable file `cp` not found in $PATH: No such file or directory
 12m         Normal    Pulled    pod/openshift-gitops-server-5f87669c76-cfg9w                     Container image "registry.redhat.io/openshift-gitops-1/argocd-rhel8@sha256:69b28d34a1a6ae55a6fcca5980bf4f6bdf5f901240b411c8a56205027671e4af" already present on machine
 2m44s       Warning   BackOff   pod/openshift-gitops-server-5f87669c76-cfg9w                     Back-off restarting failed container argocd-server in pod openshift-gitops-server-5f87669c76-cfg9w_openshift-gitops(f9a45aac-1af4-47a2-9397-24dc97364798)
 
 $ oc get pods
 NAME                                        READY   STATUS                 RESTARTS   AGE
 cluster-logging-operator-5bff654c87-64xbd   1/1     Running                0          2d
 instance-dqg5g                              1/1     Running                0          3d
 instance-grktk                              1/1     Running                0          2d
 instance-nkjc5                              0/1     CreateContainerError   0          2d
 instance-nsvpw                              0/1     CreateContainerError   0          2d
 instance-rnbhs                              1/1     Running                0          2d
 
 $ oc get events
 LAST SEEN   TYPE      REASON   OBJECT               MESSAGE
 1m37s       Normal    Pulled   pod/instance-nkjc5   Container image "registry.redhat.io/openshift-logging/vector-rhel9@sha256:fa2cfa2ed336ce105c8dea5bfe0825407e37ef296193ae162f515213fe43c8d5" already present on machine
 56m         Warning   Failed   pod/instance-nkjc5   Error: container create failed: executable file `sh` not found in $PATH: No such file or directory
 2m23s       Normal    Pulled   pod/instance-nsvpw   Container image "registry.redhat.io/openshift-logging/vector-rhel9@sha256:fa2cfa2ed336ce105c8dea5bfe0825407e37ef296193ae162f515213fe43c8d5" already present on machine
 17m         Warning   Failed   pod/instance-nsvpw   Error: container create failed: executable file `sh` not found in $PATH: No such file or directory
 
 Similar to this was also experienced in v4.6 for which bug: [https://bugzilla.redhat.com/show_bug.cgi?id=1950536] and is closed.
 
 The temporary workaround  is to delete all the images from {{/var/lib/containers/storage}} directories and reboot which doesn't helps as a permanent solution.
Status: New
#OCPBUGS-77974issue15 hours ago[openshift-4.14] ose-cluster-api can't be built hermetically ON_QA
Issue 17967569: [openshift-4.14] ose-cluster-api can't be built hermetically
Description: This is a clone of issue OCPBUGS-77973. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-cluster-api hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: Verified
#OCPBUGS-64929issue7 days agocoreos-installer iso kargs show broken on OVE Agent ISO Verified
Issue 17560657: coreos-installer iso kargs show broken on OVE Agent ISO
Description: In the 4.20.2 build, attempts to inspect or modify kernel arguments (kargs) in the generated OVE ISO fail.
 {code:none}
 $ coreos-installer iso kargs show agent-ove.x86_64.iso
 Error: No karg embed areas found; old or corrupted CoreOS ISO image.{code}
 Previously, the appliance build did not use the ISO directly as built, it was unpacked and repacked using osirrox / xorriso. The recent change now uses the ISO produced by the appliance directly, which appears to expose an existing issue in how karg embed areas are handled.
 
 [https://github.com/openshift/agent-installer-utils/commit/6d64a0891291701aa0b38b57c8ac9cfa81a69fe4] 
 
 Version-Release number of selected component (if applicable):
 {code:none}
 4.20.2{code}
 How reproducible:
 {code:none}
 Always  {code}
Status: Verified
#OCPBUGS-77975issue40 hours ago[openshift-4.13] ose-cluster-api can't be built hermetically New
Issue 17967457: [openshift-4.13] ose-cluster-api can't be built hermetically
Description: This is a clone of issue OCPBUGS-77974. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77973. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-cluster-api hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: POST
#OCPBUGS-73848issue6 weeks agooauth-proxy e2e tests fail due to ambiguous container image short names on RHEL 9 Verified
Issue 17769418: oauth-proxy e2e tests fail due to ambiguous container image short names on RHEL 9
Description: The oauth-proxy e2e-component tests fail on OCP 4.22 because RHEL 9 enforces fully qualified container image names ("short name mode").
 
 The backend test image {{nginxdemos/nginx-hello:plain-text}} is rejected by Podman since it cannot determine which registry to pull from.
 h2. Version-Release number
 
 4.22
 h2. How reproducible
 
 Always
 h2. Steps to Reproduce
 
 Run the oauth-proxy e2e-component tests on OCP 4.22
 
 The test creates a pod with a backend container using image {{nginxdemos/nginx-hello:plain-text}}
 
 Observe the pod fails to start.
 h2. Actual results
 
 The backend container fails with:
 {code:java}
 ImageInspectError: short name mode is enforcing, but image name 
 nginxdemos/nginx-hello:plain-text returns ambiguous list
 {code}
 The pod never reaches Running state, and all e2e tests fail.
 h2. Expected results
 
 The e2e tests should pass. The backend container should start successfully using a fully qualified image reference.
 h2. Additional info
 
 _Root cause:_ RHEL 9 enforces "short name mode" for container images, requiring fully qualified registry paths.
 
 _Fix:_ Replace {{nginxdemos/nginx-hello:plain-text}} with {{{}registry.k8s.io/e2e-test-images/agnhost:2.53{}}}, which is:
  * Fully qualified (satisfies RHEL 9 short name policy)
  * Multi-arch (supports ARM64)
  * The standard test image adopted across OpenShift repositories ([openshift/origin PR #29672|https://github.com/openshift/origin/pull/29672])
  * Mirrored in OpenShift CI infrastructure
 
 _PR:_ [https://github.com/openshift/oauth-proxy/pull/343]
Status: Verified
#OCPBUGS-77764issue6 days ago[openshift-4.20] hermetic build failure - node-feature-discovery Verified
Issue 17953456: [openshift-4.20] hermetic build failure - node-feature-discovery
Description: This is a clone of issue OCPBUGS-77763. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77651. The following is the description of the original issue:  --- Description of problem:
 {code:none}
   Node Feature Discovery build fails in hermetic mode due to non-deterministic binary path handling in Dockerfile. The build uses "go install" command which places binaries in unpredictable GOPATH locations, and the Dockerfile copies from multiple inconsistent paths causing binary path conflicts.
   {code}
 Version-Release number of selected component (if applicable):
 {code:none}
   OpenShift Container Platform 4.19+
   node-feature-discovery component across all affected versions
   {code}
 How reproducible:
 {code:none}
   Always - when building node-feature-discovery in hermetic mode
   {code}
 Steps to Reproduce:
 {code:none}
   1. Build node-feature-discovery container in hermetic mode
   2. Dockerfile copies binaries from multiple paths: /go/bin/* and /go/node-feature-discovery/bin/*
   3. "go install" command places binaries in non-deterministic GOPATH locations during hermetic build process
   {code}
 Actual results:
 {code:none}
   Build fails because binaries are placed in unpredictable locations by "go install" and Dockerfile attempts to copy from multiple conflicting paths
 
   Error: Inconsistent binary placement causing copy failures during Docker image build process when switching from "go install" to "go build" in
   hermetic mode
   {code}
 Expected results:
 {code:none}
   Build succeeds using deterministic binary paths. Should use "go build" with explicit output paths instead of "go install" and copy binaries from single consistent location in Dockerfile.
   {code}
 Additional info:
 {code:none}
   Root cause: "go install" command places binaries in GOPATH-dependent locations that are not deterministic in hermetic builds, unlike "go build"
    which allows explicit output path control.
 
   Fix implemented: Remove redundant /go/bin/* copy command from Dockerfile and keep only the specific /go/node-feature-discovery/bin/* path. This
    ensures consistent binary placement and eliminates path conflicts during hermetic builds.
 
   Reference upstream work: https://github.com/openshift/node-feature-discovery/pull/163/changes
   Related hermetic conversion tracking: ART-14586
   {code}
Status: Verified
#OCPBUGS-67364issue7 weeks agoSignatureValidationFailed when creating a pod after applying an image policy with PKI. New
Issue 17670663: SignatureValidationFailed when creating a pod after applying an image policy with PKI.
Description: Description of problem:
 {code:none}
     SignatureValidationFailed when creating a pod after applying an image policy with PKI. 
 The image used is signed with cosign PKI. (refer to https://polarion.engineering.redhat.com/polarion/#/project/OSE/workitem?id=OCP-86101){code}
 Version-Release number of selected component (if applicable):
 {code:none}
     4.21.0-0.nightly-2025-12-10-193741{code}
 How reproducible:
 {code:none}
     Always {code}
 Steps to Reproduce:
 {code:none}
     1. Set crio log level as "debug"
 #oc create -f - << EOF
  apiVersion: machineconfiguration.openshift.io/v1
  kind: ContainerRuntimeConfig
  metadata:
   name: set-loglevel
  spec:
   machineConfigPoolSelector:
     matchLabels:
      pools.operator.machineconfiguration.openshift.io/worker: ""
   containerRuntimeConfig:
     logLevel: debug
  EOF
 
     2. create a user namespace
  % oc new-project testnamespace 
 
     3. create an imagepolicy crd
 apiVersion: config.openshift.io/v1
  kind: ImagePolicy
  metadata:
    name: byopkipolicy   # BYOPKI with Root CA and Intermediate CA
    namespace: testnamespace
  spec:
    scopes:
    - quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3
    policy:
      rootOfTrust:
        policyType: PKI
        pki:
          caRootsData: LS0tLS1CRUdJT....Q0FURS0tLS0t
          caIntermediatesData: LS0tLS1CRUd....JQ0FURS0tLS0t
          pkiCertificateSubject:
             email: minmli@redhat.com
             hostname: myhost.example.com
 
       4. wait a few minutes, check the policy take effect on worker
 sh-5.1# cat /etc/crio/policies/testnamespace.json
 
 sh-5.1# cat /etc/containers/registries.d/sigstore-registries.yaml  
 docker:
    quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3:
      use-sigstore-attachments: true
 
       5. create a pod with the signed image
 #oc apply -f - << EOF
  apiVersion: v1
  kind: Pod
  metadata:
    name: testpod-1
    labels:
      app: testpod
  spec:
    securityContext:
      runAsNonRoot: true
      seccompProfile:
        type: RuntimeDefault
    containers:
    - name: hello-pod
      image: quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3
      securityContext:
        allowPrivilegeEscalation: false
        capabilities:
          drop: ["ALL"]
      ports:
      - containerPort: 80
  EOF
 
          {code}
 Actual results:
 {code:none}
     5 failed to pull image due to error : SignatureValidationFailed
 % oc get pod 
 NAME        READY   STATUS                      RESTARTS   AGE
 testpod-1   0/1     SignatureValidationFailed   0          5s
 
 % oc describe pod testpod-1
 ...
 Events:
   Type     Reason          Age               From     Message
   ----     ------          ----              ----     -------
   Normal   AddedInterface  24s               multus   Add eth0 [10.128.2.22/23] from ovn-kubernetes
   Normal   BackOff         23s               kubelet  Back-off pulling image "quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3"
   Warning  Failed          23s               kubelet  Error: ImagePullBackOff
   Normal   Pulling         9s (x2 over 24s)  kubelet  Pulling image "quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3"
   Warning  Failed          8s (x2 over 24s)  kubelet  Failed to pull image "quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3": SignatureValidationFailed: unable to pull image or OCI artifact: pull image err: Source image rejected: Signature for identity "quay.io/minmli/quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3" is not accepted; artifact err: provided artifact is a container image
   Warning  Failed          8s (x2 over 24s)  kubelet  Error: SignatureValidationFailed
  {code}
 Expected results:
 {code:none}
     5 the image can be pulled successfully and the pod became running{code}
 Additional info:
 {code:none}
     On the pod's node, I can pull image successfully using podman or crictl. 
 sh-5.1# crictl pull quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3
 Image is up to date for quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3
 sh-5.1#     
 sh-5.1# crictl images | grep testpkisignedimage-1
 quay.io/minmli/testpkisignedimage-1              <none>              230662f80b00e       1.39MB
 
 sh-5.1# podman pull quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3
 Trying to pull quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3...
 Getting image source signatures
 Checking if image destination supports signatures
 Copying blob f70adabe43c0 done   | 
 Copying config 230662f80b done   | 
 Writing manifest to image destination
 Storing signatures
 230662f80b00efe44d7f65d4d111f5fe9914f3dad02e92a9120995059f860408
 
 
 The crio logs show: 
 Dec 12 09:04:28 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:04:28.302403078Z" level=info msg="Checking image status: quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3" file="server/image_status.go:31" id=2dca236a-6acc-497f-b2f3-16f5a20e2496 name=/runtime.v1.ImageService/ImageStatus
 Dec 12 09:04:28 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:04:28.30259278Z" level=debug msg="reference \"[overlay@/var/lib/containers/storage+/run/containers/storage:overlay.skip_mount_home=true]quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3\" does not resolve to an image ID" file="storage/storage_reference.go:154"
 Dec 12 09:04:28 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:04:28.302723181Z" level=debug msg="Can't find quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3" file="server/image_status.go:126" id=2dca236a-6acc-497f-b2f3-16f5a20e2496 name=/runtime.v1.ImageService/ImageStatus
 Dec 12 09:04:28 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:04:28.302769481Z" level=info msg="Image quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3 not found" file="server/image_status.go:145" id=2dca236a-6acc-497f-b2f3-16f5a20e2496 name=/runtime.v1.ImageService/ImageStatus
 Dec 12 09:04:28 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:04:28.302860582Z" level=info msg="Neither image nor artfiact quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3 found" file="server/image_status.go:47" id=2dca236a-6acc-497f-b2f3-16f5a20e2496 name=/runtime.v1.ImageService/ImageStatus
 ...
 Dec 12 09:08:38 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:08:38.708198128Z" level=info msg="Pulling image: quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3" file="server/image_pull.go:44" id=5f889081-5de4-4b13-b48e-1a4a2e2540ce name=/runtime.v1.ImageService/PullImage
 Dec 12 09:08:38 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:08:38.708241229Z" level=debug msg="Using pull policy path for image quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3: \"\"" file="server/image_pull.go:147" id=5f889081-5de4-4b13-b48e-1a4a2e2540ce name=/runtime.v1.ImageService/PullImage
 Dec 12 09:08:38 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:08:38.750278117Z" level=info msg="Trying to access \"quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3\"" file="docker/docker_image_src.go:96"
 Dec 12 09:08:38 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:08:38.750438218Z" level=debug msg="Found credentials for quay.io/minmli/testpkisignedimage-1 in credential helper containers-auth.json in file /var/lib/kubelet/config.json" file="config/config.go:263"
 Dec 12 09:08:38 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:08:38.750478118Z" level=debug msg=" Lookaside configuration: using \"docker\" namespace quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3" file="docker/registries_d.go:210"
 Dec 12 09:08:38 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:08:38.750513919Z" level=debug msg=" No signature storage configuration found for quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3, using built-in default file:///var/lib/containers/sigstore" file="docker/registries_d.go:183"
 Dec 12 09:08:38 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:08:38.750597619Z" level=debug msg=" Sigstore attachments: using \"docker\" namespace quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3" file="docker/registries_d.go:243"
 Dec 12 09:08:38 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:08:38.771852216Z" level=debug msg="GET https://quay.io/v2/auth?account=openshift-release-dev%2Bart_dev_priv_visitor_pull&scope=repository%3Aminmli%2Ftestpkisignedimage-1%3Apull&service=quay.io" file="docker/docker_client.go:845"
 Dec 12 09:08:38 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:08:38.919422177Z" level=debug msg="GET https://quay.io/v2/minmli/testpkisignedimage-1/manifests/sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3" file="docker/docker_client.go:616"
 Dec 12 09:08:39 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:08:39.462189484Z" level=debug msg="IsRunningImageAllowed for image docker:quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3" file="signature/policy_eval.go:274"
 Dec 12 09:08:39 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:08:39.462468686Z" level=debug msg="Downloading /v2/minmli/testpkisignedimage-1/blobs/sha256:230662f80b00efe44d7f65d4d111f5fe9914f3dad02e92a9120995059f860408" file="docker/docker_client.go:1051"
 Dec 12 09:08:39 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:08:39.462543987Z" level=debug msg="GET https://quay.io/v2/minmli/testpkisignedimage-1/blobs/sha256:230662f80b00efe44d7f65d4d111f5fe9914f3dad02e92a9120995059f860408" file="docker/docker_client.go:616"
 Dec 12 09:08:39 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:08:39.598953145Z" level=debug msg="Reading /var/lib/containers/sigstore/minmli/testpkisignedimage-1@sha256=7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3/signature-1" file="docker/docker_image_src.go:543"
 Dec 12 09:08:39 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:08:39.599085247Z" level=debug msg="Looking for sigstore attachments in quay.io/minmli/testpkisignedimage-1:sha256-7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3.sig" file="docker/docker_client.go:1142"
 Dec 12 09:08:39 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:08:39.599138947Z" level=debug msg="GET https://quay.io/v2/minmli/testpkisignedimage-1/manifests/sha256-7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3.sig" file="docker/docker_client.go:616"
 Dec 12 09:08:39 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:08:39.678931183Z" level=debug msg="Downloading /v2/minmli/testpkisignedimage-1/blobs/sha256:7381d1e6c50aa97c5934e473e226db9a7c159754f7ad0d578f4e64c2852b395c" file="docker/docker_client.go:1051"
 Dec 12 09:08:39 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:08:39.678972183Z" level=debug msg="GET https://quay.io/v2/minmli/testpkisignedimage-1/blobs/sha256:7381d1e6c50aa97c5934e473e226db9a7c159754f7ad0d578f4e64c2852b395c" file="docker/docker_client.go:616"
 Dec 12 09:08:39 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:08:39.820416688Z" level=debug msg="Downloading /v2/minmli/testpkisignedimage-1/blobs/sha256:f70adabe43c0cccffbae8785406d490e26855b8748fc982d14bc2b20c778b929" file="docker/docker_client.go:1051"
 Dec 12 09:08:39 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:08:39.820485589Z" level=debug msg="GET https://quay.io/v2/minmli/testpkisignedimage-1/blobs/sha256:f70adabe43c0cccffbae8785406d490e26855b8748fc982d14bc2b20c778b929" file="docker/docker_client.go:616"
 Dec 12 09:08:40 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:08:40.178773594Z" level=debug msg="added name \"quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3\" to image \"230662f80b00efe44d7f65d4d111f5fe9914f3dad02e92a9120995059f860408\"" file="storage/storage_dest.go:1624"
 Dec 12 09:08:40 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:08:40.179327099Z" level=info msg="Pulled image: quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3" file="server/image_pull.go:126" id=5f889081-5de4-4b13-b48e-1a4a2e2540ce name=/runtime.v1.ImageService/PullImage
 Dec 12 09:08:40 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:08:40.1794169Z" level=debug msg="Response: *v1.PullImageResponse: image_ref:\"quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3\"" file="interceptors/interceptors.go:76" id=5f889081-5de4-4b13-b48e-1a4a2e2540ce name=/runtime.v1.ImageService/PullImage
 ...
 Dec 12 09:08:52 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:08:52.317771557Z" level=debug msg="IsRunningImageAllowed for image docker:quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3" file="signature/policy_eval.go:274"
 Dec 12 09:08:52 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:08:52.317809258Z" level=debug msg=" Using transport \"docker\" policy section \"quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3\"" file="signature/policy_eval.go:143"
 Dec 12 09:08:52 minmli-121201-grdms-worker-eastus-1 crio[2538]: time="2025-12-12T09:08:52.319713675Z" level=debug msg="Response error: checking signature of \"quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3\": verifying signatures: SignatureValidationFailed: Signature for identity \"quay.io/minmli/quay.io/minmli/testpkisignedimage-1@sha256:7fe37303835533b708fd69c8a4aa1b48b532917561ad545c290c9297791338e3\" is not accepted" file="interceptors/interceptors.go:73" id=91708134-7f8b-4453-9f09-5e881b9fc1ca name=/runtime.v1.RuntimeService/CreateContainer
 {code}
Status: New
#OCPBUGS-77007issue7 days ago[4.19] openshift-cluster-csi-drivers vmware-vsphere-csi-driver-operator-metrics endpoint 8445 port should return 401 instead 500 Verified
Issue 17895615: [4.19] openshift-cluster-csi-drivers vmware-vsphere-csi-driver-operator-metrics endpoint 8445 port should return 401 instead 500
Description: This is a clone of issue OCPBUGS-75912. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-74569. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-60159. The following is the description of the original issue:  --- Description of problem:
 
 see similar bug OCPBUGS-57585
 
 Stole the testing result from [https://github.com/openshift/origin/pull/30014#issuecomment-3153130188]
 
 8445 port is for openshift-cluster-csi-drivers vmware-vsphere-csi-driver-operator-metrics endpoint, did not see this port on other IAAS(example AWS,OSP)checked on vsphere, curl 8445 without auth returns 500(actually it's forbidden),
 {noformat}
 $ oc get infrastructure/cluster -o jsonpath={.spec.platformSpec.type}
 VSphere
 
 $ oc -n openshift-cluster-csi-drivers get ep
 Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice
 NAME                                           ENDPOINTS                                                        AGE
 vmware-vsphere-csi-driver-controller-metrics   10.129.0.17:9203,10.130.0.21:9203,10.129.0.17:9202 + 9 more...   28m
 vmware-vsphere-csi-driver-operator-metrics     10.130.0.17:8445                                                 28m
 vmware-vsphere-csi-driver-webhook-svc          10.128.0.45:8443,10.130.0.23:8443                                28m
 
 $ oc -n openshift-monitoring exec -c prometheus prometheus-k8s-0 -- curl -k -s -o /dev/null -w '%{http_code}' https://10.130.0.17:8445/metrics
 500
 {noformat}
 return code is 500, expecting 401 forbidden instead of 500.
 {code:java}
 $ oc -n openshift-monitoring exec -c prometheus prometheus-k8s-0 -- curl -k https://10.130.0.17:8445/metrics
 
 Internal Server Error: "/metrics": subjectaccessreviews.authorization.k8s.io is forbidden: User "system:serviceaccount:openshift-cluster-csi-drivers:vmware-vsphere-csi-driver-operator" cannot create resource "subjectaccessreviews" in API group "authorization.k8s.io" at the cluster scope  {code}
 curl with auth, no issue
 {noformat}
 $ token=`oc create token prometheus-k8s -n openshift-monitoring`
 $ oc -n openshift-monitoring exec -c prometheus prometheus-k8s-0 -- curl -k -H "Authorization: Bearer $token" 'https://10.130.0.17:8445/metrics' | head -n3
 # HELP aggregator_discovery_aggregation_count_total [ALPHA] Counter of number of times discovery was aggregated
 # TYPE aggregator_discovery_aggregation_count_total counter
 aggregator_discovery_aggregation_count_total 0 {noformat}
 [https://github.com/openshift/origin/pull/30014#issuecomment-3154900379]
 
  
 
 When the bug is fixed, please remove it from this code snippet to avoid regression.
 
 [https://github.com/openshift/origin/blob/4f183dd3427cffd8d97b44557caa782d65726416/test/extended/prometheus/prometheus.go#L64-L71]
Status: Verified
#OCPBUGS-62010issue8 days agoBMC F/W update fails on Dell R740, server stuck in power off ASSIGNED
Issue 17410716: BMC F/W update fails on Dell R740, server stuck in power off
Description: h4. Description of problem:
 
 Follow on from OCPBUGS-62009.
 
 I deleted the BMH and restarted the metal3 pods to recover. 
 
 I then reprovisioned the BMH and waited for inspection to finish. 
 
 I then updated the hostfirmwarecomponents to just update the BMH. 
 
 It actually did update the firmware on the server but metal3 failed to recover the server.
 
 apiVersion: metal3.io/v1alpha1
 kind: HostFirmwareComponents
 metadata:
   creationTimestamp: "2025-09-21T23:12:18Z"
   generation: 2
   name: r740xdg1
   namespace: r740-pool
   ownerReferences:
   - apiVersion: metal3.io/v1alpha1
     kind: BareMetalHost
     name: r740xdg1
     uid: 9494c650-ad12-4a75-a304-3e9a91b90a1c
   resourceVersion: "18510819"
   uid: d1077d36-c7cd-4647-a4d1-8c7eab631e5e
 spec:
   updates:
   - component: bmc
     url: http://hv14.telco5gran.eng.rdu2.redhat.com:8888/firmware/r740/iDRAC-with-Lifecycle-Controller_Firmware_XTFXJ_WN64_7.00.00.173_A00.EXE
 status:
   components:
   - component: bios
     currentVersion: 2.22.2
     initialVersion: 2.22.2
   - component: bmc
     currentVersion: 7.00.00.181
     initialVersion: 7.00.00.181
   - component: nic:NIC.Integrated.1
     currentVersion: 14.32.20.04
     initialVersion: 14.32.20.04
   conditions:
   - lastTransitionTime: "2025-09-21T23:12:19Z"
     message: ""
     observedGeneration: 2
     reason: OK
     status: "True"
     type: Valid
   - lastTransitionTime: "2025-09-21T23:26:16Z"
     message: ""
     observedGeneration: 2
     reason: OK
     status: "False"
     type: ChangeDetected
   lastUpdated: "2025-09-21T23:26:16Z"
   updates:
   - component: bmc
     url: http://hv14.telco5gran.eng.rdu2.redhat.com:8888/firmware/r740/iDRAC-with-Lifecycle-Controller_Firmware_XTFXJ_WN64_7.00.00.173_A00.EXE
 
 The ser ver is stuck in power off and is powered off again even if I manually power cycle it.
 
 The following was seen in the logs 
 
 {"level":"info","ts":1758498122.6741993,"logger":"provisioner.ironic","msg":"current provision state","host":"r740-pool~r740xdg1","lastError":"Node ea1c7224-e159-4201-84b0-c28cad19046b failed step \{'args': {'settings': [{'component': 'bmc', 'url': 'http://hv14.telco5gran.eng.rdu2.redhat.com:8888/firmware/r740/iDRAC-with-Lifecycle-Controller_Firmware_XTFXJ_WN64_7.00.00.173_A00.EXE'}]}, 'interface': 'firmware', 'step': 'update', 'abortable': False, 'priority': 0}: Redfish exception occurred. Error: HTTP POST https://10.6.36.10/redfish/v1/SessionService/Sessions returned code 401. Base.1.12.GeneralError: Unable to complete the operation because an invalid username and/or password is entered, and therefore authentication failed. Extended information: [\{'Message': 'Unable to complete the operation because an invalid username and/or password is entered, and therefore authentication failed.', 'MessageArgs': [], 'MessageArgs@odata.count': 0, 'MessageId': 'IDRAC.2.8.SYS415', 'RelatedProperties': [], 'RelatedProperties@odata.count': 0, 'Resolution': 'Enter valid user name and password and retry the operation.', 'Severity': 'Warning'}]","current":"manageable","target":""}
 
 Credentials are obviously ok as we would have not gotten this far with invalid ones.
 
 BMC logs 
 
     2025-09-22 00:06:51     LOG007     The previous log entry was repeated 1 times.    
         2025-09-21 23:41:18     RAC0720     Unable to locate the ISO or IMG image file or folder in the network share location because the file or folder path or the user credentials entered are incorrect.    
         2025-09-21 23:41:17     USR0030     Successfully logged in using telemetry, from 127.0.0.1 and REDFISH.    
         2025-09-21 23:41:15     RAC0717     Remote share unmounted successfully.    
         2025-09-21 23:40:56     USR0031     Unable to log in for NULL from 10.22.88.128 using eHTML5 Virtual Console.    
         2025-09-21 23:40:56     LOG007     The previous log entry was repeated 1 times.    
         2025-09-21 23:40:49     USR0030     Successfully logged in using root, from 10.8.53.44 and REDFISH.    
         2025-09-21 23:40:37     DIS002     Auto Discovery feature disabled.    
         2025-09-21 23:40:37     RAC0182     The iDRAC firmware was rebooted with the following reason: user initiated.    
         2025-09-21 23:40:27     IPA0100     The iDRAC IP Address changed from :: to 2620:52:9:1624:f602:70ff:fee4:f7f4.    
         2025-09-21 23:40:27     IPA0100     The iDRAC IP Address changed from 0.0.0.0 to 10.6.36.10.    
         2025-09-21 23:40:07     PR36     Version change detected for Lifecycle Controller firmware. Previous version:7.00.00.181, Current version:7.00.00.173    
         2025-09-21 23:40:02     THRM0008     The UNC Warning threshold limit of the server board inlet temperature sensor is changed to 38.    
         2025-09-21 23:39:58     PSU0800     Power Supply 2: Status = 0x1, IOUT = 0x0, VOUT= 0x0, TEMP= 0x0, FAN = 0x0, INPUT= 0x0.    
         2025-09-21 23:39:58     PSU0800     Power Supply 1: Status = 0x1, IOUT = 0x0, VOUT= 0x0, TEMP= 0x0, FAN = 0x0, INPUT= 0x0.    
         2025-09-21 23:37:38     USR0032     The session for root from 10.8.53.44 using REDFISH is logged off.    
         2025-09-21 23:37:36     SYS1001     System is turning off.    
         2025-09-21 23:37:36     SYS1003     System CPU Resetting.    
         2025-09-21 23:37:29     JCP037     The (installation or configuration) job JID_585155450820 is successfully completed.    
         2025-09-21 23:37:29     RED063     The iDRAC firmware updated successfully. Previous version: 7.00.00.181, Current version: 7.00.00.173    
         2025-09-21 23:37:29     RAC0704     Requested system powerdown.    
         2025-09-21 23:37:26     SUP1906     Firmware update successful.    
         2025-09-21 23:36:34     SUP1905     Firmware update programming flash.    
         2025-09-21 23:36:18     SUP1903     Firmware update verify image headers.    
         2025-09-21 23:36:18     SUP1904     Firmware update checksumming image.    
         2025-09-21 23:36:17     SUP1911     Firmware update initialization complete.    
         2025-09-21 23:36:17     SUP1901     Firmware update initializing.    
         2025-09-21 23:34:34     USR0032     The session for root from 10.8.53.44 using REDFISH is logged off.    
         2025-09-21 23:33:23     RED002     Package successfully downloaded.    
         2025-09-21 23:32:46     RED111     Successfully downloaded the update package details 228.097 MB in 16.9493 secs at 13.4576 MBps (107.661 Mbps) [iDRAC-with-Lifecycle-Controller_Firmware_XTFXJ_WN64_7.00.00.173_A00.EXE].    
         2025-09-21 23:32:29     RED110     Downloading the iDRAC-with-Lifecycle-Controller_Firmware_XTFXJ_WN64_7.00.00.173_A00.EXE update package.    
         2025-09-21 23:32:25     JCP027     The (installation or configuration) job JID_585155450820 is successfully created on iDRAC.    
         2025-09-21 23:29:01     CTL129     The boot media of the Controller RAID Controller in Slot 6 is Disk.Virtual.0:RAID.Slot.6-1.    
         2025-09-21 23:26:06     SYS1003     System CPU Resetting.    
         2025-09-21 23:25:43     SYS1000     System is turning on.    
         2025-09-21 23:25:42     RAC0701     Requested system powerup.
 
 Must Gather: [https://drive.google.com/file/d/1TlEg50A2NqIU8gYCBtwxg_MT1P6Jb8iu/view?usp=drive_link]
 
 if this is the same root cause as  OCPBUGS-62009,  feel free to mark as a dup.
 h4. Version-Release number of selected component (if applicable):
 
 4.20-rc.2 
 h4. How reproducible:
 
 Have not had a BMC upgrade complete on rc2 yet 
 h4. Steps to Reproduce:
  #  
  #  
  # ...
 
 h4. Actual results:
 h4. Expected results:
 h4. Additional info:
Status: ASSIGNED
#OCPBUGS-76988issue5 days agoBMH stuck in deleting state while performing IBBF Verified
Issue 17893050: BMH stuck in deleting state while performing IBBF
Description: This is a clone of issue OCPBUGS-59133. The following is the description of the original issue:  --- Description of problem:
 {code:none}
     While attempting to perform IBBF with the siteconfig operator, the BMH for the spoke is stuck in deleting state for too long due to the old BMH being unreachable. This results in the clusterinstance timing out while waiting for the BMH resource to be removed. We should find a way to mitigate this.
 {code}
     Version-Release number of selected component (if applicable):
 {code:java}
 4.18.0-0.nightly-2025-04-02-011053 {code}
 How reproducible:
 {code:none}
     100%{code}
 Steps to Reproduce:
 {code:none}
     1. Install an IBI spoke cluster using siteconfig operator
     2. Remove old node and attempt to replace with new node through IBBF
     3.
     {code}
 Actual results:
 {code:none}
     Reinstallation fails due to timeout{code}
 Expected results:
 {code:none}
     Reinstallation succeeds{code}
 Additional info:
 {code:none}
 $ oc get clusterinstance target-0 -o json | jq -r '.status.reinstall.conditions[0]'
 {
   "lastTransitionTime": "2025-04-04T17:24:45Z",
   "message": "Encountered error executing task: Deleting rendered manifests. Error: deletion timeout exceeded for object (BareMetalHost:target-0/target-0-0): Timed out waiting to delete object (BareMetalHost:target-0/target-0-0)",
   "reason": "Failed",
   "status": "False",
   "type": "ReinstallRequestProcessed"
 }
 $ oc get bmh target-0-0 -o json | jq -r .status.errorMessage
 Failed to get power state for node 7e063860-56c7-45ca-8bfa-d8f63997bd10. Error: Redfish exception occurred. Error: Resource https://[fd2e:6f44:5dd8:4::1]:8000/redfish/v1/Systems/ac8d214a-ae10-46f5-a179-208341bcb688 not found {code}
Status: Verified
#OCPBUGS-78114issue16 hours ago[openshift-4.12] ose-agent-installer-api-server can't be built hermetically POST
Issue 17972393: [openshift-4.12] ose-agent-installer-api-server can't be built hermetically
Description: This is a clone of issue OCPBUGS-78113. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78112. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78111. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-agent-installer-api-server hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: POST
#OCPBUGS-77028issue6 days ago[4.19] Agent service is failing with input/output error CLOSED
Issue 17900200: [4.19] Agent service is failing with input/output error
Description: *Description of the problem:*
 
 - Unable to discover hosts using BMC because the agent service is failing to start with this error:
 
 Error: netavark: failed to start process: fork/exec /usr/libexec/podman/netavark: input/output error
 
 
 *How reproducible:* Always
 
 *Steps to reproduce:*
 
 1. Boot a system using BMC/BMH
 
 2. ssh the node and check the agent service logs
 
 *Actual results:*
 The agent service fails 
  
 
 *Expected results:*
 The Agent service should start and the hosts should get discovered.
 
 *Additional Info:*
 
 This issue is not faced when:
 
 1. Hosts are discovered using a discovery iso
 2. When we disable converged_flow
Status: CLOSED
#OCPBUGS-57118issue8 days agocertified-operators are failing regularly due to startup probe timing out frequently and generating alert for KubePodCrashLooping New
Comment 28008468 by Jian Zhang at 2025-09-08T09:33:59.903+0000
{quote}4h14m       Warning   Failed           pod/redhat-operators-rr59t      Error: ImagePullBackOff
2h31m       Warning   Failed           pod/redhat-operators-rr59t      Failed to pull image "registry.redhat.io/redhat/redhat-operator-index:v4.16": copying system image from manifest list: reading signatures: reading signature from [https://registry.redhat.io/containers/sigstore/redhat/redhat-operator-index@sha256=XXX/signature-1:] status 503 (Service Unavailable)
#OCPBUGS-76929issue2 days agoGCP: Provider produced inconsistent result after apply - google_service_account Verified
Issue 17884788: GCP: Provider produced inconsistent result after apply - google_service_account
Description: Description of problem:
 {code:none}
     The fix given for issue https://issues.redhat.com/browse/OCPBUGS-42014(pointing to same problem) was at service account key creation. but the actual problem lies at service account creation itself.  
 
 We suggest upgrading terraform and provider to the latest version in openshift-installer for 4.12+. The underlying issue is already fixed upstream and released in v6.45.0.
 
 Note: we see this problem more often in lower openshift versions 4.12, 4.13 than in 4.16.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
 Intermittent{code}
 Steps to Reproduce:
 {code:none}
     1.Install the UPI openshift cluster in GCP platform
     2.
     3.
     {code}
 Actual results:
 {code:none}
     level=error msg=Error: Provider produced inconsistent result after apply
 level=error
 level=error msg=When applying changes to
 level=error msg=module.master.google_service_account.master-node-sa[0], provider
 level=error msg="provider[\"openshift/local/google\"]" produced an unexpected new value: Root
 level=error msg=resource was present, but now absent.
 level=error
 level=error msg=This is a bug in the provider, which should be reported in the provider's own{code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: Verified
#OCPBUGS-77768issue6 days ago[openshift-4.12] hypershift can't be built hermetically New
Issue 17953646: [openshift-4.12] hypershift can't be built hermetically
Description: Description of problem:
 
 ART can't build hypershift hermetically.
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.    {code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#TRT-2568issueSerial and techpreview-serial jobs failing with Prometheus auth check on CCO metrics endpoint in 4.22 nightly amd64 payload CLOSED
Issue 17966625: Serial and techpreview-serial jobs failing with Prometheus auth check on CCO metrics endpoint in 4.22 nightly amd64 payload
Description: h2. Payload Regression  PR https://github.com/openshift/cloud-credential-operator/pull/976 is causing blocking job failures in the nightly amd64 payload.  h2. Evidence * Payload: [4.22.0-0.nightly-2026-03-08-004901|https://amd64.ocp.releases.ci.openshift.org/releasestream/4.22.0-0.nightly/release/4.22.0-0.nightly-2026-03-08-004901] * Failing blocking jobs: ** [aws-ovn-serial-1of2|https://prow.ci.openshift.org/view/gs/test-platform-results/logs/periodic-ci-openshift-release-main-nightly-4.22-e2e-aws-ovn-serial-1of2/2030491767962341376] ** [aws-ovn-serial-2of2|https://prow.ci.openshift.org/view/gs/test-platform-results/logs/periodic-ci-openshift-release-main-nightly-4.22-e2e-aws-ovn-serial-2of2/2030494424521248768] ** [aws-ovn-techpreview-serial-1of3|https://prow.ci.openshift.org/view/gs/test-platform-results/logs/periodic-ci-openshift-release-main-ci-4.22-e2e-aws-ovn-techpreview-serial-1of3/2030513102306414592] ** [aws-ovn-techpreview-serial-2of3|https://prow.ci.openshift.org/view/gs/test-platform-results/logs/periodic-ci-openshift-release-main-ci-4.22-e2e-aws-ovn-techpreview-serial-2of3/2030495813175611392] ** [aws-ovn-techpreview-serial-3of3|https://prow.ci.openshift.org/view/gs/test-platform-results/logs/periodic-ci-openshift-release-main-ci-4.22-e2e-aws-ovn-techpreview-serial-3of3/2030495769554849792] ** [overall-analysis-all|https://prow.ci.openshift.org/view/gs/test-platform-results/logs/periodic-ci-openshift-release-main-nightly-4.22-overall-analysis-all/2030574039981887488] (downstream) * Originating payload: 4.22.0-0.nightly-2026-03-07-161029 (first rejected after last accepted 4.22.0-0.nightly-2026-03-07-091925) * Test failure: {{[sig-instrumentation][Late] Platform Prometheus targets should not be accessible without auth}} * Error: {{failed to ensure scraping target of pod openshift-cloud-credential-operator/cco-metrics requires auth: context deadline exceeded}} * The PR removed kube-rbac-proxy from the CCO metrics endpoint and reimplemented RBAC via controller-runtime. The Prometheus auth conformance test detects that the CCO metrics endpoint no longer requires authentication properly. * Revert PR: https://github.com/openshift/cloud-credential-operator/pull/982  h2. Action  Revert PR https://github.com/openshift/cloud-credential-operator/pull/976 to restore payload acceptance.
Status: CLOSED
#OCPBUGS-32745issue2 weeks ago2 minutes time out degrades machineconfignode resources when pinning release images New
Issue 15961872: 2 minutes time out degrades machineconfignode resources when pinning release images
Description: Description of problem:{code:none}
 
     {code}
     Version-Release number of selected component (if applicable):{code:none}
 
     {code}
     How reproducible:{code:none}
 
     {code}
     Steps to Reproduce:{code:none}
     1Description of problem:{code:none}
 If all the images in the pinnedimageset are not pulled before 2 minutes the machineconfignode is degraded
 
 
   - lastTransitionTime: "2024-04-23T10:30:46Z"
     message: 'Error: requeue: prefetching images incomplete after: 2m0s'
     reason: PrefetchFailed
     status: "True"
     type: PinnedImageSetsDegraded
 
     {code}
     Version-Release number of selected component (if applicable):{code:none}
     pre-merge https://github.com/openshift/machine-config-operator/pull/4303
     {code}
     How reproducible:{code:none}
 Always
     {code}
     Steps to Reproduce:{code:none}
     1.  execute this command to fill in the following pinnedimageset
 
 
 $ oc adm release info   -o pullspec | awk '{print "   - name: " $1}'
    - name: registry.build03.ci.openshift.org/ci-ln-d30q8sk/stable@sha256:7b1b1255645d6fe7a...
    - name: registry.build03.ci.openshift.org/ci-ln-d30q8sk/stable@sha256:42141b60a0ecaaab...
 ....
 
 
 
 apiVersion: machineconfiguration.openshift.io/v1alpha1
 kind: PinnedImageSet
 metadata:
   [?4ms:
     machineconfiguration.openshift.io/role: worker
   name: my-worker-pinned-images-all
 spec:
   pinnedImages:
    - name: registry.build03.ci.openshift.org/ci-ln-d30q8sk/stable@sha256:7b1b1255645d6fe7a...
    - name: registry.build03.ci.openshift.org/ci-ln-d30q8sk/stable@sha256:42141b60a0ecaaab...
 ......
 
     {code}
     Actual results:{code:none}
 The pinnedimageset will take longer than 2 minutes and the machineconfignodes will be degraded.
 
 
     {code}
     Expected results:{code:none}
 It is highly unlikely that MCO will be able to pull all the images in a relase in less than 2 minutes, so this timeout will always degrade MCO when we try to pin the release images.
 
 We expect MCO to be able to pull all the images in a relase without being degraded. 
 
     {code}
     Additional info:{code:none}
 This is already planned to increase the timeout in follow-up PRs. This issue will not block the merge. The main goal of this ticket is not to forget that change and it's verification.
 
 
     {code}
     
     2.
     3.
     {code}
     Actual results:{code:none}
 
     {code}
     Expected results:{code:none}
 
     {code}
     Additional info:{code:none}
 
     {code}
Status: New
#OCPBUGS-78013issue34 hours ago[openshift-4.13] sriov-cni can't be build hermetically POST
Issue 17969334: [openshift-4.13] sriov-cni can't be build hermetically
Description: This is a clone of issue OCPBUGS-78011. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build sriov-cni hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: POST
#OCPBUGS-72604issue7 weeks ago"Observe - Metrics" unable to show metrics results Verified
Issue 17752877: "Observe - Metrics" unable to show metrics results
Description: Description of problem:
 
 checked with 4.22.0-0.nightly-2026-01-11-222452 on  chrome Version 143.0.7499.169 (Official Build) (64-bit), admin user goes to "Observe - Metrics", search metrics cluster_infrastructure_provider first, it can show results, search another metrics, example prometheus_build_info, the result page is empty, see screen recording(sometimes always see the error for any metrics): 
 [https://drive.google.com/file/d/1LhRIueFTfsI1jG3a7f2ptO-Ap7MAqRN_/view?usp=drive_link]
 
 error is [https://drive.google.com/file/d/1O0RZT5qlwylTQaBkN1O3ylgdH37t1hjZ/view?usp=drive_link]
 {code:none}
 5643-chunk-6a634255311d2aaad42a.min.js:1 Caught error in a child component: Error: Minified React error #301; visit https://reactjs.org/docs/error-decoder.html?invariant=301 for the full message or use the non-minified dev environment for full errors and additional helpful warnings.
     at sa (react-dom.production.min.js:157:181)
     at Va (react-dom.production.min.js:180:154)
     at Go (react-dom.production.min.js:269:343)
     at Ec (react-dom.production.min.js:250:347)
     at Sc (react-dom.production.min.js:250:278)
     at xc (react-dom.production.min.js:250:138)
     at gc (react-dom.production.min.js:243:163)
     at react-dom.production.min.js:123:115
     at t.unstable_runWithPriority (scheduler.production.min.js:18:343)
     at Gi (react-dom.production.min.js:122:325) {componentStack: '\n    at Ce (https://console-openshift-console.apps...9a68b1aa4dc782775.min.js:2:29335)\n    at Suspense'}{code}
 this is a 4.22 regression, no issue with 4.21 and the latest accepted 4.22 build 4.22.0-0.nightly-2026-01-06-164201
 
 Version-Release number of selected component (if applicable):
 {code:none}
 4.22 only{code}
 How reproducible:
 {code:none}
 always{code}
 Steps to Reproduce:
 {code:none}
 1. see the description{code}
 Actual results:
 {code:none}
 "Observe - Metrics" unable to show metrics results    {code}
 Expected results:
 {code:none}
 show results{code}
Status: Verified
#OCPBUGS-77992issue35 hours ago[openshift-4.16] openshift-enterprise-console-operator can't be built hermetically New
Issue 17968937: [openshift-4.16] openshift-enterprise-console-operator can't be built hermetically
Description: This is a clone of issue OCPBUGS-77862. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77861. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77860. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build openshift-enterprise-console-operator hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-14473issue5 weeks agoCannot read properties of undefined (reading 'description') error when visiting Overview page ASSIGNED
Issue 15295508: Cannot read properties of undefined (reading 'description') error when visiting Overview page
Description: Description of problem:
 {code:none}
 visit Home -> Overview page, keep the window open for a long time, then we can see an error page displayed
 
 We have seen this error several times, just not sure what is exactly the trigger{code}
 Version-Release number of selected component (if applicable):
 {code:none}
 4.14.0-0.nightly-2023-05-31-080250{code}
 How reproducible:
 {code:none}
 Sometimes, not sure what triggers the error{code}
 Steps to Reproduce:
 {code:none}
 1. Visit Home -> Overview page and keep the window open for a long time and do nothing
 2.
 3.
 {code}
 Actual results:
 {code:none}
 We can see an error page displayed after sometime
 
 TypeErrorDescription:Cannot read properties of undefined (reading 'description')
 
 Component trace:
 at https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:1352496
     at https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/dashboards~metal3-baremetalhost~namespace~node~node-detail-chunk-92927db820425366ad3d.min.js:1:22265
     at p.memo.t.children.e.children.t.reduxes.e.reduxes.t.reduxes.every.reduxID (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:634756)
     at S (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendors~main-chunk-24c590a6d12fc8fa6bbf.min.js:87:86676)
     at <anonymous> (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:636622)
     at S (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendors~main-chunk-24c590a6d12fc8fa6bbf.min.js:87:86676)
     at <anonymous> (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/dashboards~metal3-baremetalhost~namespace~node~node-detail-chunk-92927db820425366ad3d.min.js:1:588)
     at S (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendors~main-chunk-24c590a6d12fc8fa6bbf.min.js:87:86676)
     at div
     at u (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendor-patternfly-core-chunk-73fa94b476a411a945fb.min.js:1:120940)
     at div
     at p (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:1341337)
     at article
     at d (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendor-patternfly-core-chunk-73fa94b476a411a945fb.min.js:1:75090)
     at C (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/dashboards~metal3-baremetalhost~namespace~node~node-detail-chunk-92927db820425366ad3d.min.js:1:23649)
     at div
     at u (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendor-patternfly-core-chunk-73fa94b476a411a945fb.min.js:1:57168)
     at div
     at u (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendor-patternfly-core-chunk-73fa94b476a411a945fb.min.js:1:120940)
     at div
     at u (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendor-patternfly-core-chunk-73fa94b476a411a945fb.min.js:1:57168)
     at div
     at u (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendor-patternfly-core-chunk-73fa94b476a411a945fb.min.js:1:120940)
     at div
     at t.default (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:1343061)
     at div
     at t.default (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:222744)
     at h (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/dashboards-chunk-8ac1ba95f2636c5fc2db.min.js:1:353)
     at i (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:367513)
     at t (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendors~main-chunk-24c590a6d12fc8fa6bbf.min.js:21:69403)
     at t (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendors~main-chunk-24c590a6d12fc8fa6bbf.min.js:21:71448)
     at Suspense
     at div
     at https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:772795
     at https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/dashboards-chunk-8ac1ba95f2636c5fc2db.min.js:1:15657
     at S (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendors~main-chunk-24c590a6d12fc8fa6bbf.min.js:87:86676)
     at s (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:162531)
     at t (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendors~main-chunk-24c590a6d12fc8fa6bbf.min.js:21:69403)
     at T
     at t (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendors~main-chunk-24c590a6d12fc8fa6bbf.min.js:21:71448)
     at Suspense
     at i (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:367513)
     at section
     at m (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendor-patternfly-core-chunk-73fa94b476a411a945fb.min.js:1:170067)
     at div
     at div
     at t.a (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:1261973)
     at div
     at div
     at c (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendor-patternfly-core-chunk-73fa94b476a411a945fb.min.js:1:699285)
     at d (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendor-patternfly-core-chunk-73fa94b476a411a945fb.min.js:1:219102)
     at div
     at d (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendor-patternfly-core-chunk-73fa94b476a411a945fb.min.js:1:89497)
     at c (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:1026224)
     at https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:374347
     at S (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendors~main-chunk-24c590a6d12fc8fa6bbf.min.js:87:86676)
     at main
     at div
     at v (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendor-patternfly-core-chunk-73fa94b476a411a945fb.min.js:1:466913)
     at div
     at div
     at c (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendor-patternfly-core-chunk-73fa94b476a411a945fb.min.js:1:311318)
     at div
     at div
     at c (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendor-patternfly-core-chunk-73fa94b476a411a945fb.min.js:1:699285)
     at d (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendor-patternfly-core-chunk-73fa94b476a411a945fb.min.js:1:219102)
     at div
     at d (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendor-patternfly-core-chunk-73fa94b476a411a945fb.min.js:1:89497)
     at Jn (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendors~main-chunk-24c590a6d12fc8fa6bbf.min.js:36:185603)
     at t.default (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:781345)
     at t.default (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/quick-start-chunk-65e076b79c5e124b90d0.min.js:1:1261)
     at s (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:162531)
     at t.a (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:1340377)
     at ee (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:1358378)
     at St (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendors~main-chunk-24c590a6d12fc8fa6bbf.min.js:36:142711)
     at ee (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:1358378)
     at ee (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:1358378)
     at i (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:737017)
     at t.a (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:1334610)
     at t.a (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:1334799)
     at t.a (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:1332284)
     at te (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:1358509)
     at https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:1361613
     at r (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendors~main-chunk-24c590a6d12fc8fa6bbf.min.js:36:122340)
     at t (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendors~main-chunk-24c590a6d12fc8fa6bbf.min.js:21:69403)
     at t (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendors~main-chunk-24c590a6d12fc8fa6bbf.min.js:21:71448)
     at t (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendors~main-chunk-24c590a6d12fc8fa6bbf.min.js:21:66008)
     at ae (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:1361877)
     at t.a (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:711369)
     at t.a (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:959460)
     at s (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:567305)
     at t.a (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendors~main-chunk-24c590a6d12fc8fa6bbf.min.js:135:257438)
     at Suspense
 
 
 Stack trace:
 TypeError: Cannot read properties of undefined (reading 'description')
     at https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:1353099
     at https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/main-chunk-d5c8c19ae7b84e8e6960.min.js:1:1353338
     at na (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendors~main-chunk-24c590a6d12fc8fa6bbf.min.js:248:58879)
     at za (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendors~main-chunk-24c590a6d12fc8fa6bbf.min.js:248:68397)
     at La (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendors~main-chunk-24c590a6d12fc8fa6bbf.min.js:248:67731)
     at Hs (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendors~main-chunk-24c590a6d12fc8fa6bbf.min.js:248:114591)
     at xl (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendors~main-chunk-24c590a6d12fc8fa6bbf.min.js:248:98327)
     at Cl (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendors~main-chunk-24c590a6d12fc8fa6bbf.min.js:248:98255)
     at _l (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendors~main-chunk-24c590a6d12fc8fa6bbf.min.js:248:98118)
     at pl (https://console-openshift-console.apps.qe-daily1-414-0602.qe.devcluster.openshift.com/static/vendors~main-chunk-24c590a6d12fc8fa6bbf.min.js:248:95105){code}
 Expected results:
 {code:none}
 no errors happen{code}
 Additional info:
 {code:none}
  {code}
Status: ASSIGNED
#OCPBUGS-75912issue5 days ago[4.20] openshift-cluster-csi-drivers vmware-vsphere-csi-driver-operator-metrics endpoint 8445 port should return 401 instead 500 Verified
Issue 17839406: [4.20] openshift-cluster-csi-drivers vmware-vsphere-csi-driver-operator-metrics endpoint 8445 port should return 401 instead 500
Description: This is a clone of issue OCPBUGS-74569. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-60159. The following is the description of the original issue:  --- Description of problem:
 
 see similar bug OCPBUGS-57585
 
 Stole the testing result from [https://github.com/openshift/origin/pull/30014#issuecomment-3153130188]
 
 8445 port is for openshift-cluster-csi-drivers vmware-vsphere-csi-driver-operator-metrics endpoint, did not see this port on other IAAS(example AWS,OSP)checked on vsphere, curl 8445 without auth returns 500(actually it's forbidden),
 {noformat}
 $ oc get infrastructure/cluster -o jsonpath={.spec.platformSpec.type}
 VSphere
 
 $ oc -n openshift-cluster-csi-drivers get ep
 Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice
 NAME                                           ENDPOINTS                                                        AGE
 vmware-vsphere-csi-driver-controller-metrics   10.129.0.17:9203,10.130.0.21:9203,10.129.0.17:9202 + 9 more...   28m
 vmware-vsphere-csi-driver-operator-metrics     10.130.0.17:8445                                                 28m
 vmware-vsphere-csi-driver-webhook-svc          10.128.0.45:8443,10.130.0.23:8443                                28m
 
 $ oc -n openshift-monitoring exec -c prometheus prometheus-k8s-0 -- curl -k -s -o /dev/null -w '%{http_code}' https://10.130.0.17:8445/metrics
 500
 {noformat}
 return code is 500, expecting 401 forbidden instead of 500.
 {code:java}
 $ oc -n openshift-monitoring exec -c prometheus prometheus-k8s-0 -- curl -k https://10.130.0.17:8445/metrics
 
 Internal Server Error: "/metrics": subjectaccessreviews.authorization.k8s.io is forbidden: User "system:serviceaccount:openshift-cluster-csi-drivers:vmware-vsphere-csi-driver-operator" cannot create resource "subjectaccessreviews" in API group "authorization.k8s.io" at the cluster scope  {code}
 curl with auth, no issue
 {noformat}
 $ token=`oc create token prometheus-k8s -n openshift-monitoring`
 $ oc -n openshift-monitoring exec -c prometheus prometheus-k8s-0 -- curl -k -H "Authorization: Bearer $token" 'https://10.130.0.17:8445/metrics' | head -n3
 # HELP aggregator_discovery_aggregation_count_total [ALPHA] Counter of number of times discovery was aggregated
 # TYPE aggregator_discovery_aggregation_count_total counter
 aggregator_discovery_aggregation_count_total 0 {noformat}
 [https://github.com/openshift/origin/pull/30014#issuecomment-3154900379]
 
  
 
 When the bug is fixed, please remove it from this code snippet to avoid regression.
 
 [https://github.com/openshift/origin/blob/4f183dd3427cffd8d97b44557caa782d65726416/test/extended/prometheus/prometheus.go#L64-L71]
Status: Verified
#OCPBUGS-76596issue2 weeks agoFailed to configure pod interface: timed out waiting for OVS port binding New
Issue 17866577: Failed to configure pod interface: timed out waiting for OVS port binding
Description: Description of problem:
 
 After the OCP cluster upgrade from version 4.17.38 to 4.18.30. We are observing several pods stuck is container creating phase. Following is the sample log message for one of those pods.
 {code:java}
 Jan 28 10:51:44 helixtest-6w28l-worker-0-q6mwl kubenswrapper[14632]: E0128 10:51:44.775553   14632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"redis-redis-ha-server-1_helix-itom-test(a82254b6-5c69-4c3b-bb3c-bd6e8aa43c44)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"redis-redis-ha-server-1_helix-itom-test(a82254b6-5c69-4c3b-bb3c-bd6e8aa43c44)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_redis-redis-ha-server-1_helix-itom-test_a82254b6-5c69-4c3b-bb3c-bd6e8aa43c44_0(7b9dfc5ae370376d3e9dff1608bc40a130848aad4c7474c1bad8dfd06aec8b20): error adding pod helix-itom-test_redis-redis-ha-server-1 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"7b9dfc5ae370376d3e9dff1608bc40a130848aad4c7474c1bad8dfd06aec8b20\\\" Netns:\\\"/var/run/netns/453c2a54-81b6-4c59-93ec-f6f613a2bee4\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=helix-itom-test;K8S_POD_NAME=redis-redis-ha-server-1;K8S_POD_INFRA_CONTAINER_ID=7b9dfc5ae370376d3e9dff1608bc40a130848aad4c7474c1bad8dfd06aec8b20;K8S_POD_UID=a82254b6-5c69-4c3b-bb3c-bd6e8aa43c44\\\" Path:\\\"\\\" ERRORED: error configuring pod [helix-itom-test/redis-redis-ha-server-1] networking: [helix-itom-test/redis-redis-ha-server-1/a82254b6-5c69-4c3b-bb3c-bd6e8aa43c44:ovn-kubernetes]: error adding container to network \\\"ovn-kubernetes\\\": CNI request failed with status 400: '[helix-itom-test/redis-redis-ha-server-1 7b9dfc5ae370376d3e9dff1608bc40a130848aad4c7474c1bad8dfd06aec8b20 network default NAD default] [helix-itom-test/redis-redis-ha-server-1 7b9dfc5ae370376d3e9dff1608bc40a130848aad4c7474c1bad8dfd06aec8b20 network default NAD default] failed to configure pod interface: timed out waiting for OVS port binding (ovn-installed) for 0a:58:0a:54:02:3d [10.84.2.61/23]\\n'\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="helix-itom-test/redis-redis-ha-server-1" podUID="a82254b6-5c69-4c3b-bb3c-bd6e8aa43c44"{code}
 After the cluster upgrade, the issue was observed on a few nodes, but not across all nodes. The problem was resolved after restarting the {{ovnkube-node}} pods.
 
 We have observed the same issue on clusters upgraded from version 4.17.38 to 4.18.30.
 
  
 
  
 
 Version-Release number of selected component (if applicable):
 
 How reproducible:
 
 Steps to Reproduce:
 
 1. Upgrade the cluster from 4.17.38 to 4.18.30
 
 2.
 
 3.
 
 Actual results: The pods remain stuck in the creation phase until the {{ovnkube-node}} pods are restarted.
 
 Expected results: The pods should not stuck in creation phase.
 
 Additional info: I'll share the additional details in comments section.
 
 Please fill in the following template while reporting a bug and provide as much relevant information as possible. Doing so will give us the best chance to find a prompt resolution.
 
 Affected Platforms:
 
 Is it an
  # internal CI failure
  # customer issue / SD
  # internal RedHat testing failure
 
 If it is an internal RedHat testing failure:
  * Please share a kubeconfig or creds to a live cluster for the assignee to debug/troubleshoot along with reproducer steps (specially if it's a telco use case like ICNI, secondary bridges or BM+kubevirt).
 
 If it is a CI failure:
  * Did it happen in different CI lanes? If so please provide links to multiple failures with the same error instance
  * Did it happen in both sdn and ovn jobs? If so please provide links to multiple failures with the same error instance
  * Did it happen in other platforms (e.g. aws, azure, gcp, baremetal etc) ? If so please provide links to multiple failures with the same error instance
  * When did the failure start happening? Please provide the UTC timestamp of the networking outage window from a sample failure run
  * If it's a connectivity issue,
  * What is the srcNode, srcIP and srcNamespace and srcPodName?
  * What is the dstNode, dstIP and dstNamespace and dstPodName?
  * What is the traffic path? (examples: pod2pod? pod2external?, pod2svc? pod2Node? etc)
 
 If it is a customer / SD issue:
  * Provide enough information in the bug description that Engineering doesn’t need to read the entire case history.
  * Don’t presume that Engineering has access to Salesforce.
  * Do presume that Engineering will access attachments through supportshell.
  * Describe what each relevant attachment is intended to demonstrate (failed pods, log errors, OVS issues, etc).
  * Referring to the attached must-gather, sosreport or other attachment, please provide the following details:
  ** If the issue is in a customer namespace then provide a namespace inspect.
  ** If it is a connectivity issue:
  *** What is the srcNode, srcNamespace, srcPodName and srcPodIP?
  *** What is the dstNode, dstNamespace, dstPodName and dstPodIP?
  *** What is the traffic path? (examples: pod2pod? pod2external?, pod2svc? pod2Node? etc)
  *** Please provide the UTC timestamp networking outage window from must-gather
  *** Please provide tcpdump pcaps taken during the outage filtered based on the above provided src/dst IPs
  ** If it is not a connectivity issue:
  *** Describe the steps taken so far to analyze the logs from networking components (cluster-network-operator, OVNK, SDN, openvswitch, ovs-configure etc) and the actual component where the issue was seen based on the attached must-gather. Please attach snippets of relevant logs around the window when problem has happened if any.
 
  * When showing the results from commands, include the entire command in the output.  
  * For OCPBUGS in which the issue has been identified, label with “sbr-triaged”
  * For OCPBUGS in which the issue has not been identified and needs Engineering help for root cause, label with “sbr-untriaged”
  * Do not set the priority, that is owned by Engineering and will be set when the bug is evaluated
  * Note: bugs that do not meet these minimum standards will be closed with label “SDN-Jira-template”
  * For guidance on using this template please see
 [OCPBUGS Template Training for Networking  components|https://source.redhat.com/groups/public/gss-openshift/cee_openshift_wiki/sbr_shift_networking#jive_content_id_Training_Materials]
Status: New
#OCPBUGS-78106issue17 hours ago[openshift-4.13] cluster-network-operator can't be built hermetically New
Issue 17972074: [openshift-4.13] cluster-network-operator can't be built hermetically
Description: This is a clone of issue OCPBUGS-78105. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78104. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build cluster-network-operator hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
  
 
 Version-Release number of selected component (if applicable):
 
 How reproducible:
 
 Steps to Reproduce:
 
 1.
 
 2.
 
 3.
 
 Actual results:
 
 Expected results:
 
 Additional info:
 
 Please fill in the following template while reporting a bug and provide as much relevant information as possible. Doing so will give us the best chance to find a prompt resolution.
 
 Affected Platforms:
 
 Is it an
  # internal CI failure
  # customer issue / SD
  # internal RedHat testing failure
 
 If it is an internal RedHat testing failure:
  * Please share a kubeconfig or creds to a live cluster for the assignee to debug/troubleshoot along with reproducer steps (specially if it's a telco use case like ICNI, secondary bridges or BM+kubevirt).
 
 If it is a CI failure:
  * Did it happen in different CI lanes? If so please provide links to multiple failures with the same error instance
  * Did it happen in both sdn and ovn jobs? If so please provide links to multiple failures with the same error instance
  * Did it happen in other platforms (e.g. aws, azure, gcp, baremetal etc) ? If so please provide links to multiple failures with the same error instance
  * When did the failure start happening? Please provide the UTC timestamp of the networking outage window from a sample failure run
  * If it's a connectivity issue,
  * What is the srcNode, srcIP and srcNamespace and srcPodName?
  * What is the dstNode, dstIP and dstNamespace and dstPodName?
  * What is the traffic path? (examples: pod2pod? pod2external?, pod2svc? pod2Node? etc)
 
 If it is a customer / SD issue:
  * Provide enough information in the bug description that Engineering doesn’t need to read the entire case history.
  * Don’t presume that Engineering has access to Salesforce.
  * Do presume that Engineering will access attachments through supportshell.
  * Describe what each relevant attachment is intended to demonstrate (failed pods, log errors, OVS issues, etc).
  * Referring to the attached must-gather, sosreport or other attachment, please provide the following details:
  ** If the issue is in a customer namespace then provide a namespace inspect.
  ** If it is a connectivity issue:
  *** What is the srcNode, srcNamespace, srcPodName and srcPodIP?
  *** What is the dstNode, dstNamespace, dstPodName and dstPodIP?
  *** What is the traffic path? (examples: pod2pod? pod2external?, pod2svc? pod2Node? etc)
  *** Please provide the UTC timestamp networking outage window from must-gather
  *** Please provide tcpdump pcaps taken during the outage filtered based on the above provided src/dst IPs
  ** If it is not a connectivity issue:
  *** Describe the steps taken so far to analyze the logs from networking components (cluster-network-operator, OVNK, SDN, openvswitch, ovs-configure etc) and the actual component where the issue was seen based on the attached must-gather. Please attach snippets of relevant logs around the window when problem has happened if any.
 
  * When showing the results from commands, include the entire command in the output.  
  * For OCPBUGS in which the issue has been identified, label with “sbr-triaged”
  * For OCPBUGS in which the issue has not been identified and needs Engineering help for root cause, label with “sbr-untriaged”
  * Do not set the priority, that is owned by Engineering and will be set when the bug is evaluated
  * Note: bugs that do not meet these minimum standards will be closed with label “SDN-Jira-template”
  * For guidance on using this template please see
 [OCPBUGS Template Training for Networking  components|https://source.redhat.com/groups/public/gss-openshift/cee_openshift_wiki/sbr_shift_networking#jive_content_id_Training_Materials]
Status: New
#OCPBUGS-73849issue7 weeks agooauth-proxy e2e tests fail due to ambiguous container image short names on RHEL 9 New
Issue 17769514: oauth-proxy e2e tests fail due to ambiguous container image short names on RHEL 9
Description: The oauth-proxy e2e-component tests fail on OCP 4.21 because RHEL 9 enforces fully qualified container image names ("short name mode"). The backend test image {{nginxdemos/nginx-hello:plain-text}} is rejected by Podman since it cannot determine which registry to pull from.
 h2. Version-Release number
 
 4.21
 h2. How reproducible
 
 Always
 h2. Steps to Reproduce
 
 Run the oauth-proxy e2e-component tests on OCP 4.21
 
 The test creates a pod with a backend container using image {{nginxdemos/nginx-hello:plain-text}}
 
 Observe the pod fails to start
 h2. Actual results
 
 The backend container fails with:
 {code:java}
 ImageInspectError: short name mode is enforcing, but image name 
 nginxdemos/nginx-hello:plain-text returns ambiguous list
 {code}
 The pod never reaches Running state, and all e2e tests fail.
 h2. Expected results
 
 The e2e tests should pass. The backend container should start successfully using a fully qualified image reference.
 h2. Additional info
 
 _Root cause:_ RHEL 9 enforces "short name mode" for container images, requiring fully qualified registry paths.
 
 _Fix:_ Replace {{nginxdemos/nginx-hello:plain-text}} with {{{}registry.k8s.io/e2e-test-images/agnhost:2.53{}}}.
 
 _Backport of:_ [OCPBUGS-73848|https://issues.redhat.com/browse/OCPBUGS-73848]
Status: New
#OCPBUGS-65559issue6 days agoobservability tests are failing as loki is no longer a valid exporter for OTEL config file CLOSED
Issue 17570261: observability tests are failing as loki is no longer a valid exporter for OTEL config file
Description: Description of problem:{code:none}
 observability tests are failing as loki is no longer a valid exporter for OTEL config file
     {code}
     Version-Release number of selected component (if applicable):{code:none}
      4.21
     {code}
     How reproducible:{code:none}
     
     {code}
     Steps to Reproduce:{code:none}
     1. 
     2.
     3.
     {code}
     Actual results:{code:none}
     Nov 11 16:35:38 el96-src-optional-bootc-host1 opentelemetry-collector[18443]: Error: failed to get config: cannot unmarshal the configuration: decoding failed due to the following error(s):
 Nov 11 16:35:38 el96-src-optional-bootc-host1 opentelemetry-collector[18443]: 'exporters' unknown type: "loki" for id: "loki" (valid values: [otlp prometheus prometheusremotewrite kafka awscloudwatchlogs awsemf awsxray loadbalancing debug otlphttp file])
 Nov 11 16:35:38 el96-src-optional-bootc-host1 opentelemetry-collector[18443]: 2025/11/11 16:35:38 collector server run finished with error: failed to get config: cannot unmarshal the configuration: decoding failed due to the following error(s):
 Nov 11 16:35:38 el96-src-optional-bootc-host1 opentelemetry-collector[18443]: 'exporters' unknown type: "loki" for id: "loki" (valid values: [otlp prometheus prometheusremotewrite kafka awscloudwatchlogs awsemf awsxray loadbalancing debug otlphttp file])
 Nov 11 16:35:38 el96-src-optional-bootc-host1 systemd[1]: microshift-observability.service: Main process exited, code=exited, status=1/FAILURE
     {code}
     Expected results:{code:none}
      No failures.
     {code}
     Additional info:{code:none}
 
     {code}
Status: CLOSED
#OCPBUGS-77193issue2 weeks ago[openshift-4.19] ose-libvirt-machine-controllers can't be build hermetically New
Issue 17925857: [openshift-4.19] ose-libvirt-machine-controllers can't be build hermetically
Description: Description of problem:
 
 ART can't build ose-libvirt-machine-controllers hermetically.
 The error is the following in the prefetch-dependencies konflux task:
 {code:none}
     Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.  Please try running `go mod vendor` and committing the changes.  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
  
 
  
 
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-77858issue2 days ago[openshift-4.15] oc-mirror can't be built hermetically Verified
Issue 17959717: [openshift-4.15] oc-mirror can't be built hermetically
Description: This is a clone of issue OCPBUGS-77857. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build oc-mirror hermetically
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: Verified
#OCPBUGS-75954issue13 days ago[4.18] Improper namespace could affect PPC POST
Issue 17841296: [4.18] Improper namespace could affect PPC
Description: This is a clone of issue OCPBUGS-60218. The following is the description of the original issue:  --- Description of problem:
 {code:none}
 $ podman run --entrypoint performance-profile-creator -v "/root/ppc/must-gather.local.3879427019288622939/":/must-gather:z registry.redhat.io/openshift4/ose-cluster-node-tuning-rhel9-operator:v4.18 info --must-gather-dir-path /must-gather level=info msg="Nodes names targeted by master pool are: " level=info msg="Multiple matches for the specified must gather directory path: /must-gather and suffix: nodes" Error: failed to parse the cluster data: failed to load node's GHW snapshot: can't obtain the node path worker2.ocp4.terryhu.cc: Multiple matches for the specified must gather directory path: /must-gather and suffix: nodes.  Expected only one performance-addon-operator-must-gather* directory, please check the must-gather tarball{code}
 Version-Release number of selected component (if applicable):
 {code:none}
 4.18{code}
 How reproducible:
 {code:none}
 100%{code}
 Steps to Reproduce:
 {code:none}
     1. Collect must-gather
     2. Create a namespace that has "nodes" as its suffix, such as
 $ mkdir /root/ppc/must-gather.local.3879427019288622939/quay-io-openshift-release-dev-ocp-v4-0-art-dev-sha256-a81f2df09a1b9a7ed63a20ced2182b9ba00a4b0ce49a8598d167ba900c87b12a/namespaces/three-nodes     
     3. Run the PPC info
     {code}
 Actual results:
 {code:none}
 Multiple matches for the specified must gather directory path: /must-gather and suffix: nodes.  Expected only one performance-addon-operator-must-gather* directory, please check the must-gather tarball{code}
 Expected results:
 {code:none}
 The PPC should only filter the "nodes" under cluster-scoped-resources/core, not cluster-scoped-resources.{code}
 Additional info:
 {code:none}
     {code}
Status: POST
#OCPBUGS-62309issue2 weeks ago[4.22] Unable to provision NVIDIA DGX B200 nodes via ACM, Redfish VirtualMedia InsertMedia request fails with missing UserName/Password parameters Verified
Issue 17431143: [4.22] Unable to provision NVIDIA DGX B200 nodes via ACM, Redfish VirtualMedia InsertMedia request fails with missing UserName/Password parameters
Description: *Description of problem:*
 
 
 When attempting to provision NVIDIA DGX B200 nodes through Advanced Cluster Management (ACM), the BareMetalHost (BMH) status hangs at {_}Provisioning{_}.
 
 The Bare Metal Operator (BMO) and Ironic logs show Redfish VirtualMedia mounting failures. The BMC rejects the {{VirtualMedia.InsertMedia}} request with an HTTP 400 error, requiring {{UserName}} and {{Password}} parameters in the request body.
 
 This behavior prevents automated provisioning of DGX B200 nodes.
 ----
 *Version-Release number of selected component (if applicable):*
  * OpenShift: 4.18.3 (hub cluster)
 
  * ACM: 2.13
 
  * Ironic/BMO components from OCP 4.18.3 release
 
 ----
 *How reproducible:*
 Always – occurs consistently when attempting to provision NVIDIA DGX B200 nodes.
 ----
 *Steps to Reproduce:*
  # Register NVIDIA DGX B200 node in ACM using BareMetalHost (via BMC/Redfish).
 
  # Observe provisioning workflow.
 
  # Check Ironic/BMO logs when provisioning fails.
 
 ----
 *Actual results:*
  * BMH stuck in _Provisioning_ state.
 
  * ISO not mounted via Redfish VirtualMedia.
 
  * BMO/Ironic logs show:
 
  
 {code:java}
 HTTP request: POST https://<bmc>/redfish/v1/Managers/BMC/VirtualMedia/CD1/Actions/VirtualMedia.InsertMedia body: {'Image': 'https://<iso-server>/boot-<uuid>.iso', 'TransferProtocolType': 'HTTPS'}{code}
 {{ }}
 Response:
  
 {code:java}
 status code: 400 error: Base.1.12.GeneralError: The action ... requires the parameter UserName to be present in the request body. MessageId: Base.1.12.ActionParameterMissing{code}
 {{ }}
 ----
 *Expected results:*
  * Redfish VirtualMedia ISO should mount successfully.
 
  * BMH should proceed past _Provisioning_ and complete host registration.
 
 ----
 *Additional info:*
  * The fix for handling {{UserName}} in Redfish InsertMedia was merged upstream (sushy) and included in OCP 4.18.1 ({{{}Provide vmedia username and password if required{}}}).
 
  * However, in this scenario, the firmware’s response differs: the error lacks a structured {{code}} field, causing credential detection logic to fail.
 
  * Engineering suspects a firmware bug, but BMO/Ironic may require additional handling (e.g., checking {{MessageId=ActionParameterMissing}} or the {{error}} field directly) as a workaround.
Status: Verified
#OCPBUGS-77167issue7 days ago[openshift-4.22] hermetic build failure - ose-ibm-vpc-block-csi-driver Verified
Issue 17919616: [openshift-4.22] hermetic build failure - ose-ibm-vpc-block-csi-driver
Description: Description of problem:
 {code:none}
 IBM VPC Block CSI Driver build fails in hermetic mode due to non-deterministic GOPATH handling in Makefile. The build system rewrites $GOPATH causing the binary output path -o ${GOPATH}/bin/${EXE_DRIVER_NAME} to become invalid, resulting in build failures.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
 OpenShift 4.19+ (affects hermetic build conversion) {code}
 How reproducible:
 {code:none}
 Always - when building ibm-vpc-block-csi-driver in hermetic mode{code}
 Steps to Reproduce:
 {code:none}
   1. Build ibm-vpc-block-csi-driver container in hermetic mode
   2. Makefile uses absolute GOPATH-dependent output path: -o ${GOPATH}/bin/${EXE_DRIVER_NAME}
   3. Build system rewrites $GOPATH during hermetic build process     {code}
 Actual results:
 {code:none}
 Build fails because binary cannot be found at expected GOPATH location after build system modifies $GOPATH
 
 Error: building at STEP "COPY --from=builder /cachi2/output/deps/gomod/bin/ibm-vpc-block-csi-driver /bin/ibm-vpc-block-csi-driver": checking on sources under [...]: copier: stat: "/cachi2/output/deps/gomod/bin/ibm-vpc-block-csi-driver": no such file or directory{code}
 Expected results:
 {code:none}
 Build succeeds using relative path bin/${EXE_DRIVER_NAME} that doesn't depend on GOPATH manipulation, consistent with other CSI drivers like aws-ebs-csi-driver  {code}
 Additional info:
 {code:none}
 Fix implemented: Change Makefile to use relative path -o bin/${EXE_DRIVER_NAME} instead of absolute GOPATH-dependent path. This aligns with other CSI drivers and makes builds hermetic-compatible. Comment added to indicate this is a permanent carry patch for OpenShift builds.{code}
Status: Verified
#OCPBUGS-77766issue6 days ago[openshift-4.19] hermetic build failure - node-feature-discovery Verified
Issue 17953518: [openshift-4.19] hermetic build failure - node-feature-discovery
Description: This is a clone of issue OCPBUGS-77764. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77763. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77651. The following is the description of the original issue:  --- Description of problem:
 {code:none}
   Node Feature Discovery build fails in hermetic mode due to non-deterministic binary path handling in Dockerfile. The build uses "go install" command which places binaries in unpredictable GOPATH locations, and the Dockerfile copies from multiple inconsistent paths causing binary path conflicts.
   {code}
 Version-Release number of selected component (if applicable):
 {code:none}
   OpenShift Container Platform 4.19+
   node-feature-discovery component across all affected versions
   {code}
 How reproducible:
 {code:none}
   Always - when building node-feature-discovery in hermetic mode
   {code}
 Steps to Reproduce:
 {code:none}
   1. Build node-feature-discovery container in hermetic mode
   2. Dockerfile copies binaries from multiple paths: /go/bin/* and /go/node-feature-discovery/bin/*
   3. "go install" command places binaries in non-deterministic GOPATH locations during hermetic build process
   {code}
 Actual results:
 {code:none}
   Build fails because binaries are placed in unpredictable locations by "go install" and Dockerfile attempts to copy from multiple conflicting paths
 
   Error: Inconsistent binary placement causing copy failures during Docker image build process when switching from "go install" to "go build" in
   hermetic mode
   {code}
 Expected results:
 {code:none}
   Build succeeds using deterministic binary paths. Should use "go build" with explicit output paths instead of "go install" and copy binaries from single consistent location in Dockerfile.
   {code}
 Additional info:
 {code:none}
   Root cause: "go install" command places binaries in GOPATH-dependent locations that are not deterministic in hermetic builds, unlike "go build"
    which allows explicit output path control.
 
   Fix implemented: Remove redundant /go/bin/* copy command from Dockerfile and keep only the specific /go/node-feature-discovery/bin/* path. This
    ensures consistent binary placement and eliminates path conflicts during hermetic builds.
 
   Reference upstream work: https://github.com/openshift/node-feature-discovery/pull/163/changes
   Related hermetic conversion tracking: ART-14586
   {code}
Status: Verified
#OCPBUGS-77185issue7 days agoBMH stuck in deleting state while performing IBBF Verified
Issue 17925312: BMH stuck in deleting state while performing IBBF
Description: This is a clone of issue OCPBUGS-76988. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-59133. The following is the description of the original issue:  --- Description of problem:
 {code:none}
     While attempting to perform IBBF with the siteconfig operator, the BMH for the spoke is stuck in deleting state for too long due to the old BMH being unreachable. This results in the clusterinstance timing out while waiting for the BMH resource to be removed. We should find a way to mitigate this.
 {code}
     Version-Release number of selected component (if applicable):
 {code:java}
 4.18.0-0.nightly-2025-04-02-011053 {code}
 How reproducible:
 {code:none}
     100%{code}
 Steps to Reproduce:
 {code:none}
     1. Install an IBI spoke cluster using siteconfig operator
     2. Remove old node and attempt to replace with new node through IBBF
     3.
     {code}
 Actual results:
 {code:none}
     Reinstallation fails due to timeout{code}
 Expected results:
 {code:none}
     Reinstallation succeeds{code}
 Additional info:
 {code:none}
 $ oc get clusterinstance target-0 -o json | jq -r '.status.reinstall.conditions[0]'
 {
   "lastTransitionTime": "2025-04-04T17:24:45Z",
   "message": "Encountered error executing task: Deleting rendered manifests. Error: deletion timeout exceeded for object (BareMetalHost:target-0/target-0-0): Timed out waiting to delete object (BareMetalHost:target-0/target-0-0)",
   "reason": "Failed",
   "status": "False",
   "type": "ReinstallRequestProcessed"
 }
 $ oc get bmh target-0-0 -o json | jq -r .status.errorMessage
 Failed to get power state for node 7e063860-56c7-45ca-8bfa-d8f63997bd10. Error: Redfish exception occurred. Error: Resource https://[fd2e:6f44:5dd8:4::1]:8000/redfish/v1/Systems/ac8d214a-ae10-46f5-a179-208341bcb688 not found {code}
Status: Verified
#OCPBUGS-61601issue6 days agonaming and the actual interpretation appear to be misleading for images and image-tags in limitrange New
Issue 17385394: naming and the actual interpretation appear to be misleading for images and image-tags in limitrange
Description: Description of problem:
 {code:none}
 naming and the actual interpretation appear to be misleading for images and image-tags in limitrange{code}
 Version-Release number of selected component (if applicable):
 {code:none}
 4.x{code}
 How reproducible:
 {code:none}
 100 %{code}
 Steps to Reproduce:
 
 1- create a limitrange using below yaml
 
 
 {code:java}
 apiVersion: v1
 kind: LimitRange
 metadata:
   name: openshift-resource-limits
 spec:
   limits:
     - type: openshift.io/ImageStream
       max:
         openshift.io/image-tags: 1
         openshift.io/images: 2  
 
 oc describe limitrange
 Name:                     openshift-resource-limits
 Namespace:                amuhamme
 Type                      Resource                 Min  Max  Default Request  Default Limit  Max Limit/Request Ratio
 ----                      --------                 ---  ---  ---------------  -------------  -----------------------
 openshift.io/ImageStream  openshift.io/image-tags  -    1    -                -              -
 openshift.io/ImageStream  openshift.io/images      -    2    -                -              -{code}
  
 
 2- As per above limitrange conf, it should allow maximum one tag's and 2 images per tag
 
 3- Try to create more than one tag and able to create more than one tag
 {code:java}
  oc get is
 NAME    IMAGE REPOSITORY                                                  TAGS         UPDATED
 test    image-registry.openshift-image-registry.svc:5000/amuhamme/test    new,latest   6 seconds ago
 test1   image-registry.openshift-image-registry.svc:5000/amuhamme/test1  {code}
 3- if i create one more tag (3rd one) its failing with permission denied. 
 {code:java}
 [root@worker-0 ~]# podman push image-registry.openshift-image-registry.svc:5000/amuhamme/test:hi
 Getting image source signatures
 Copying blob b64aaade7c59 skipped: already exists 
 Copying blob dc9835b180d8 skipped: already exists 
 Copying blob ae0badd53767 skipped: already exists 
 Copying blob 210e6a098c9f skipped: already exists 
 Copying config a250ef31d2 done   | 
 Writing manifest to image destination
 Error: writing manifest: uploading manifest hi to image-registry.openshift-image-registry.svc:5000/amuhamme/test: denied {code}
 4- try to push another image (second one) to the tag and see its failing
 {code:java}
 [root@worker-0 ~]# podman push image-registry.openshift-image-registry.svc:5000/amuhamme/test:latest
 Getting image source signatures
 Copying blob 210e6a098c9f skipped: already exists 
 Copying blob ae0badd53767 skipped: already exists 
 Copying blob c91bbdb1dcb2 done   | 
 Copying blob dc9835b180d8 skipped: already exists 
 Copying config 6cd179e2e2 done   | 
 Writing manifest to image destination
 Error: writing manifest: uploading manifest latest to image-registry.openshift-image-registry.svc:5000/amuhamme/test: denied {code}
 It seems that _"image-tags"_ are being treated as the total number of images within a tag, while _"images"_ are being counted as the number of tags per imagestream. This naming and the actual interpretation appear to be misleading.
 
 Actual results:
 {code:none}
 name and actual considerations are misleading{code}
 Expected results:
 {code:none}
  name and actual considerations should not misleading   {code}
 Additional info:
 {code:none}
     {code}
Status: New
Checking if image destination supports signatures
Error: Copying this image would require changing layer representation, which we cannot do: "Would invalidate signatures"
[root@worker-0 ~]# podman push image-registry.openshift-image-registry.svc:5000/amuhamme/test:latest --remove-signatures
Writing manifest to image destination
Error: writing manifest: uploading manifest latest to image-registry.openshift-image-registry.svc:5000/amuhamme/test: denied {code}
Also, seems like this is not respecting by oc image-import command, even manual image push is denied more than one image, i am able to import using oc image-import. see below
#OCPBUGS-74407issue38 hours ago[4.18] Unable to provision NVIDIA DGX B200 nodes via ACM, Redfish VirtualMedia InsertMedia request fails with missing UserName/Password parameters ASSIGNED
Issue 17801336: [4.18] Unable to provision NVIDIA DGX B200 nodes via ACM, Redfish VirtualMedia InsertMedia request fails with missing UserName/Password parameters
Description: *Description of problem:*
 
 
 When attempting to provision NVIDIA DGX B200 nodes through Advanced Cluster Management (ACM), the BareMetalHost (BMH) status hangs at {_}Provisioning{_}.
 
 The Bare Metal Operator (BMO) and Ironic logs show Redfish VirtualMedia mounting failures. The BMC rejects the {{VirtualMedia.InsertMedia}} request with an HTTP 400 error, requiring {{UserName}} and {{Password}} parameters in the request body.
 
 This behavior prevents automated provisioning of DGX B200 nodes.
 ----
 *Version-Release number of selected component (if applicable):*
  * OpenShift: 4.18.3 (hub cluster)
 
  * ACM: 2.13
 
  * Ironic/BMO components from OCP 4.18.3 release
 
 ----
 *How reproducible:*
 Always – occurs consistently when attempting to provision NVIDIA DGX B200 nodes.
 ----
 *Steps to Reproduce:*
  # Register NVIDIA DGX B200 node in ACM using BareMetalHost (via BMC/Redfish).
 
  # Observe provisioning workflow.
 
  # Check Ironic/BMO logs when provisioning fails.
 
 ----
 *Actual results:*
  * BMH stuck in _Provisioning_ state.
 
  * ISO not mounted via Redfish VirtualMedia.
 
  * BMO/Ironic logs show:
 
  
 {code:java}
 HTTP request: POST https://<bmc>/redfish/v1/Managers/BMC/VirtualMedia/CD1/Actions/VirtualMedia.InsertMedia body: {'Image': 'https://<iso-server>/boot-<uuid>.iso', 'TransferProtocolType': 'HTTPS'}{code}
 {{ }}
 Response:
  
 {code:java}
 status code: 400 error: Base.1.12.GeneralError: The action ... requires the parameter UserName to be present in the request body. MessageId: Base.1.12.ActionParameterMissing{code}
 {{ }}
 ----
 *Expected results:*
  * Redfish VirtualMedia ISO should mount successfully.
 
  * BMH should proceed past _Provisioning_ and complete host registration.
 
 ----
 *Additional info:*
  * The fix for handling {{UserName}} in Redfish InsertMedia was merged upstream (sushy) and included in OCP 4.18.1 ({{{}Provide vmedia username and password if required{}}}).
 
  * However, in this scenario, the firmware’s response differs: the error lacks a structured {{code}} field, causing credential detection logic to fail.
 
  * Engineering suspects a firmware bug, but BMO/Ironic may require additional handling (e.g., checking {{MessageId=ActionParameterMissing}} or the {{error}} field directly) as a workaround.
Status: ASSIGNED
#OCPBUGS-77266issue13 days agodpu-cni, dpu-daemon and dpu-operator can't be built hermetically POST
Issue 17931079: dpu-cni, dpu-daemon and dpu-operator can't be built hermetically
Description: Description of problem:
 
 ART can't build dpu-operator, dpu-cni and dpu-daemon hermetically. Hermetic builds on Konflux require that go mod vendor does not produce a diff.
 {code:none}
   diff --git a/vendor/modules.txt b/vendor/modules.txt  
 index 2950f8c0..c651edc8 100644  
 --- a/vendor/modules.txt  
 +++ b/vendor/modules.txt  
 @@ -747,7 +747,6 @@ k8s.io/client-go/util/flowcontrol   
 k8s.io/client-go/util/homedir   
 k8s.io/client-go/util/keyutil   
 k8s.io/client-go/util/retry  
 -k8s.io/client-go/util/testing   
 k8s.io/client-go/util/workqueue   
 # k8s.io/component-base v0.29.0   
 ## explicit; go 1.21  2026-02-24 
 11:51:31,718 ERROR PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.  Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.    Please try running `go mod vendor` and committing the changes.    Note that you may need to `git add --force` ignored files in the vendor/ dir.  {code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: POST
#OCPBUGS-78176issue9 hours ago[openshift-4.16] ose-azure-cluster-api-controllers can't be built hermetically New
Issue 17974845: [openshift-4.16] ose-azure-cluster-api-controllers can't be built hermetically
Description: This is a clone of issue OCPBUGS-77983. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77905. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77904. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-azure-cluster-api-controllers
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-71189issue7 weeks agocluster install failed and worker machines stuck in Provisioned on proxy + techpreview cluster Verified
{code:java}
[~/must-gather.local.7303639993172620028/quay-io-openshift-release-dev-ocp-v4-0-art-dev-sha256-727500d1498ec6cb417e7f7fe80eaaaae551c877c5e456b58c776fb051150431] $ omc -n openshift-machine-config-operator logs machine-config-operator-6f6cb95d74-8blrn -c machine-config-operator 2026-01-07T03:30:39.182782649Z I0107 03:30:39.182742       1 start.go:52] Version: 4.22.0-0.nightly-2026-01-03-160855 (Raw: 6e3697a8, Hash: 9333c5a69d9e8159b98908561d7fb2269b7a9197) 2026-01-07T03:30:39.184776890Z I0107 03:30:39.184732       1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}. 2026-01-07T03:30:39.185543101Z I0107 03:30:39.185513       1 metrics.go:92] Registering Prometheus metrics 2026-01-07T03:30:39.185599617Z I0107 03:30:39.185581       1 metrics.go:99] Starting metrics listener on 127.0.0.1:8797 2026-01-07T03:30:39.248066417Z I0107 03:30:39.248028       1 leaderelection.go:257] attempting to acquire leader lease openshift-machine-config-operator/machine-config... 2026-01-07T03:30:39.259065477Z I0107 03:30:39.259034       1 leaderelection.go:271] successfully acquired lease openshift-machine-config-operator/machine-config 2026-01-07T03:30:39.290104920Z I0107 03:30:39.290068       1 simple_featuregate_reader.go:171] Starting feature-gate-detector 2026-01-07T03:30:39.298691190Z I0107 03:30:39.298644       1 event.go:377] Event(v1.ObjectReference{Kind:"Node", Namespace:"openshift-machine-config-operator", Name:"ip-10-0-53-145.us-east-2.compute.internal", UID:"4a35db35-5a95-46f2-8ef8-791c8e61f61d", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'FeatureGatesInitialized' FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AWSClusterHostedDNSInstall", "AWSDedicatedHosts", "AWSDualStackInstall", "AWSServiceLBNetworkSecurityGroup", "AdditionalRoutingCapabilities", "AdminNetworkPolicy", "AlibabaPlatform", "AutomatedEtcdBackup", "AzureClusterHostedDNSInstall", "AzureDedicatedHosts", "AzureDualStackInstall", "AzureMultiDisk", "AzureWorkloadIdentity", "BootImageSkewEnforcement", "BootcNodeManagement", "BuildCSIVolumes", "CBORServingAndStorage", "CPMSMachineNamePrefix", "CRDCompatibilityRequirementOperator", "ClientsPreferCBOR", "ClusterAPIInstallIBMCloud", "ClusterAPIMachineManagement", "ClusterMonitoringConfig", "ClusterVersionOperatorConfiguration", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DualReplica", "DyanmicServiceEndpointIBMCloud", "EtcdBackendQuota", "EventTTL", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GCPClusterHostedDNSInstall", "GCPCustomAPIEndpoints", "GCPCustomAPIEndpointsInstall", "GCPDualStackInstall", "GatewayAPI", "GatewayAPIController", "HighlyAvailableArbiter", "ImageModeStatusReporting", "ImageStreamImportMode", "ImageVolume", "InsightsConfig", "InsightsOnDemandDataGather", "IrreconcilableMachineConfig", "KMSEncryptionProvider", "KMSv1", "MachineAPIMigration", "MachineConfigNodes", "ManagedBootImages", "ManagedBootImagesAWS", "ManagedBootImagesAzure", "ManagedBootImagesCPMS", "ManagedBootImagesvSphere", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiDiskSetup", "MutableCSINodeAllocatableCount", "MutatingAdmissionPolicy", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NewOLMCatalogdAPIV1Metas", "NewOLMOwnSingleNamespace", "NewOLMPreflightPermissionChecks", "NewOLMWebhookProviderOpenshiftServiceCA", "NoRegistryClusterInstall", "NutanixMultiSubnets", "OSStreams", "OVNObservability", "OnPremDNSRecords", "OpenShiftPodSecurityAdmission", "PinnedImages", "PreconfiguredUDNAddresses", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "SELinuxMount", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "SigstoreImageVerificationPKI", "StoragePerformantSecurityPolicy", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereConfigurableMaxAllowedBlockVolumesPerNode", "VSphereHostVMGroupZonal", "VSphereMixedNodeEnv", "VSphereMultiDisk", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}, Disabled:[]v1.FeatureGateName{"ClientsAllowCBOR", "ClusterAPIInstall", "ClusterAPIMachineManagementVSphere", "EventedPLEG", "Example2", "ExternalSnapshotMetadata", "HyperShiftOnlyDynamicResourceAllocation", "IngressControllerDynamicConfigurationManager", "MachineAPIOperatorDisableMachineHealthCheckController", "MultiArchInstallAzure", "NetworkConnect", "NewOLMBoxCutterRuntime", "ProvisioningRequestAvailable", "ShortCertRotation"}} 2026-01-07T03:30:39.298746779Z I0107 03:30:39.298690       1 featuregates.go:112] FeatureGates initialized: enabled=[AWSClusterHostedDNS AWSClusterHostedDNSInstall AWSDedicatedHosts AWSDualStackInstall AWSServiceLBNetworkSecurityGroup AdditionalRoutingCapabilities AdminNetworkPolicy AlibabaPlatform AutomatedEtcdBackup AzureClusterHostedDNSInstall AzureDedicatedHosts AzureDualStackInstall AzureMultiDisk AzureWorkloadIdentity BootImageSkewEnforcement BootcNodeManagement BuildCSIVolumes CBORServingAndStorage CPMSMachineNamePrefix CRDCompatibilityRequirementOperator ClientsPreferCBOR ClusterAPIInstallIBMCloud ClusterAPIMachineManagement ClusterMonitoringConfig ClusterVersionOperatorConfiguration ConsolePluginContentSecurityPolicy DNSNameResolver DualReplica DyanmicServiceEndpointIBMCloud EtcdBackendQuota EventTTL Example ExternalOIDC ExternalOIDCWithUIDAndExtraClaimMappings GCPClusterHostedDNS GCPClusterHostedDNSInstall GCPCustomAPIEndpoints GCPCustomAPIEndpointsInstall GCPDualStackInstall GatewayAPI GatewayAPIController HighlyAvailableArbiter ImageModeStatusReporting ImageStreamImportMode ImageVolume InsightsConfig InsightsOnDemandDataGather IrreconcilableMachineConfig KMSEncryptionProvider KMSv1 MachineAPIMigration MachineConfigNodes ManagedBootImages ManagedBootImagesAWS ManagedBootImagesAzure ManagedBootImagesCPMS ManagedBootImagesvSphere MaxUnavailableStatefulSet MetricsCollectionProfiles MinimumKubeletVersion MixedCPUsAllocation MultiDiskSetup MutableCSINodeAllocatableCount MutatingAdmissionPolicy NetworkDiagnosticsConfig NetworkLiveMigration NetworkSegmentation NewOLM NewOLMCatalogdAPIV1Metas NewOLMOwnSingleNamespace NewOLMPreflightPermissionChecks NewOLMWebhookProviderOpenshiftServiceCA NoRegistryClusterInstall NutanixMultiSubnets OSStreams OVNObservability OnPremDNSRecords OpenShiftPodSecurityAdmission PinnedImages PreconfiguredUDNAddresses ProcMountType RouteAdvertisements RouteExternalCertificate SELinuxMount ServiceAccountTokenNodeBinding SignatureStores SigstoreImageVerification SigstoreImageVerificationPKI StoragePerformantSecurityPolicy UpgradeStatus UserNamespacesPodSecurityStandards UserNamespacesSupport VSphereConfigurableMaxAllowedBlockVolumesPerNode VSphereHostVMGroupZonal VSphereMixedNodeEnv VSphereMultiDisk VSphereMultiNetworks VolumeAttributesClass VolumeGroupSnapshot], disabled=[ClientsAllowCBOR ClusterAPIInstall ClusterAPIMachineManagementVSphere EventedPLEG Example2 ExternalSnapshotMetadata HyperShiftOnlyDynamicResourceAllocation IngressControllerDynamicConfigurationManager MachineAPIOperatorDisableMachineHealthCheckController MultiArchInstallAzure NetworkConnect NewOLMBoxCutterRuntime ProvisioningRequestAvailable ShortCertRotation] 2026-01-07T03:30:39.768680812Z I0107 03:30:39.768598       1 operator.go:421] Starting MachineConfigOperator 2026-01-07T03:30:39.769692277Z I0107 03:30:39.769670       1 osimagestream_ocp.go:43] Starting building of the OSImageStream instance 2026-01-07T03:30:39.769922691Z I0107 03:30:39.769903       1 event.go:377] Event(v1.ObjectReference{Kind:"ClusterOperator", Namespace:"openshift-machine-config-operator", Name:"machine-config", UID:"17c41ffe-dda1-4394-9647-e05eb8076050", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'OperatorVersionChanged' clusteroperator/machine-config started a version change from [] to [{operator 4.22.0-0.nightly-2026-01-03-160855} {operator-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:114347e569e8c93398dd56ec97f1a4a6be310296e889e2f8e9d4e2c8aa00bfea}] 2026-01-07T03:31:09.781088597Z time="2026-01-07T03:31:09Z" level=warning msg="Failed, retrying in 1s ... (1/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp: lookup quay.io: i/o timeout" 2026-01-07T03:31:09.781088597Z time="2026-01-07T03:31:09Z" level=warning msg="Failed, retrying in 1s ... (1/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp: lookup quay.io: i/o timeout" 2026-01-07T03:31:29.784262097Z time="2026-01-07T03:31:29Z" level=warning msg="Failed, retrying in 2s ... (2/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp: lookup quay.io on 172.30.0.10:53: read udp 10.130.0.25:38704->172.30.0.10:53: i/o timeout" 2026-01-07T03:31:29.784293967Z time="2026-01-07T03:31:29Z" level=warning msg="Failed, retrying in 2s ... (2/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp: lookup quay.io on 172.30.0.10:53: read udp 10.130.0.25:38704->172.30.0.10:53: i/o timeout" 2026-01-07T03:31:41.855214478Z time="2026-01-07T03:31:41Z" level=warning msg="Failed, retrying in 4s ... (3/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp: lookup quay.io on 172.30.0.10:53: read udp 10.130.0.25:58883->172.30.0.10:53: read: connection refused" 2026-01-07T03:31:41.855393194Z time="2026-01-07T03:31:41Z" level=warning msg="Failed, retrying in 4s ... (3/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp: lookup quay.io on 172.30.0.10:53: read udp 10.130.0.25:58883->172.30.0.10:53: read: connection refused" 2026-01-07T03:31:45.909631456Z time="2026-01-07T03:31:45Z" level=warning msg="Failed, retrying in 8s ... (4/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp: lookup quay.io on 172.30.0.10:53: read udp 10.130.0.25:42137->172.30.0.10:53: read: connection refused" 2026-01-07T03:31:45.909661238Z time="2026-01-07T03:31:45Z" level=warning msg="Failed, retrying in 8s ... (4/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp: lookup quay.io on 172.30.0.10:53: read udp 10.130.0.25:42137->172.30.0.10:53: read: connection refused" 2026-01-07T03:32:23.911790239Z time="2026-01-07T03:32:23Z" level=warning msg="Failed, retrying in 16s ... (5/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 34.239.22.218:443: i/o timeout" 2026-01-07T03:32:23.911893557Z time="2026-01-07T03:32:23Z" level=warning msg="Failed, retrying in 16s ... (5/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 34.239.22.218:443: i/o timeout" 2026-01-07T03:33:09.913513194Z time="2026-01-07T03:33:09Z" level=warning msg="Failed, retrying in 32s ... (6/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 54.225.186.3:443: i/o timeout" 2026-01-07T03:33:09.913575620Z time="2026-01-07T03:33:09Z" level=warning msg="Failed, retrying in 32s ... (6/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 54.225.186.3:443: i/o timeout" 2026-01-07T03:34:11.915140223Z time="2026-01-07T03:34:11Z" level=warning msg="Failed, retrying in 1m4s ... (7/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 98.88.205.176:443: i/o timeout" 2026-01-07T03:34:11.917330967Z time="2026-01-07T03:34:11Z" level=warning msg="Failed, retrying in 1m4s ... (7/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 98.88.205.176:443: i/o timeout" 2026-01-07T03:34:13.746831592Z I0107 03:34:13.746788       1 operator.go:441] Change observed to kube-apiserver-server-ca 2026-01-07T03:35:39.771942499Z time="2026-01-07T03:35:39Z" level=warning msg="Failed, retrying in 2m8s ... (8/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": context deadline exceeded" 2026-01-07T03:35:39.772002892Z time="2026-01-07T03:35:39Z" level=warning msg="Failed, retrying in 2m8s ... (8/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": context deadline exceeded" 2026-01-07T03:35:39.772002892Z W0107 03:35:39.771978       1 osimagestream.go:170] error fetching OSImageStreams from a source: error inspecting OS image error getting image source for quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4446453d8ae3c37706f8982e65b03c661cb6aa664fa2c93e280b6fefee05a00f: error getting image source for //quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4446453d8ae3c37706f8982e65b03c661cb6aa664fa2c93e280b6fefee05a00f: pinging container registry quay.io: Get "https://quay.io/v2/": context deadline exceeded 2026-01-07T03:35:39.772385463Z W0107 03:35:39.772370       1 osimagestream.go:170] error fetching OSImageStreams from a source: pinging container registry registry.ci.openshift.org: Get "https://registry.ci.openshift.org/v2/": context deadline exceeded 2026-01-07T03:35:39.772606165Z I0107 03:35:39.772581       1 event.go:377] Event(v1.ObjectReference{Kind:"ClusterOperator", Namespace:"openshift-machine-config-operator", Name:"machine-config", UID:"17c41ffe-dda1-4394-9647-e05eb8076050", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Warning' reason: 'OperatorDegraded: OSImageStreamFailed' Unable to apply 4.22.0-0.nightly-2026-01-03-160855: error building the OSImageStream: unable to retrieve any OSImageStream from the configured sources 2026-01-07T03:35:39.780735277Z I0107 03:35:39.780635       1 osimagestream_ocp.go:43] Starting building of the OSImageStream instance 2026-01-07T03:35:39.780926375Z I0107 03:35:39.780899       1 event.go:377] Event(v1.ObjectReference{Kind:"ClusterOperator", Namespace:"openshift-machine-config-operator", Name:"machine-config", UID:"17c41ffe-dda1-4394-9647-e05eb8076050", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'OperatorVersionChanged' clusteroperator/machine-config started a version change from [] to [{operator 4.22.0-0.nightly-2026-01-03-160855} {operator-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:114347e569e8c93398dd56ec97f1a4a6be310296e889e2f8e9d4e2c8aa00bfea}] 2026-01-07T03:36:09.788441114Z time="2026-01-07T03:36:09Z" level=warning msg="Failed, retrying in 1s ... (1/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 98.89.121.154:443: i/o timeout" 2026-01-07T03:36:09.788441114Z time="2026-01-07T03:36:09Z" level=warning msg="Failed, retrying in 1s ... (1/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 98.89.121.154:443: i/o timeout" 2026-01-07T03:36:40.790501478Z time="2026-01-07T03:36:40Z" level=warning msg="Failed, retrying in 2s ... (2/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 98.89.121.154:443: i/o timeout" 2026-01-07T03:36:40.790656389Z time="2026-01-07T03:36:40Z" level=warning msg="Failed, retrying in 2s ... (2/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 98.89.121.154:443: i/o timeout" 2026-01-07T03:37:12.792224170Z time="2026-01-07T03:37:12Z" level=warning msg="Failed, retrying in 4s ... (3/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 54.225.186.3:443: i/o timeout" 2026-01-07T03:37:12.792224170Z time="2026-01-07T03:37:12Z" level=warning msg="Failed, retrying in 4s ... (3/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 54.225.186.3:443: i/o timeout" 2026-01-07T03:37:46.794170911Z time="2026-01-07T03:37:46Z" level=warning msg="Failed, retrying in 8s ... (4/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 98.89.119.17:443: i/o timeout" 2026-01-07T03:37:46.794170911Z time="2026-01-07T03:37:46Z" level=warning msg="Failed, retrying in 8s ... (4/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 98.89.119.17:443: i/o timeout" 2026-01-07T03:38:24.795203614Z time="2026-01-07T03:38:24Z" level=warning msg="Failed, retrying in 16s ... (5/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 54.225.186.3:443: i/o timeout" 2026-01-07T03:38:24.795203614Z time="2026-01-07T03:38:24Z" level=warning msg="Failed, retrying in 16s ... (5/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 54.225.186.3:443: i/o timeout" 2026-01-07T03:39:01.469345834Z E0107 03:39:01.469301       1 leaderelection.go:441] Failed to update lock optimistically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path 2026-01-07T03:39:10.796420874Z time="2026-01-07T03:39:10Z" level=warning msg="Failed, retrying in 32s ... (6/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 54.225.186.3:443: i/o timeout" 2026-01-07T03:39:10.796420874Z time="2026-01-07T03:39:10Z" level=warning msg="Failed, retrying in 32s ... (6/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 54.225.186.3:443: i/o timeout" 2026-01-07T03:40:01.470755282Z E0107 03:40:01.470714       1 leaderelection.go:448] error retrieving resource lock openshift-machine-config-operator/machine-config: the server was unable to return a response in the time allotted, but may still be processing the request (get leases.coordination.k8s.io machine-config) 2026-01-07T03:40:12.798360642Z time="2026-01-07T03:40:12Z" level=warning msg="Failed, retrying in 1m4s ... (7/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 44.219.13.75:443: i/o timeout" 2026-01-07T03:40:12.798360642Z time="2026-01-07T03:40:12Z" level=warning msg="Failed, retrying in 1m4s ... (7/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 44.219.13.75:443: i/o timeout" 2026-01-07T03:40:39.787068848Z W0107 03:40:39.787016       1 osimagestream.go:170] error fetching OSImageStreams from a source: error inspecting OS image error getting image source for quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4446453d8ae3c37706f8982e65b03c661cb6aa664fa2c93e280b6fefee05a00f: error getting image source for //quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4446453d8ae3c37706f8982e65b03c661cb6aa664fa2c93e280b6fefee05a00f: pinging container registry quay.io: Get "https://quay.io/v2/": dial tcp 44.219.13.75:443: i/o timeout 2026-01-07T03:40:39.787522935Z W0107 03:40:39.787485       1 osimagestream.go:170] error fetching OSImageStreams from a source: pinging container registry registry.ci.openshift.org: Get "https://registry.ci.openshift.org/v2/": context deadline exceeded 2026-01-07T03:40:39.787734479Z I0107 03:40:39.787700       1 event.go:377] Event(v1.ObjectReference{Kind:"ClusterOperator", Namespace:"openshift-machine-config-operator", Name:"machine-config", UID:"17c41ffe-dda1-4394-9647-e05eb8076050", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Warning' reason: 'OperatorDegraded: OSImageStreamFailed' Unable to apply 4.22.0-0.nightly-2026-01-03-160855: error building the OSImageStream: unable to retrieve any OSImageStream from the configured sources 2026-01-07T03:40:39.846570394Z I0107 03:40:39.846530       1 osimagestream_ocp.go:43] Starting building of the OSImageStream instance 2026-01-07T03:41:09.853142075Z time="2026-01-07T03:41:09Z" level=warning msg="Failed, retrying in 1s ... (1/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 98.89.119.17:443: i/o timeout" 2026-01-07T03:41:09.853142075Z time="2026-01-07T03:41:09Z" level=warning msg="Failed, retrying in 1s ... (1/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 98.89.119.17:443: i/o timeout" 2026-01-07T03:41:40.855420954Z time="2026-01-07T03:41:40Z" level=warning msg="Failed, retrying in 2s ... (2/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 44.219.13.75:443: i/o timeout" 2026-01-07T03:41:40.855420954Z time="2026-01-07T03:41:40Z" level=warning msg="Failed, retrying in 2s ... (2/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 44.219.13.75:443: i/o timeout" 2026-01-07T03:42:12.857492517Z time="2026-01-07T03:42:12Z" level=warning msg="Failed, retrying in 4s ... (3/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 44.219.13.75:443: i/o timeout" 2026-01-07T03:42:12.857492517Z time="2026-01-07T03:42:12Z" level=warning msg="Failed, retrying in 4s ... (3/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 44.219.13.75:443: i/o timeout" 2026-01-07T03:42:46.860653030Z time="2026-01-07T03:42:46Z" level=warning msg="Failed, retrying in 8s ... (4/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 98.89.121.154:443: i/o timeout" 2026-01-07T03:42:46.860858926Z time="2026-01-07T03:42:46Z" level=warning msg="Failed, retrying in 8s ... (4/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 98.89.121.154:443: i/o timeout" 2026-01-07T03:43:24.862495147Z time="2026-01-07T03:43:24Z" level=warning msg="Failed, retrying in 16s ... (5/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 98.89.121.154:443: i/o timeout" 2026-01-07T03:43:24.862495147Z time="2026-01-07T03:43:24Z" level=warning msg="Failed, retrying in 16s ... (5/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 98.89.121.154:443: i/o timeout" 2026-01-07T03:44:10.864415197Z time="2026-01-07T03:44:10Z" level=warning msg="Failed, retrying in 32s ... (6/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 98.91.37.92:443: i/o timeout" 2026-01-07T03:44:10.864415197Z time="2026-01-07T03:44:10Z" level=warning msg="Failed, retrying in 32s ... (6/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 98.91.37.92:443: i/o timeout" 2026-01-07T03:45:12.867431892Z time="2026-01-07T03:45:12Z" level=warning msg="Failed, retrying in 1m4s ... (7/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 3.208.138.95:443: i/o timeout" 2026-01-07T03:45:12.867431892Z time="2026-01-07T03:45:12Z" level=warning msg="Failed, retrying in 1m4s ... (7/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 3.208.138.95:443: i/o timeout" 2026-01-07T03:45:39.851781849Z W0107 03:45:39.851735       1 osimagestream.go:170] error fetching OSImageStreams from a source: error inspecting OS image error getting image source for quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4446453d8ae3c37706f8982e65b03c661cb6aa664fa2c93e280b6fefee05a00f: error getting image source for //quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4446453d8ae3c37706f8982e65b03c661cb6aa664fa2c93e280b6fefee05a00f: pinging container registry quay.io: Get "https://quay.io/v2/": dial tcp 3.208.138.95:443: i/o timeout 2026-01-07T03:45:39.852161055Z W0107 03:45:39.852146       1 osimagestream.go:170] error fetching OSImageStreams from a source: pinging container registry registry.ci.openshift.org: Get "https://registry.ci.openshift.org/v2/": context deadline exceeded 2026-01-07T03:45:39.852454419Z I0107 03:45:39.852431       1 osimagestream_ocp.go:43] Starting building of the OSImageStream instance 2026-01-07T03:46:09.856170272Z time="2026-01-07T03:46:09Z" level=warning msg="Failed, retrying in 1s ... (1/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 44.219.13.75:443: i/o timeout" 2026-01-07T03:46:09.856266863Z time="2026-01-07T03:46:09Z" level=warning msg="Failed, retrying in 1s ... (1/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 44.219.13.75:443: i/o timeout" 2026-01-07T03:46:40.858003632Z time="2026-01-07T03:46:40Z" level=warning msg="Failed, retrying in 2s ... (2/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 44.219.13.75:443: i/o timeout" 2026-01-07T03:46:40.858038529Z time="2026-01-07T03:46:40Z" level=warning msg="Failed, retrying in 2s ... (2/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 44.219.13.75:443: i/o timeout" 2026-01-07T03:47:12.859537756Z time="2026-01-07T03:47:12Z" level=warning msg="Failed, retrying in 4s ... (3/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 3.208.138.95:443: i/o timeout" 2026-01-07T03:47:12.859537756Z time="2026-01-07T03:47:12Z" level=warning msg="Failed, retrying in 4s ... (3/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 3.208.138.95:443: i/o timeout" 2026-01-07T03:47:46.862650766Z time="2026-01-07T03:47:46Z" level=warning msg="Failed, retrying in 8s ... (4/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 3.208.138.95:443: i/o timeout" 2026-01-07T03:47:46.862650766Z time="2026-01-07T03:47:46Z" level=warning msg="Failed, retrying in 8s ... (4/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 3.208.138.95:443: i/o timeout" 2026-01-07T03:48:24.863428543Z time="2026-01-07T03:48:24Z" level=warning msg="Failed, retrying in 16s ... (5/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 98.91.37.92:443: i/o timeout" 2026-01-07T03:48:24.863428543Z time="2026-01-07T03:48:24Z" level=warning msg="Failed, retrying in 16s ... (5/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 98.91.37.92:443: i/o timeout" 2026-01-07T03:49:10.865865310Z time="2026-01-07T03:49:10Z" level=warning msg="Failed, retrying in 32s ... (6/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 98.91.37.92:443: i/o timeout" 2026-01-07T03:49:10.865865310Z time="2026-01-07T03:49:10Z" level=warning msg="Failed, retrying in 32s ... (6/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 98.91.37.92:443: i/o timeout" 2026-01-07T03:50:12.869451737Z time="2026-01-07T03:50:12Z" level=warning msg="Failed, retrying in 1m4s ... (7/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 3.208.138.95:443: i/o timeout" 2026-01-07T03:50:12.869451737Z time="2026-01-07T03:50:12Z" level=warning msg="Failed, retrying in 1m4s ... (7/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 3.208.138.95:443: i/o timeout" 2026-01-07T03:50:39.856418285Z W0107 03:50:39.856376       1 osimagestream.go:170] error fetching OSImageStreams from a source: error inspecting OS image error getting image source for quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4446453d8ae3c37706f8982e65b03c661cb6aa664fa2c93e280b6fefee05a00f: error getting image source for //quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4446453d8ae3c37706f8982e65b03c661cb6aa664fa2c93e280b6fefee05a00f: pinging container registry quay.io: Get "https://quay.io/v2/": dial tcp 3.208.138.95:443: i/o timeout 2026-01-07T03:50:39.856923760Z W0107 03:50:39.856890       1 osimagestream.go:170] error fetching OSImageStreams from a source: pinging container registry registry.ci.openshift.org: Get "https://registry.ci.openshift.org/v2/": context deadline exceeded 2026-01-07T03:50:39.859295024Z I0107 03:50:39.857256       1 osimagestream_ocp.go:43] Starting building of the OSImageStream instance 2026-01-07T03:51:09.864502095Z time="2026-01-07T03:51:09Z" level=warning msg="Failed, retrying in 1s ... (1/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 98.88.205.176:443: i/o timeout" 2026-01-07T03:51:09.864502095Z time="2026-01-07T03:51:09Z" level=warning msg="Failed, retrying in 1s ... (1/50). Error: pinging container registry quay.io: Get \"https://quay.io/v2/\": dial tcp 98.88.205.176:443: i/o timeout" ...
{code}
#OCPBUGS-64665issue3 months agokube-descheduler-operator e2e tests failure on s390x CLOSED
Issue 17541721: kube-descheduler-operator e2e tests failure on s390x
Description: Description of problem:
 {code:none}
     Observing TestRemoveDuplicates, TestLiveMigrationInBackground, TestFailedPods, TestLowNodeUtilizationKubernetesMetrics, TestTooManyRestarts, TestTopologySpreadConstraint tests failed in e2e tests.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
  {code}
 How reproducible:
 {code:none}
     Every time{code}
 Steps to Reproduce:
 {code:none}
     1. Install descheduler operator.
     2. Clone https://github.ibm.com/OpenShift-on-Z/ocp_test_automation -b madeel-pipes
  branch
     3. Run the descheduler e2e test suit
     {code}
 Actual results:
 {code:none}
      TestRemoveDuplicates, TestLiveMigrationInBackground, TestFailedPods, TestLowNodeUtilizationKubernetesMetrics, TestTooManyRestarts, TestTopologySpreadConstraint failed{code}
 Expected results:
 {code:none}
     All the tests in e2e suit should pass.{code}
 Additional info:
 {code:none}
  1. TestRemoveDuplicates - exec container process `/pause`: Exec format error
 
 2. TestLiveMigrationInBackground - Unable to get kubevirt/kubevirt: the server could not find the requested resource (get kubevirts.kubevirt.io kubevirt)
 
 3. TestFailedPods - Error from server (BadRequest): container "pause" in pod "test-failed-pods-default-args-5ll2b" is waiting to start: CreateContainerError - Error: container create failed: executable file `/bin/false` not found: No such file or directory
 
 4. TestLowNodeUtilizationKubernetesMetrics - Error from server (BadRequest): container "pause" in pod "lownodeutilization-kubernetes-metrics-pod-86cc8687d5-8fzkv" is waiting to start: ImageInspectError - Failed to inspect image "": rpc error: code = Unknown desc = short name mode is enforcing, but image name narmidm/k8s-pod-cpu-stressor:latest returns ambiguous list
 
 5. TestTooManyRestarts - Error from server (BadRequest): container "pause" in pod "restart-pod-5f974fbc98-l8kp8" is waiting to start: CreateContainerError - Error: container create failed: executable file `/bin/false` not found: No such file or directory
 
 6. TestTopologySpreadConstraint - exec container process `/pause`: Exec format error{code}
Status: CLOSED
#OCPBUGS-74405issue38 hours ago[4.20] Unable to provision NVIDIA DGX B200 nodes via ACM, Redfish VirtualMedia InsertMedia request fails with missing UserName/Password parameters Verified
Issue 17801333: [4.20] Unable to provision NVIDIA DGX B200 nodes via ACM, Redfish VirtualMedia InsertMedia request fails with missing UserName/Password parameters
Description: *Description of problem:*
 
 
 When attempting to provision NVIDIA DGX B200 nodes through Advanced Cluster Management (ACM), the BareMetalHost (BMH) status hangs at {_}Provisioning{_}.
 
 The Bare Metal Operator (BMO) and Ironic logs show Redfish VirtualMedia mounting failures. The BMC rejects the {{VirtualMedia.InsertMedia}} request with an HTTP 400 error, requiring {{UserName}} and {{Password}} parameters in the request body.
 
 This behavior prevents automated provisioning of DGX B200 nodes.
 ----
 *Version-Release number of selected component (if applicable):*
  * OpenShift: 4.18.3 (hub cluster)
 
  * ACM: 2.13
 
  * Ironic/BMO components from OCP 4.18.3 release
 
 ----
 *How reproducible:*
 Always – occurs consistently when attempting to provision NVIDIA DGX B200 nodes.
 ----
 *Steps to Reproduce:*
  # Register NVIDIA DGX B200 node in ACM using BareMetalHost (via BMC/Redfish).
 
  # Observe provisioning workflow.
 
  # Check Ironic/BMO logs when provisioning fails.
 
 ----
 *Actual results:*
  * BMH stuck in _Provisioning_ state.
 
  * ISO not mounted via Redfish VirtualMedia.
 
  * BMO/Ironic logs show:
 
  
 {code:java}
 HTTP request: POST https://<bmc>/redfish/v1/Managers/BMC/VirtualMedia/CD1/Actions/VirtualMedia.InsertMedia body: {'Image': 'https://<iso-server>/boot-<uuid>.iso', 'TransferProtocolType': 'HTTPS'}{code}
 {{ }}
 Response:
  
 {code:java}
 status code: 400 error: Base.1.12.GeneralError: The action ... requires the parameter UserName to be present in the request body. MessageId: Base.1.12.ActionParameterMissing{code}
 {{ }}
 ----
 *Expected results:*
  * Redfish VirtualMedia ISO should mount successfully.
 
  * BMH should proceed past _Provisioning_ and complete host registration.
 
 ----
 *Additional info:*
  * The fix for handling {{UserName}} in Redfish InsertMedia was merged upstream (sushy) and included in OCP 4.18.1 ({{{}Provide vmedia username and password if required{}}}).
 
  * However, in this scenario, the firmware’s response differs: the error lacks a structured {{code}} field, causing credential detection logic to fail.
 
  * Engineering suspects a firmware bug, but BMO/Ironic may require additional handling (e.g., checking {{MessageId=ActionParameterMissing}} or the {{error}} field directly) as a workaround.
Status: Verified
#OCPBUGS-77853issue5 days ago[openshift-4.18] multus-cni can't be built hermetically POST
Issue 17959374: [openshift-4.18] multus-cni can't be built hermetically
Description: Description of problem:
 
 ART can't build multus-cni hermetically. Same applies for multus-cni-container-microshift
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.    {code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: POST
#OCPBUGS-61298issue8 hours ago[ACM] ACM cannot access ironic-python-agent on dual-stack environment but only IPv6 is unreachable ON_QA
Issue 17365535: [ACM] ACM cannot access ironic-python-agent on dual-stack environment but only IPv6 is unreachable
Description: Description of problem:
 {code:none}
 Baremetal host is not detected randomly when we meet the following conditions:
 - ACM Hub Cluster is IPv4/IPv6 dual-stack
 - Managed cluster and NMStateConfig is IPv4/IPv6 dual-stack
 - Managed nodes are added by BMH resource, not discovery ISO
 - IPv4 is reachable between ACM and managed nodes
 - IPv6 is unreachable between ACM and managed nodes
 
 In this case, when we create a BMH resource, the managed node is not detected well randomly.
 ironic-python-agent has some issues, agent.service doesn't start, and Agent resource is not created.
 
 ironic-python-agent sometimes advertises an IPv6 URL to ACM:
 
 ~~~
 ironic-agent[XXX]: 2025-01-01 00:00:00.000 1 DEBUG ironic_python_agent.ironic_api_client [-] Heartbeat: announcing callback URL https://[2001:db8::1]:9999, API version is 1.68 heartbeat /usr/lib/python3.9/site-packages/ironic_python_agent/ironic_api_client.py:186
 ~~~
 
 But as IPv6 is unreachable, ACM metal3-ironic shows the following error:
 ~~~
 2025-01-01 00:00:00.000 1 ERROR ironic.drivers.modules.agent_client [-] Failed to connect to the agent running on node XXXXXXXX to collect commands status. Error: HTTPSConnectionPool(host='2001:db8::1', port=9999): Max retries exceeded with url: /v1/commands/ (Caused by ConnectTimeoutError(<urllib3.connection.HTTPSConnection object at 0xXXXXXX>, 'Connection to 2001:db8::1 timed out. (connect timeout=60)')): requests.exceptions.ConnectTimeout: HTTPSConnectionPool(host='2001:db8::1', port=9999): Max retries exceeded with url: /v1/commands/ (Caused by ConnectTimeoutError(<urllib3.connection.HTTPSConnection object at 0xXXXXXX>, 'Connection to 2001:db8::1 timed out. (connect timeout=60)'))
 ~~~
 
 Then agent.service doesn't start on the managed node and the node is not detected by ACM.
 
 But sometimes ironic-python-agent advertises an IPv4 URL to ACM.
 In this case, everything works well.
 The issue occurrence is random.
 ~~~
 ironic-agent[XXX]: 2025-01-01 00:00:00.000 1 DEBUG ironic_python_agent.ironic_api_client [-] Heartbeat: announcing callback URL https://192.0.2.1:9999, API version is 1.68 heartbeat /usr/lib/python3.9/site-packages/ironic_python_agent/ironic_api_client.py:186
 ~~~{code}
 Version-Release number of selected component (if applicable):
 {code:none}
 OCP 4.18
 ACM 2.13{code}
 How reproducible:
 
 Steps to Reproduce:
 {code:none}
 1. Prepare ACM hub cluster and managed nodes. IPv4 is reachable between ACM and managed nodes, but IPv6 is not reachable.
 2. Create an ACM Hub Cluster with IPv4/IPv6 dual-stack 3. Create NMStateConfig resource with IPv4/IPv6 dual-stack setting for managed nodes
 4. Create InfraEnv resource
 5. Create a BMH resoruce
 6. The managed node starts automatically, but the node is not detected by the above errors randomly{code}
 Actual results:
 {code:none}
 ACM metal3-ironic cannot connect to the managed nodes{code}
 Expected results:
 {code:none}
 It works with IPv4 even if IPv6 is not reachable{code}
 Workaround:
 {code:none}
 We can specify the advertised IP address using advertise_host setting.
 The issue can be solved by manually setting it on the managed node.
 
 ~~~
 # vi /etc/ironic-python-agent.conf
    :
 [DEFAULT]
 advertise_host = 192.0.2.1
     :
 
 # systemctl restart ironic-agent.service
 ~~~
 
 However, we need to make this config change manually on each managed node after the issue occurs.{code}
 Additional info:
 {code:none}
 I checked the source code.
 The advertised IP address is determined by the following code:
 
 https://github.com/openstack/ironic-python-agent/blob/master/ironic_python_agent/agent.py#L344
 https://github.com/openstack/ironic-python-agent/blob/master/ironic_python_agent/agent.py#L305-L324
 ~~~
     def _find_routable_addr(self):
         ips = set()
         for api_url in self.api_urls:
             ironic_host = urlparse.urlparse(api_url).hostname
             # Try resolving it in case it's not an IP address
             try:
                 addrs = socket.getaddrinfo(ironic_host, 0)
             except socket.gaierror:
                 LOG.debug('Could not resolve %s, maybe no DNS', ironic_host)
                 ips.add(ironic_host)
                 continue
             ips.update(addr for _, _, _, _, (addr, *_) in addrs)
 
         for attempt in range(self.ip_lookup_attempts):
             for ironic_host in ips:
                 found_ip = self._get_route_source(ironic_host)
                 if found_ip:
                     return found_ip
 
             time.sleep(self.ip_lookup_sleep)
 ~~~
 
 api_urls has both IPv4 address and IPv6 address in dual-stack environment
 
 ~~~
 # cat /etc/ironic-python-agent.conf
 
 [DEFAULT]
 api_url = https://192.0.2.2:6385,https://[2001:db8::2]:6385
   :
 ~~~
 
 The above code determines the advertised IP address using "ip route" command.
 It doesn't check the actual connectivity.
 That's why unreachable IPv6 address can be selected.
 
 Additionally, it has "ips = set()".
 As set() doesn't keep the order of its elements, it randomly returns IPv4 address or IPv6 address in the subsequent "for" loop.
 If it returns IPv4 first, IPv4 URL is advertised and it works.
 If it returns IPv6 first, IPv6 URL is advertised and it doesn't work.
 That's why this issue occurrence is random.{code}
 Fix ideas:
 {code:none}
 I came up with some ideas to fix this issue.
 What do you think?
 
 A: In the _find_routable_addr() function, check the actual network connectivity, not only checking the routing table.
    If it's not reachable, use another IP address.
    I think this is the easiest and best fix.
 
 B: Add some parameters to InfraEnv CR to specify which network is used for the provisioning.
    If we can specify the reachable network address in the CR, it will avoid the issue.
    But I understand that this will be an RFE.
 
 C: Add some parameters to add KernelArguments in BMH resource.
    Ironic-python-agent can get setting parameters from kernel arguments, and currently InfraEnv can add KernelArguments to DisocveryISO.
    But to avoid the issue, we need to set different values to ipa-advertise-host per BMH.
    InfraEnv cannot set different settings per node.
    If BMH can have a parameter to set KernelArgs, we will be able to set different ipa-advertise-host per host, and we can avoid the issue.
    But I understand that this is also an RFE.{code}
Status: ON_QA
#OCPBUGS-77988issue31 hours agoComponent Readiness: failed to pull test image for DRA test New
Issue 17968761: Component Readiness: failed to pull test image for DRA test
Description: I0309 06:23:43.998820 3396 dump.go:53] At 2026-03-09 06:13:10 +0000 UTC - event for sample-device-plugin-beta: {daemonset-controller } SuccessfulCreate: Created pod: sample-device-plugin-beta-v7dgf I0309 06:23:43.998837 3396 dump.go:53] At 2026-03-09 06:13:12 +0000 UTC - event for sample-device-plugin-beta-v7dgf: {kubelet worker-0.ostest.test.metalkube.org} Failed: Failed to pull image "virthost.ostest.test.metalkube.org:5000/localimages/local-test-image:e2e-registry-k8s-io-e2e-test-images-sample-device-plugin-1-7-ULwza-sZKxhdAQs1": unable to pull image or OCI artifact: pull image err: initializing source docker://virthost.ostest.test.metalkube.org:5000/localimages/local-test-image:e2e-registry-k8s-io-e2e-test-images-sample-device-plugin-1-7-ULwza-sZKxhdAQs1: reading manifest e2e-registry-k8s-io-e2e-test-images-sample-device-plugin-1-7-ULwza-sZKxhdAQs1 in virthost.ostest.test.metalkube.org:5000/localimages/local-test-image: manifest unknown; artifact err: get manifest: build image source: reading manifest e2e-registry-k8s-io-e2e-test-images-sample-device-plugin-1-7-ULwza-sZKxhdAQs1 in virthost.ostest.test.metalkube.org:5000/localimages/local-test-image: manifest unknown I0309 06:23:43.998855 3396 dump.go:53] At 2026-03-09 06:13:12 +0000 UTC - event for sample-device-plugin-beta-v7dgf: {kubelet worker-0.ostest.test.metalkube.org} Pulling: Pulling image "virthost.ostest.test.metalkube.org:5000/localimages/local-test-image:e2e-registry-k8s-io-e2e-test-images-sample-device-plugin-1-7-ULwza-sZKxhdAQs1" I0309 06:23:43.998873 3396 dump.go:53] At 2026-03-09 06:13:12 +0000 UTC - event for sample-device-plugin-beta-v7dgf: {kubelet worker-0.ostest.test.metalkube.org} Failed: Error: ErrImagePull I0309 06:23:43.998890 3396 dump.go:53] At 2026-03-09 06:13:12 +0000 UTC - event for sample-device-plugin-beta-v7dgf: {multus } AddedInterface: Add eth0 [10.128.2.46/23 fd01:0:0:5::2e/64] from ovn-kubernetes I0309 06:23:43.998908 3396 dump.go:53] At 2026-03-09 06:13:13 +0000 UTC - event for sample-device-plugin-beta-v7dgf: {kubelet worker-0.ostest.test.metalkube.org} BackOff: Back-off pulling image "virthost.ostest.test.metalkube.org:5000/localimages/local-test-image:e2e-registry-k8s-io-e2e-test-images-sample-device-plugin-1-7-ULwza-sZKxhdAQs1" I0309 06:23:43.998923 3396 dump.go:53] At 2026-03-09 06:13:13 +0000 UTC - event for sample-device-plugin-beta-v7dgf: {kubelet worker-0.ostest.test.metalkube.org} Failed: Error: ImagePullBackOff I0309 06:23:43.998942 3396 dump.go:53] At 2026-03-09 06:23:12 +0000 UTC - event for dra-test-driver: {replicaset-controller } SuccessfulDelete: Deleted pod: dra-test-driver-hnqkt I0309 06:23:43.998957 3396 dump.go:53] At 2026-03-09 06:23:12 +0000 UTC - event for dra-test-driver-hnqkt: {kubelet worker-0.ostest.test.metalkube.org} Killing: Stopping container pause   (_Feel free to update this bug's summary to be more specific._) Component Readiness has found a potential regression in the following test:  {code:none}[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] on single node supports extended resources together with ResourceClaim [Serial]{code}  Test has a 0.00% pass rate, but 95.00% is required.  Sample (being evaluated) Release: 4.22 Start Time: 2026-03-02T00:00:00Z End Time: 2026-03-09T12:00:00Z Success Rate: 0.00% Successes: 0 Failures: 9 Flakes: 0 Base (historical) Release: 4.21 Start Time: 2026-01-04T00:00:00Z End Time: 2026-02-03T23:59:59Z Success Rate: 0.00% Successes: 0 Failures: 0 Flakes: 0  View the [test details report|https://sippy-auth.dptools.openshift.org/sippy-ng/component_readiness/test_details?Architecture=amd64&FeatureSet=techpreview&Installer=ipi&LayeredProduct=none&Network=ovn&Platform=metal&Suite=serial&Topology=ha&Upgrade=none&baseEndTime=2026-02-03T23%3A59%3A59Z&baseRelease=4.21&baseStartTime=2026-01-04T00%3A00%3A00Z&capability=DynamicResourceAllocation&columnGroupBy=Network%2CPlatform%2CTopology&component=Node+%2F+Kubelet&confidence=95&dbGroupBy=Architecture%2CFeatureSet%2CInstaller%2CLayeredProduct%2CNetwork%2CPlatform%2CSuite%2CTopology%2CUpgrade&environment=Architecture%3Aamd64+FeatureSet%3Atechpreview+Installer%3Aipi+LayeredProduct%3Anone+Network%3Aovn+Platform%3Ametal+Suite%3Aserial+Topology%3Aha+Upgrade%3Anone&flakeAsFailure=false&ignoreDisruption=true&ignoreMissing=false&includeMultiReleaseAnalysis=true&includeVariant=Architecture%3Aamd64&includeVariant=Architecture%3Aarm64&includeVariant=Architecture%3Amulti&includeVariant=CGroupMode%3Av2&includeVariant=ContainerRuntime%3Acrun&includeVariant=ContainerRuntime%3Arunc&includeVariant=FeatureSet%3Adefault&includeVariant=FeatureSet%3Atechpreview&includeVariant=Installer%3Ahypershift&includeVariant=Installer%3Aipi&includeVariant=Installer%3Aupi&includeVariant=JobTier%3Ablocking&includeVariant=JobTier%3Ainforming&includeVariant=JobTier%3Astandard&includeVariant=LayeredProduct%3Anone&includeVariant=LayeredProduct%3Avirt&includeVariant=Network%3Aovn&includeVariant=OS%3Arhcos9&includeVariant=Owner%3Aeng&includeVariant=Owner%3Aservice-delivery&includeVariant=Platform%3Aaws&includeVariant=Platform%3Aazure&includeVariant=Platform%3Agcp&includeVariant=Platform%3Ametal&includeVariant=Platform%3Arosa&includeVariant=Platform%3Avsphere&includeVariant=Topology%3Aexternal&includeVariant=Topology%3Aha&includeVariant=Topology%3Amicroshift&includeVariant=Upgrade%3Amicro&includeVariant=Upgrade%3Aminor&includeVariant=Upgrade%3Anone&minFail=3&passRateAllTests=0&passRateNewTests=95&pity=5&sampleEndTime=2026-03-09T12%3A00%3A00Z&sampleRelease=4.22&sampleStartTime=2026-03-02T00%3A00%3A00Z&testId=openshift-tests%3A822a302164b8b59711e2e93707ba2cf8] for additional context.               Filed by: [~cmeadors@redhat.com]
Status: ASSIGNED
#OCPBUGS-76318issue5 days agoConfusing prompt "groups.user.openshift.io are not enabled on this cluster" Verified
Issue 17846259: Confusing prompt "groups.user.openshift.io are not enabled on this cluster"
Description: Description of problem:
 {code:none}
 Confusing prompt "groups.user.openshift.io are not enabled on this cluster"
 {code}
 Version-Release number of selected component (if applicable):
 {code:none}
 $ oc version
 Client Version: 4.22.0-0.nightly-2026-02-06-032407
 Kustomize Version: v5.7.1
 Server Version: 4.22.0-0.nightly-2026-02-03-002928
 Kubernetes Version: ...
 {code}
 How reproducible:
 {code:none}
 Always
 {code}
 Steps to Reproduce:
 {code:none}
 1. Launch a 4.22 fresh env
 
 2. $ oc adm groups new testgroup
 group.user.openshift.io/testgroup created
 
 3. $ oc adm groups add-users testgruop testuser-1
 error: Error: groups.user.openshift.io are not enabled on this cluster.
 Is the embeedded OAuth server disabled?.
 
 Original error:
 %!w(<nil>)
 
 {code}
 Actual results:
 {code:none}
 Confusing prompt. The Groups are indeed enabled.
 {code}
 Expected results:
 {code:none}
 Message should not be incorrect and confusing
 {code}
 Additional info:
 {code:none}
  {code}
Status: Verified
#OCPBUGS-52324issue8 days agoInspection error when installing baremetal cluster on HP ProLiant DL360 Gen10 hardware running iLO5 v3.08 CLOSED
Issue 16713141: Inspection error when installing baremetal cluster on HP ProLiant DL360 Gen10 hardware running iLO5 v3.08
Description: Description of problem:
 {code:none}
 Customer attempted to install baremetal cluster on HP ProLiant DL360 Gen10 hardware running iLO5 v3.08. The cluster install status was hanging on baremetalhost:cp0: inspecting. 
 
 The status of the baremetal resource was: 'Failed to inspect hardware. Reason: unable to start inspection: Redfish exception occurred. Error: Setting boot mode to uefi failed for node 813b617e-42f9-48db-99e9-1b2f3a35f94b. Error: HTTP PATCH https://10.64.143.135/redfish/v1/Systems/1 returned code 400. iLO.0.10.ExtendedInfo: See @Message.ExtendedInfo for more information. Extended information: [{''MessageId'': ''iLO.2.24.UnsupportedOperation''}]'     
 
 They tried the workaround to manually set the BIOS mode to UEFI first on each node before running the installer (here’s a Bugzilla bug for it https://bugzilla.redhat.com/show_bug.cgi?id=1820337) which says it is resolved but they would like to check if it was tested using a hardware that already had the mode set to UEFI. 
 
 HPE support advised customer that ilo API endpoint is READONLY and does not work in any version of ILO5 because ILO 5 implements `BootSourceOverrideMode` as a read-only item that reflects the current boot mode. 
 Reference: https://servermanagementportal.ext.hpe.com/docs/redfishservices/ilos/ilo5/ilo5_adaptation/#boot-source-override 
 
 Slack discussion: https://redhat-internal.slack.com/archives/CFP6ST0A3/p1734579771973629{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: CLOSED
#OCPBUGS-77444issue8 days agoTopology page loading with error after installed Virtualization operator ASSIGNED
Issue 17940088: Topology page loading with error after installed Virtualization operator
Description: Description of problem:
 {code:none}
     Topology page loading with error after installed Virtualization operator{code}
 Version-Release number of selected component (if applicable):
 {code:none}
 4.21.0-0.nightly-2026-02-26-211439    {code}
 How reproducible:
 {code:none}
     Always{code}
 Steps to Reproduce:
 {code:none}
     Make sure Virtualization operator has been installed on the cluster
     1. Navigate to Workloads -> Topology page
     2. Select a Project (eg: openshift-cnv, or others)
     3. Choose list view, Click the resources that list on the page 
     {code}
 Actual results:
 {code:none}
     Error: Something wrong happened(0 , r.getModifyApplicationAction) is not a function
 
 TypeErrorDescription:(0 , r.getModifyApplicationAction) is not a functionComponent trace:at o (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:186784)
     at v (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:187346)
     at S (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:184138)
     at f (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:1596403)
     at div
     at o (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:3216981)
     at div
     at o (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:3216626)
     at div
     at o (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:3666721)
     at div
     at o (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:3666277)
     at div
     at l (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:3662004)
     at div
     at l (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:3660900)
     at section
     at v (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:3439814)
     at g (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:3199640)
     at div
     at p (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:342999)
     at div
     at d (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/15646-chunk-d51f2bcad5f4e16934fd.min.js:1:57545)
     at div
     at a (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/15646-chunk-d51f2bcad5f4e16934fd.min.js:1:54932)
     at c (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/15646-chunk-d51f2bcad5f4e16934fd.min.js:1:52460)
     at div
     at div
     at l (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:52:160317)
     at div
     at div
     at a (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:3609865)
     at v (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:3284786)
     at d (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/15646-chunk-d51f2bcad5f4e16934fd.min.js:1:54256)
     at div
     at o (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:3283992)
     at h (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:3282768)
     at div
     at d (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:3281064)
     at div
     at div
     at o (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:3667597)
     at div
     at o (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:3667206)
     at div
     at https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/15646-chunk-d51f2bcad5f4e16934fd.min.js:1:44828
     at S (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/72133-bundle-9dbb1ef6b0b5f1bf2c3e.min.js:2:18753)
     at I (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:2414861)
     at https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/15646-chunk-d51f2bcad5f4e16934fd.min.js:1:38140
     at https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:2392338
     at div
     at A (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:471365)
     at j (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:471805)
     at f (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:1596403)
     at https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/15646-chunk-d51f2bcad5f4e16934fd.min.js:1:41548
     at https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:1175966
     at div
     at div
     at i (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/15646-chunk-d51f2bcad5f4e16934fd.min.js:1:5900)
     at f (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/15646-chunk-d51f2bcad5f4e16934fd.min.js:1:65177)
     at d (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:854733)
     at https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/15646-chunk-d51f2bcad5f4e16934fd.min.js:1:41875
     at i (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:277483)
     at withFallback()
     at r (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/dev-console/code-refs/topologyPage-chunk-e08f224937cee6e37e3e.min.js:1:2180)
     at b (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:75336)
     at f (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:75681)
     at S (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:75855)
     at _ (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/72133-bundle-9dbb1ef6b0b5f1bf2c3e.min.js:2:51105)
     at ne (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/72133-bundle-9dbb1ef6b0b5f1bf2c3e.min.js:2:56537)
     at Suspense
     at i (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:277483)
     at m
     at section
     at v (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:3439814)
     at W (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:933627)
     at main
     at div
     at div
     at div
     at o (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:3283992)
     at h (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:3282768)
     at div
     at d (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:3281064)
     at div
     at div
     at w (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:3431896)
     at div
     at l (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:3660900)
     at c (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:875511)
     at div
     at o (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:3283313)
     at div
     at div
     at o (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:3283992)
     at h (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:3282768)
     at div
     at d (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:3281064)
     at Sn (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:3115697)
     at i (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:54467)
     at a (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:54036)
     at fe (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:946219)
     at tt (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:3074630)
     at fe (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:946219)
     at fe (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:946219)
     at u (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/api/plugins/kubevirt-plugin/exposed-modalProvider-chunk-5f1caebfdb41b6e7076c.min.js:1:423)
     at fe (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:946219)
     at fe (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:946219)
     at Suspense
     at f (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:112989)
     at b (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:111873)
     at a (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:23099)
     at u (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:25343)
     at ye (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:946335)
     at Se (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:949610)
     at _ (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/72133-bundle-9dbb1ef6b0b5f1bf2c3e.min.js:2:51105)
     at ne (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/72133-bundle-9dbb1ef6b0b5f1bf2c3e.min.js:2:56537)
     at _ (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/72133-bundle-9dbb1ef6b0b5f1bf2c3e.min.js:2:51105)
     at ne (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/72133-bundle-9dbb1ef6b0b5f1bf2c3e.min.js:2:56537)
     at te (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/72133-bundle-9dbb1ef6b0b5f1bf2c3e.min.js:2:55950)
     at G (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/72133-bundle-9dbb1ef6b0b5f1bf2c3e.min.js:2:43270)
     at t (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/72133-bundle-9dbb1ef6b0b5f1bf2c3e.min.js:2:65457)
     at xe (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:949902)
     at g (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:478028)
     at c (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:86080)
     at e (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:2594154)
     at v (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/main-bundle-6e23ac1e136e272a2f6a.min.js:1:926590)
     at c (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/72133-bundle-9dbb1ef6b0b5f1bf2c3e.min.js:2:16683)
     at SuspenseStack trace:TypeError: (0 , r.getModifyApplicationAction) is not a function
     at https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/api/plugins/kubevirt-plugin/exposed-topology-chunk-04d21a7cb3ef7251ab91.min.js:1:65957
     at Object.useMemo (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/vendors~main-chunk-37548-cefc1afc0027db257b90.min.js:2:2517670)
     at t.useMemo (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/static/72133-bundle-9dbb1ef6b0b5f1bf2c3e.min.js:2:78415)
     at m (https://console-openshift-console.apps.qe-uidaily-0227.qe.azure.devcluster.openshift.com/api/plugins/kubevirt-plugin/exposed-{code}
 Expected results:
 {code:none}
     No errors, resources can loaded normally{code}
 Additional info:
 {code:none}
     {code}
Status: ASSIGNED
#OCPBUGS-78162issue9 hours ago[openshift-4.13] ose-metallb-operator can't be built hermetically POST
Issue 17974682: [openshift-4.13] ose-metallb-operator can't be built hermetically
Description: This is a clone of issue OCPBUGS-78161. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-metallb-operator hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: POST
#OCPBUGS-77214issue12 days ago[openshift-4.12] hermetic build failure - ose-ibm-vpc-block-csi-driver ASSIGNED
Issue 17926364: [openshift-4.12] hermetic build failure - ose-ibm-vpc-block-csi-driver
Description: This is a clone of issue OCPBUGS-77213. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77212. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77211. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77210. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77209. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77207. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77206. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77205. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77203. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77167. The following is the description of the original issue:  --- Description of problem:
 {code:none}
 IBM VPC Block CSI Driver build fails in hermetic mode due to non-deterministic GOPATH handling in Makefile. The build system rewrites $GOPATH causing the binary output path -o ${GOPATH}/bin/${EXE_DRIVER_NAME} to become invalid, resulting in build failures.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
 OpenShift 4.19+ (affects hermetic build conversion) {code}
 How reproducible:
 {code:none}
 Always - when building ibm-vpc-block-csi-driver in hermetic mode{code}
 Steps to Reproduce:
 {code:none}
   1. Build ibm-vpc-block-csi-driver container in hermetic mode
   2. Makefile uses absolute GOPATH-dependent output path: -o ${GOPATH}/bin/${EXE_DRIVER_NAME}
   3. Build system rewrites $GOPATH during hermetic build process     {code}
 Actual results:
 {code:none}
 Build fails because binary cannot be found at expected GOPATH location after build system modifies $GOPATH
 
 Error: building at STEP "COPY --from=builder /cachi2/output/deps/gomod/bin/ibm-vpc-block-csi-driver /bin/ibm-vpc-block-csi-driver": checking on sources under [...]: copier: stat: "/cachi2/output/deps/gomod/bin/ibm-vpc-block-csi-driver": no such file or directory{code}
 Expected results:
 {code:none}
 Build succeeds using relative path bin/${EXE_DRIVER_NAME} that doesn't depend on GOPATH manipulation, consistent with other CSI drivers like aws-ebs-csi-driver  {code}
 Additional info:
 {code:none}
 Fix implemented: Change Makefile to use relative path -o bin/${EXE_DRIVER_NAME} instead of absolute GOPATH-dependent path. This aligns with other CSI drivers and makes builds hermetic-compatible. Comment added to indicate this is a permanent carry patch for OpenShift builds.{code}
Status: ASSIGNED
#OCPBUGS-39360issue8 days agoMany tests failing on DualStack IPv6 primary POST
{  fail [k8s.io/kubernetes/test/e2e/network/endpointslice.go:611]: EndpointSlice.discovery.k8s.io "e2e-custom-slicel8znj" is invalid: endpoints[0].addresses[0]: Invalid value: "fd01:0:0:4::951": must be a valid IPv4 address
Error: exit with code 1 {code}
We seem to not be informing the tests that v6 is primary.
#OCPBUGS-76625issue11 days agoMicroshift pods fails to start with error container create failed: systemd failed to install eBPF device filter on cgroup New
Issue 17870004: Microshift pods fails to start with error container create failed: systemd failed to install eBPF device filter on cgroup
Description: Description of problem:{code:none}
 When running iso installs and ostree z or y upgrades i see microshift pods goes into createContainerError state and doing a describe on the pods shows error below
 
 openshift-ovn-kubernetes               3h20m       Warning   Failed              pod/ovnkube-node-gscpl                          (combined from similar events): Error: container create failed: systemd failed to install eBPF device filter on cgroup `/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2535c48_e30b_4620_93c2_afc19d1685c1.slice/crio-34e6ac00324e13e50481a795f54b5a1ace01bf2e8c186a75b407700d0ca66e5b.scope`
     {code}
     Version-Release number of selected component (if applicable):{code:none}
      4.18.33
      microshift-4.18.33-202602090801.p0.ge2f5f13.assembly.4.18.33.el9.x86_64
     {code}
     How reproducible:{code:none}
     always
     {code}
     Steps to Reproduce:{code:none}
     1. Trigger iso installs and ostree z and y upgrades with  microshift release testing pipeline.
     2.
     3.
     {code}
     Actual results:{code:none}
 Microshift pods fails to start with error  as described in the description
     {code}
     Expected results:{code:none}
 Microshift pods should start fine.
     {code}
     Additional info:{code:none}
 https://redhat-internal.slack.com/archives/C03CJTNLKAT/p1770818291495579
 
     {code}
Status: ASSIGNED
#OCPBUGS-74573issue5 hours agoMachine creation fails on ASH when using AzureManaged boot diagnostics without proper validation Verified
Issue 17811325: Machine creation fails on ASH when using AzureManaged boot diagnostics without proper validation
Description: Description of problem:
 {code:none}
 Before the fix https://github.com/openshift/machine-api-provider-azure/pull/182 
 When creating a Machine/MachineSet on Azure Stack Hub with diagnostics.boot.storageAccountType: AzureManaged, the Machine creation succeed, but in fact, it doesn't take affect.
 
 After the fix https://github.com/openshift/machine-api-provider-azure/pull/182  When creating a Machine/MachineSet on Azure Stack Hub with diagnostics.boot.storageAccountType: AzureManaged, the Machine creation fails with a 400 error from Azure API, but there is no validation at the Machine API level to prevent this invalid configuration.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
 4.22{code}
 How reproducible:
 {code:none}
 Always {code}
 Steps to Reproduce:
 {code:none}
     1. On an Azure Stack Hub cluster, create a MachineSet with the following configuration:
 
           diagnostics:
             boot:
               storageAccountType: AzureManaged
 
     2. Check machine status
     3.
     {code}
 Actual results:
 {code:none}
 Machine create failed with error:
         RESPONSE 400: 400 Bad Request
         ERROR CODE: InvalidParameter
         --------------------------------------------------------------------------------
         {
           "error": {
             "code": "InvalidParameter",
             "message": "Required parameter 'bootDiagnostics.storageAccountUri' is missing (null).",
             "target": "bootDiagnostics.storageAccountUri"
           }
         }{code}
 Expected results:
 {code:none}
 Azure Stack Hub does not support Azure-managed boot diagnostics storage.
 So Machine/MachineSet creation should fail with validation error like "Error: Invalid Machine configuration: AzureManaged boot diagnostics is not supported on Azure Stack Hub"{code}
 Additional info:
 {code:none}
     {code}
Status: Verified
#OCPBUGS-78041issue34 hours ago[openshift-4.14] ose-powervs-machine-controllers can't be built hermetically New
Issue 17969587: [openshift-4.13] ose-powervs-machine-controllers can't be built hermetically
Description: This is a clone of issue OCPBUGS-78039. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78038. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-powervs-machine-controllers hermetically.
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-78026issue34 hours ago[openshift-4.12] ose-vsphere-cluster-api-controllers can't be built hermetically New
Issue 17969534: [openshift-4.12] ose-vsphere-cluster-api-controllers can't be built hermetically
Description: This is a clone of issue OCPBUGS-78024. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78023. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78021. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78020. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78019. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78018. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78017. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-vsphere-cluster-api-controllers hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-17155issue2 weeks agoConsole: Minified React error #306 after cluster upgrade CLOSED
Issue 15400366: Console: Minified React error #306 after cluster upgrade
Description: Description of problem:
 
 Admin Pesrpective of the cluster console -> Observe -> Metrics
 https://console.build02.ci.openshift.org/monitoring/query-browser
 
 The page does not work (see the screenshots for the error) after upgrading the cluster {{build02}} to 4.14.0-ec.4 with MacOS Chrome.
 It works with Firefox on the same laptop.
 
 The same step works for build01 with {{4.13.7}}.
 Client and Server version are listed below and the screenshots.
 
 {noformat}
 oc --context build02 version
 Client Version: 4.13.4
 Kustomize Version: v4.5.7
 Server Version: 4.14.0-ec.4
 Kubernetes Version: v1.27.3+4aaeaec
 
 ~ oc --context build01 version
 Client Version: 4.13.4
 Kustomize Version: v4.5.7
 Server Version: 4.13.7
 Kubernetes Version: v1.26.6+73ac561
 {noformat}
 
 {noformat}
 Error
 Description:4566Minified React error #306; visit https://reactjs.org/docs/error-decoder.html?invariant=306&args[]=null&args[]= for the full message or use the non-minified dev environment for full errors and additional helpful warnings.
 
 Component trace:
 at t (https://console.build02.ci.openshift.org/static/vendors~main-chunk-f91c32b721335a248446.min.js:21:71524)
     at I (https://console.build02.ci.openshift.org/static/main-chunk-db2d63d44701b9895477.min.js:1:1269039)
     at t (https://console.build02.ci.openshift.org/static/vendors~main-chunk-f91c32b721335a248446.min.js:21:73605)
     at Suspense
     at i (https://console.build02.ci.openshift.org/static/main-chunk-db2d63d44701b9895477.min.js:1:313033)
     at section
     at m (https://console.build02.ci.openshift.org/static/vendor-patternfly-core-chunk-31bef589fa2b3b6b8d69.min.js:1:170065)
     at div
     at div
     at t.a (https://console.build02.ci.openshift.org/static/main-chunk-db2d63d44701b9895477.min.js:1:1269268)
     at div
     at div
     at c (https://console.build02.ci.openshift.org/static/vendor-patternfly-core-chunk-31bef589fa2b3b6b8d69.min.js:1:699584)
     at d (https://console.build02.ci.openshift.org/static/vendor-patternfly-core-chunk-31bef589fa2b3b6b8d69.min.js:1:217958)
     at div
     at d (https://console.build02.ci.openshift.org/static/vendor-patternfly-core-chunk-31bef589fa2b3b6b8d69.min.js:1:89498)
     at c (https://console.build02.ci.openshift.org/static/main-chunk-db2d63d44701b9895477.min.js:1:1015663)
     at https://console.build02.ci.openshift.org/static/main-chunk-db2d63d44701b9895477.min.js:1:318678
     at x (https://console.build02.ci.openshift.org/static/vendors~main-chunk-f91c32b721335a248446.min.js:120:86684)
     at main
     at div
     at v (https://console.build02.ci.openshift.org/static/vendor-patternfly-core-chunk-31bef589fa2b3b6b8d69.min.js:1:466913)
     at div
     at div
     at c (https://console.build02.ci.openshift.org/static/vendor-patternfly-core-chunk-31bef589fa2b3b6b8d69.min.js:1:313082)
     at div
     at div
     at c (https://console.build02.ci.openshift.org/static/vendor-patternfly-core-chunk-31bef589fa2b3b6b8d69.min.js:1:699584)
     at d (https://console.build02.ci.openshift.org/static/vendor-patternfly-core-chunk-31bef589fa2b3b6b8d69.min.js:1:217958)
     at div
     at d (https://console.build02.ci.openshift.org/static/vendor-patternfly-core-chunk-31bef589fa2b3b6b8d69.min.js:1:89498)
     at Jn (https://console.build02.ci.openshift.org/static/vendors~main-chunk-f91c32b721335a248446.min.js:58:195416)
     at t.default (https://console.build02.ci.openshift.org/static/main-chunk-db2d63d44701b9895477.min.js:1:782052)
     at t.default (https://console.build02.ci.openshift.org/static/quick-start-chunk-807ebd742dc381028a39.min.js:1:1260)
     at s (https://console.build02.ci.openshift.org/static/main-chunk-db2d63d44701b9895477.min.js:1:149799)
     at t.a (https://console.build02.ci.openshift.org/static/main-chunk-db2d63d44701b9895477.min.js:1:1345565)
     at te (https://console.build02.ci.openshift.org/static/main-chunk-db2d63d44701b9895477.min.js:1:1363570)
     at xt (https://console.build02.ci.openshift.org/static/vendors~main-chunk-f91c32b721335a248446.min.js:58:152524)
     at te (https://console.build02.ci.openshift.org/static/main-chunk-db2d63d44701b9895477.min.js:1:1363570)
     at te (https://console.build02.ci.openshift.org/static/main-chunk-db2d63d44701b9895477.min.js:1:1363570)
     at o (https://console.build02.ci.openshift.org/static/main-chunk-db2d63d44701b9895477.min.js:1:725597)
     at t.a (https://console.build02.ci.openshift.org/static/main-chunk-db2d63d44701b9895477.min.js:1:1339708)
     at t.a (https://console.build02.ci.openshift.org/static/main-chunk-db2d63d44701b9895477.min.js:1:1339897)
     at t.a (https://console.build02.ci.openshift.org/static/main-chunk-db2d63d44701b9895477.min.js:1:1337382)
     at ne (https://console.build02.ci.openshift.org/static/main-chunk-db2d63d44701b9895477.min.js:1:1363701)
     at https://console.build02.ci.openshift.org/static/main-chunk-db2d63d44701b9895477.min.js:1:1366805
     at r (https://console.build02.ci.openshift.org/static/vendors~main-chunk-f91c32b721335a248446.min.js:58:206190)
     at t (https://console.build02.ci.openshift.org/static/vendors~main-chunk-f91c32b721335a248446.min.js:21:71524)
     at t (https://console.build02.ci.openshift.org/static/vendors~main-chunk-f91c32b721335a248446.min.js:21:73605)
     at P (https://console.build02.ci.openshift.org/static/vendors~main-chunk-f91c32b721335a248446.min.js:58:4488)
     at ne (https://console.build02.ci.openshift.org/static/vendors~main-chunk-f91c32b721335a248446.min.js:58:9932)
     at te (https://console.build02.ci.openshift.org/static/vendors~main-chunk-f91c32b721335a248446.min.js:58:9341)
     at Y (https://console.build02.ci.openshift.org/static/vendors~main-chunk-f91c32b721335a248446.min.js:115:11631)
     at t (https://console.build02.ci.openshift.org/static/vendors~main-chunk-f91c32b721335a248446.min.js:21:67954)
     at re (https://console.build02.ci.openshift.org/static/main-chunk-db2d63d44701b9895477.min.js:1:1367069)
     at t.a (https://console.build02.ci.openshift.org/static/main-chunk-db2d63d44701b9895477.min.js:1:699943)
     at t.a (https://console.build02.ci.openshift.org/static/main-chunk-db2d63d44701b9895477.min.js:1:949236)
     at s (https://console.build02.ci.openshift.org/static/main-chunk-db2d63d44701b9895477.min.js:1:511869)
     at t.a (https://console.build02.ci.openshift.org/static/vendors~main-chunk-f91c32b721335a248446.min.js:168:261025)
     at Suspense
 Stack trace:
 Error: Minified React error #306; visit https://reactjs.org/docs/error-decoder.html?invariant=306&args[]=null&args[]= for the full message or use the non-minified dev environment for full errors and additional helpful warnings.
     at Hs (https://console.build02.ci.openshift.org/static/vendors~main-chunk-f91c32b721335a248446.min.js:281:112221)
     at Sl (https://console.build02.ci.openshift.org/static/vendors~main-chunk-f91c32b721335a248446.min.js:281:98327)
     at Cl (https://console.build02.ci.openshift.org/static/vendors~main-chunk-f91c32b721335a248446.min.js:281:98255)
     at _l (https://console.build02.ci.openshift.org/static/vendors~main-chunk-f91c32b721335a248446.min.js:281:98118)
     at pl (https://console.build02.ci.openshift.org/static/vendors~main-chunk-f91c32b721335a248446.min.js:281:95105)
     at https://console.build02.ci.openshift.org/static/vendors~main-chunk-f91c32b721335a248446.min.js:281:44774
     at t.unstable_runWithPriority (https://console.build02.ci.openshift.org/static/vendors~main-chunk-f91c32b721335a248446.min.js:289:3768)
     at Uo (https://console.build02.ci.openshift.org/static/vendors~main-chunk-f91c32b721335a248446.min.js:281:44551)
     at Wo (https://console.build02.ci.openshift.org/static/vendors~main-chunk-f91c32b721335a248446.min.js:281:44719)
     at N (https://console.build02.ci.openshift.org/static/vendors~main-chunk-f91c32b721335a248446.min.js:289:2828)
 {noformat}
Status: CLOSED
#OCPBUGS-73918issue6 weeks ago[4.21][Hypershift] Ingress Operator Degraded due to 'ManagementCPUsOverride' admission rejection during initial bootstrap (0 worker nodes) ASSIGNED
  - Similarity: Same admission plugin (autoscaling.openshift.io/ManagementCPUsOverride) rejecting pod creation
  - Error: Error creating: pods "perf-node-gather-daemonset-" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the pod namespace does not allow the workload type management
  - Context: SNO with Telco DU profile during must-gather operations
#OCPBUGS-44026issue13 days agoBaremetalHosts stuck in 'provisioning' state despite deploy_step success New
Issue 16426241: BaremetalHosts stuck in 'provisioning' state despite deploy_step success
Description: h3. Description of problem:
 
 During a ZTP install (by [~rhn-gps-rspazzol]), the BMHs remain stuck in the "provisioning" state while the Agent CRs show "Pending user interaction".
 It seems possible that the cause is a restart of the metal3 Pod with a fresh ironic database during deployment.
 h3. Version-Release number of selected component (if applicable):
 
 Likely affects all extant versions of metal platform when used with the converged flow of ZTP.
 h3. How reproducible:
 
 Reportedly happens regularly; haven't done repeated tests with the reproduction steps below but likely to be 100%.
 h3. Steps to Reproduce:
  # Start provisioning a cluster with ZTP
  # While one or more BMHs are in the "provisioning" state, delete the metal3 pod
  # Observe the BMHs stuck in provisioning forever
 
 h3. Actual results:
 
 The assisted-installer completes writing CoreOS to disk successfully and stops the agent service ready to restart the host:
 {noformat}
 Oct 30 14:20:37 localhost.localdomain assisted-installer[45058]: time="2024-10-30T14:20:37Z" level=info msg="Updating node installation stage: Rebooting - Ironic will reboot the node shortly" request_id=9d003e7b-b2f1-49b2-8c72-f122c7b69a68
 Oct 30 14:20:37 localhost.localdomain assisted-installer[45058]: time="2024-10-30T14:20:37Z" level=info msg="Running systemctl stop [agent.service]"
 Oct 30 14:20:37 localhost.localdomain systemd[1]: Stopping agent.service...
 {noformat}
 However, by this point in time, IPA heartbeats have been failing for some time with the error:
 {noformat}
 Oct 30 14:19:21 localhost.localdomain podman[3035]: 2024-10-30 14:19:21.355 1 ERROR ironic_python_agent.agent [-] error sending heartbeat to ['https://10.9.49.125:6385']: ironic_python_agent.errors.HeartbeatError: Error heartbeating to agent API: Error 404: Node 6f7546a2-f49e-4d8d-88f6-a462d53868b6 could not be found.
 Oct 30 14:19:21 localhost.localdomain podman[3035]: 2024-10-30 14:19:21.355 1 ERROR ironic_python_agent.agent Traceback (most recent call last):
 Oct 30 14:19:21 localhost.localdomain podman[3035]: 2024-10-30 14:19:21.355 1 ERROR ironic_python_agent.agent   File "/usr/lib/python3.9/site-packages/ironic_python_agent/agent.py", line 148, in do_heartbeat
 Oct 30 14:19:21 localhost.localdomain podman[3035]: 2024-10-30 14:19:21.355 1 ERROR ironic_python_agent.agent     self.api.heartbeat(
 Oct 30 14:19:21 localhost.localdomain podman[3035]: 2024-10-30 14:19:21.355 1 ERROR ironic_python_agent.agent   File "/usr/lib/python3.9/site-packages/ironic_python_agent/ironic_api_client.py", line 200, in heartbeat
 Oct 30 14:19:21 localhost.localdomain podman[3035]: 2024-10-30 14:19:21.355 1 ERROR ironic_python_agent.agent     raise errors.HeartbeatError(error)
 Oct 30 14:19:21 localhost.localdomain podman[3035]: 2024-10-30 14:19:21.355 1 ERROR ironic_python_agent.agent ironic_python_agent.errors.HeartbeatError: Error heartbeating to agent API: Error 404: Node 6f7546a2-f49e-4d8d-88f6-a462d53868b6 could not be found.
 Oct 30 14:19:21 localhost.localdomain podman[3035]: 2024-10-30 14:19:21.355 1 ERROR ironic_python_agent.agent 
 {noformat}
 Checking the BMH object, it has _not_ been disconnected from ironic, and it shows a _different_ provisioning ID from the node ID that IPA is looking up in ironic (6f7546a2-f49e-4d8d-88f6-a462d53868b6). This suggests that ironic has been restarted with a fresh database and a new node created with a fresh ID.
 
 Just after the installation succeeds, we also get some heartbeat errors of the form:
 {noformat}
 Oct 30 14:20:53 localhost.localdomain podman[3035]: 2024-10-30 14:20:53.707 1 ERROR ironic_python_agent.agent [-] error sending heartbeat to ['https://10.9.49.125:6385']: ironic_python_agent.errors.HeartbeatConnectionError: Error attempting to heartbeat - Possible transitory network failure or
  blocking port may be present.: ('Connection aborted.', RemoteDisconnected('Remote end closed connection without response'))
 {noformat}
 This continues to happen intermittently, but mostly the logs are full of the "node could not be found" errors.
 h3. Expected results:
 
 After the custom deploy method successfully completes, IPA should inform ironic of this:
 {noformat}
 Oct 30 14:20:48 localhost.localdomain podman[3035]: 2024-10-30 14:20:47.939 1 DEBUG ironic_coreos_install [-] Assisted Agent is in state inactive (result success) _is_assisted_running /usr/lib/python3.9/site-packages/ironic_coreos_install.py:136
 Oct 30 14:20:48 localhost.localdomain podman[3035]: 2024-10-30 14:20:47.947 1 INFO ironic_coreos_install [-] Succesfully installed using the assisted agent
 Oct 30 14:20:48 localhost.localdomain podman[3035]: 2024-10-30 14:20:47.951 1 INFO root [-] Deploy step completed: {'args': {}, 'interface': 'deploy', 'priority': 80, 'step': 'start_assisted_install'}, result: None
 Oct 30 14:20:48 localhost.localdomain podman[3035]: 2024-10-30 14:20:47.951 1 INFO root [-] Asynchronous command execute_deploy_step completed: {'deploy_result': None, 'deploy_step': {'args': {}, 'interface': 'deploy', 'priority': 80, 'step': 'start_assisted_install'}}
 {noformat}
 Once ironic knows that deployment is complete, it will reboot the host, which will come up as a Node and join the cluster, and the BMH provisioning state will change to {{{}provisioned{}}}.
 
 This would presumably work if we either restarted IPA or looked up the node ID again when heartbeating fails. (We may want to limit this behaviour to Metal³, since changes of node ID are not expected in regular OpenStack deployments.)
 h3. Additional info:
 
 Note that when an ironic node is in the {{deploying}} state, Metal³ cannot even attempt to deprovision the host (which is also required for deleting the BMH); it [must wait|https://github.com/metal3-io/baremetal-operator/blob/v0.8.0/pkg/provisioner/ironic/ironic.go#L1697-L1698] for the state to change to {{active}} (which won't happen when this bug is triggered) or {{deploy-wait}} (which doesn't seem to happen either). So this bug is a possible trigger of other reported bugs where deprovisioning never completes, such as OCPBUGS-38253.
 
 It seems likely that the same issue would affect regular (non-assisted) baremetal deployments using the {{install_coreos}} custom deploy method also.
Status: New
2024-10-30T14:19:30.936 creating PXE enabled ironic port for node {host: 'etl7~dl380g9-10', NodeUUID: '26d24895-bc25-4f33-8db6-06e35caa9fee', MAC: '00:11:0a:68:06:4c'}
2024-10-30T14:19:31.020 current provision state {host: 'etl7~dl380g9-10', lastError: '', current: 'enroll', target: ''}
2024-10-30T14:19:31.020 changing provisioning state {host: 'etl7~dl380g9-10', current: 'enroll', existing target: '', new target: 'manage'}
2024-10-30T14:19:41.425 pending PreprovisioningImage not ready {baremetalhost: {'name': 'dl380g9-10', 'namespace': 'etl7'}, provisioningState: 'provisioning'}
2024-10-30T14:19:41.461 current provision state {host: 'etl7~dl380g9-10', lastError: '', current: 'manageable', target: ''}
2024-10-30T14:19:41.461 updating option data {host: 'etl7~dl380g9-10', option: 'instance_uuid', value: '122f0196-c2af-4343-a459-5f3142811de4', oldValue: ''}
2024-10-30T14:24:41.670 pending PreprovisioningImage not ready {baremetalhost: {'name': 'dl380g9-10', 'namespace': 'etl7'}, provisioningState: 'provisioning'}
2024-10-30T14:24:41.719 current provision state {host: 'etl7~dl380g9-10', lastError: '', current: 'manageable', target: ''}
2024-10-30T14:24:41.719 done {baremetalhost: {'name': 'dl380g9-10', 'namespace': 'etl7'}, provisioningState: 'provisioning', requeue: True, after: 300}
ster-engine.apps.hub2.ocp.rht-labs.com/boot-artifacts/kernel?arch=x86_64&version=4.16', 'ExtraKernelParams': 'coreos.live.rootfs_url=https://assisted-image-service-multicluster-engine.apps.hub2.ocp.rht-labs.com/boot-artifacts/rootfs?arch=x86_64&version=4.16 rd.bootif=0', 'Format': 'initrd'}}
2024-10-30T14:04:38.045 current provision state {host: 'etl7~dl380g9-10', lastError: '', current: 'wait call-back', target: 'active'}
2024-10-30T14:04:38.045 provisioning {baremetalhost: {'name': 'dl380g9-10', 'namespace': 'etl7'}, provisioningState: 'provisioning'}
2024-10-30T14:04:44.729 pending PreprovisioningImage not ready {baremetalhost: {'name': 'dl380g9-10', 'namespace': 'etl7'}, provisioningState: 'provisioning'}
2024-10-30T14:04:44.776 current provision state {host: 'etl7~dl380g9-10', lastError: '', current: 'wait call-back', target: 'active'}
2024-10-30T14:04:44.777 provisioning {baremetalhost: {'name': 'dl380g9-10', 'namespace': 'etl7'}, provisioningState: 'provisioning'}
#OCPBUGS-77996issue35 hours ago[openshift-4.12] openshift-enterprise-console-operator can't be built hermetically New
Issue 17969152: [openshift-4.12] openshift-enterprise-console-operator can't be built hermetically
Description: This is a clone of issue OCPBUGS-77995. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77994. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77993. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77992. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77862. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77861. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77860. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build openshift-enterprise-console-operator hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-78157issue9 hours ago[openshift-4.13] ose-nutanix-machine-controllers can't be built hermetically POST
Issue 17974657: [openshift-4.13] ose-nutanix-machine-controllers can't be built hermetically
Description: This is a clone of issue OCPBUGS-78156. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-nutanix-machine-controllers hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: POST
#OCPBUGS-74406issue38 hours ago[4.19] Unable to provision NVIDIA DGX B200 nodes via ACM, Redfish VirtualMedia InsertMedia request fails with missing UserName/Password parameters POST
Issue 17801335: [4.19] Unable to provision NVIDIA DGX B200 nodes via ACM, Redfish VirtualMedia InsertMedia request fails with missing UserName/Password parameters
Description: *Description of problem:*
 
 
 When attempting to provision NVIDIA DGX B200 nodes through Advanced Cluster Management (ACM), the BareMetalHost (BMH) status hangs at {_}Provisioning{_}.
 
 The Bare Metal Operator (BMO) and Ironic logs show Redfish VirtualMedia mounting failures. The BMC rejects the {{VirtualMedia.InsertMedia}} request with an HTTP 400 error, requiring {{UserName}} and {{Password}} parameters in the request body.
 
 This behavior prevents automated provisioning of DGX B200 nodes.
 ----
 *Version-Release number of selected component (if applicable):*
  * OpenShift: 4.18.3 (hub cluster)
 
  * ACM: 2.13
 
  * Ironic/BMO components from OCP 4.18.3 release
 
 ----
 *How reproducible:*
 Always – occurs consistently when attempting to provision NVIDIA DGX B200 nodes.
 ----
 *Steps to Reproduce:*
  # Register NVIDIA DGX B200 node in ACM using BareMetalHost (via BMC/Redfish).
 
  # Observe provisioning workflow.
 
  # Check Ironic/BMO logs when provisioning fails.
 
 ----
 *Actual results:*
  * BMH stuck in _Provisioning_ state.
 
  * ISO not mounted via Redfish VirtualMedia.
 
  * BMO/Ironic logs show:
 
  
 {code:java}
 HTTP request: POST https://<bmc>/redfish/v1/Managers/BMC/VirtualMedia/CD1/Actions/VirtualMedia.InsertMedia body: {'Image': 'https://<iso-server>/boot-<uuid>.iso', 'TransferProtocolType': 'HTTPS'}{code}
 {{ }}
 Response:
  
 {code:java}
 status code: 400 error: Base.1.12.GeneralError: The action ... requires the parameter UserName to be present in the request body. MessageId: Base.1.12.ActionParameterMissing{code}
 {{ }}
 ----
 *Expected results:*
  * Redfish VirtualMedia ISO should mount successfully.
 
  * BMH should proceed past _Provisioning_ and complete host registration.
 
 ----
 *Additional info:*
  * The fix for handling {{UserName}} in Redfish InsertMedia was merged upstream (sushy) and included in OCP 4.18.1 ({{{}Provide vmedia username and password if required{}}}).
 
  * However, in this scenario, the firmware’s response differs: the error lacks a structured {{code}} field, causing credential detection logic to fail.
 
  * Engineering suspects a firmware bug, but BMO/Ironic may require additional handling (e.g., checking {{MessageId=ActionParameterMissing}} or the {{error}} field directly) as a workaround.
Status: POST
#OCPBUGS-77207issue5 days ago[openshift-4.18] hermetic build failure - ose-ibm-vpc-block-csi-driver Verified
Issue 17926425: [openshift-4.18] hermetic build failure - ose-ibm-vpc-block-csi-driver
Description: This is a clone of issue OCPBUGS-77206. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77205. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77203. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77167. The following is the description of the original issue:  --- Description of problem:
 {code:none}
 IBM VPC Block CSI Driver build fails in hermetic mode due to non-deterministic GOPATH handling in Makefile. The build system rewrites $GOPATH causing the binary output path -o ${GOPATH}/bin/${EXE_DRIVER_NAME} to become invalid, resulting in build failures.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
 OpenShift 4.19+ (affects hermetic build conversion) {code}
 How reproducible:
 {code:none}
 Always - when building ibm-vpc-block-csi-driver in hermetic mode{code}
 Steps to Reproduce:
 {code:none}
   1. Build ibm-vpc-block-csi-driver container in hermetic mode
   2. Makefile uses absolute GOPATH-dependent output path: -o ${GOPATH}/bin/${EXE_DRIVER_NAME}
   3. Build system rewrites $GOPATH during hermetic build process     {code}
 Actual results:
 {code:none}
 Build fails because binary cannot be found at expected GOPATH location after build system modifies $GOPATH
 
 Error: building at STEP "COPY --from=builder /cachi2/output/deps/gomod/bin/ibm-vpc-block-csi-driver /bin/ibm-vpc-block-csi-driver": checking on sources under [...]: copier: stat: "/cachi2/output/deps/gomod/bin/ibm-vpc-block-csi-driver": no such file or directory{code}
 Expected results:
 {code:none}
 Build succeeds using relative path bin/${EXE_DRIVER_NAME} that doesn't depend on GOPATH manipulation, consistent with other CSI drivers like aws-ebs-csi-driver  {code}
 Additional info:
 {code:none}
 Fix implemented: Change Makefile to use relative path -o bin/${EXE_DRIVER_NAME} instead of absolute GOPATH-dependent path. This aligns with other CSI drivers and makes builds hermetic-compatible. Comment added to indicate this is a permanent carry patch for OpenShift builds.{code}
Status: Verified
#OCPBUGS-77756issue6 days ago[openshift-4.13] hypershift can't be built hermetically POST
Issue 17952763: [openshift-4.13] hypershift can't be built hermetically
Description: Description of problem:
 
 ART can't build hypershift hermetically.
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.    {code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: POST
#OCPBUGS-59320issue2 months agoSome failed bundle unpack pods remain in Error state and are not automatically cleaned up. New
Issue 17145530: Some failed bundle unpack pods remain in Error state and are not automatically cleaned up.
Description: Description of problem:
 {code:none}
     In hypershift hosted cluster, some failed bundle unpack pods remain in Error state and are not automatically cleaned up.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
      Cluster version is 4.20.0-0.nightly-2025-07-15-031201{code}
 How reproducible:
 {code:none}
     always{code}
 Steps to Reproduce:
 {code:none}
     1. create hypershift hosted cluster and create some subs 
 
     2. check pod in openshift-marketplace
 xzha@xzha1-mac 420_upgrade_test % oc get pod -n openshift-marketplace
 NAME                                                              READY   STATUS      RESTARTS   AGE
 bf4ed2748e7e3aa99a4089248dd10559a123af0a4556d4231c44159a9dc2pfc   0/1     Error       0          125m
 bf4ed2748e7e3aa99a4089248dd10559a123af0a4556d4231c44159a9dd5w6b   0/1     Completed   0          124m
 fef64e67f7ed7822d1a21a290838f02d042715e25164247c3fafdbd62ccxc6l   0/1     Completed   0          3m11s
 fef64e67f7ed7822d1a21a290838f02d042715e25164247c3fafdbd62cfwnvz   0/1     Error       0          4m12s
 fef64e67f7ed7822d1a21a290838f02d042715e25164247c3fafdbd62cw84qm   0/1     Error       0          3m46s
 
 there are some failed pods remain in Error state and are not automatically cleaned up.
 
 state:
       terminated:
         containerID: cri-o://f84d39209f18ff35913b26c8775bc04b3c743b3628cc313b44df9a23269affc6
         exitCode: 1
         finishedAt: "2025-07-15T12:01:48Z"
         message: |+
           time="2025-07-15T12:01:38Z" level=info msg="Using in-cluster kube client config"
           Error: error loading manifests from directory: Get "https://172.31.0.1:443/api/v1/namespaces/openshift-marketplace/configmaps/bf4ed2748e7e3aa99a4089248dd10559a123af0a4556d4231c44159a9d3c60a": net/http: TLS handshake timeout
           Usage:{code}
 Actual results:
 {code:none}
    some failed pods remain in Error state and are not automatically cleaned up. {code}
 Expected results:
 {code:none}
     those failed pods will be cleaned automatically{code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-78109issue16 hours ago[openshift-4.13] openshift-enterprise-haproxy-router can't be built hermetically POST
Issue 17972290: [openshift-4.13] openshift-enterprise-haproxy-router can't be built hermetically
Description: This is a clone of issue OCPBUGS-78108. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build openshift-enterprise-haproxy-router & ose-haproxy-router-base hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: POST
#OCPBUGS-72507issue4 weeks ago[JobSet] - It's not possible to create a JobSet using default options Verified
Unable to parse ALM expamples
 SyntaxError: JSON.parse: expected double-quoted property name at line 11 column 5 of the JSON data {code}
I think it was the newlines being included. YAML's | vs |-
#OCPBUGS-77435issue4 days ago"Something wrong happened" n is not iterable showcasing in Openshift web console during NNCP creation New
Issue 17939175: "Something wrong happened" n is not iterable showcasing in Openshift web console during NNCP creation
Description: Description of problem:
 {code:none}
 If we start the NNCP wizard to create a new bridge. As soon as we click next on the policy name, the page "something wrong happened" appears with the message "n is not iterable".
 {code}
 Version-Release number of selected component (if applicable):
 {code:none}
     kubernetes-nmstate-operator.4.21.0-202602180447  
     Openshift version 4.21.2{code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
    Component Trace:
 at hn (https://console-openshift-console.apps.gsphub.gsplab.local/api/plugins/nmstate-console-plugin/exposed-Topology-chunk-e190b1827ce330a09353.min.js:1:52049)
     at div
     at div
     at K (https://console-openshift-console.apps.gsphub.gsplab.local/api/plugins/nmstate-console-plugin/5662-chunk-e3a18368582cf532b0da.min.js:1:10277)
     at div
     at div
     at J (https://console-openshift-console.apps.gsphub.gsplab.local/api/plugins/nmstate-console-plugin/5662-chunk-e3a18368582cf532b0da.min.js:1:11289)
     at ee (https://console-openshift-console.apps.gsphub.gsplab.local/api/plugins/nmstate-console-plugin/5662-chunk-e3a18368582cf532b0da.min.js:1:19753)
     at div
     at L (https://console-openshift-console.apps.gsphub.gsplab.local/api/plugins/nmstate-console-plugin/5662-chunk-e3a18368582cf532b0da.min.js:1:7793)
     at Z (https://console-openshift-console.apps.gsphub.gsplab.local/api/plugins/nmstate-console-plugin/5662-chunk-e3a18368582cf532b0da.min.js:1:17418)
     at bn (https://console-openshift-console.apps.gsphub.gsplab.local/api/plugins/nmstate-console-plugin/exposed-Topology-chunk-e190b1827ce330a09353.min.js:1:54251)
     at wn (https://console-openshift-console.apps.gsphub.gsplab.local/api/plugins/nmstate-console-plugin/exposed-Topology-chunk-e190b1827ce330a09353.min.js:1:55731)
     at Hn (https://console-openshift-console.apps.gsphub.gsplab.local/api/plugins/nmstate-console-plugin/exposed-Topology-chunk-e190b1827ce330a09353.min.js:1:65714)
     at div
     at div
     at l (https://console-openshift-console.apps.gsphub.gsplab.local/static/vendors~main-chunk-37548-c2cda3d4584ab055c403.min.js:52:160317)
     at Xn (https://console-openshift-console.apps.gsphub.gsplab.local/api/plugins/nmstate-console-plugin/exposed-Topology-chunk-e190b1827ce330a09353.min.js:1:66964)
     at div
     at o (https://console-openshift-console.apps.gsphub.gsplab.local/static/vendors~main-chunk-37548-c2cda3d4584ab055c403.min.js:2:3688430)
     at div
     at o (https://console-openshift-console.apps.gsphub.gsplab.local/static/vendors~main-chunk-37548-c2cda3d4584ab055c403.min.js:2:3688039)
     at w (https://console-openshift-console.apps.gsphub.gsplab.local/static/vendors~main-chunk-37548-c2cda3d4584ab055c403.min.js:52:161445)
     at a (https://console-openshift-console.apps.gsphub.gsplab.local/static/vendors~main-chunk-37548-c2cda3d4584ab055c403.min.js:52:163267)
     at Sa (https://console-openshift-console.apps.gsphub.gsplab.local/api/plugins/nmstate-console-plugin/exposed-Topology-chunk-e190b1827ce330a09353.min.js:1:73585)
     at b (https://console-openshift-console.apps.gsphub.gsplab.local/static/main-bundle-1476fa5274b883cc891a.min.js:1:75336)
     at f (https://console-openshift-console.apps.gsphub.gsplab.local/static/main-bundle-1476fa5274b883cc891a.min.js:1:75681)
     at S (https://console-openshift-console.apps.gsphub.gsplab.local/static/main-bundle-1476fa5274b883cc891a.min.js:1:75855)
     at _ (https://console-openshift-console.apps.gsphub.gsplab.local/static/72133-bundle-9dbb1ef6b0b5f1bf2c3e.min.js:2:51105)
     at ne (https://console-openshift-console.apps.gsphub.gsplab.local/static/72133-bundle-9dbb1ef6b0b5f1bf2c3e.min.js:2:56537)
     at Suspense
     at i (https://console-openshift-console.apps.gsphub.gsplab.local/static/main-bundle-1476fa5274b883cc891a.min.js:1:277483)
     at m
     at section
     at v (https://console-openshift-console.apps.gsphub.gsplab.local/static/vendors~main-chunk-37548-c2cda3d4584ab055c403.min.js:2:3460647)
     at W (https://console-openshift-console.apps.gsphub.gsplab.local/static/main-bundle-1476fa5274b883cc891a.min.js:1:933627)
     at main
     at div
     at div
     at div
     at o (https://console-openshift-console.apps.gsphub.gsplab.local/static/vendors~main-chunk-37548-c2cda3d4584ab055c403.min.js:2:3304825)
     at h (https://console-openshift-console.apps.gsphub.gsplab.local/static/vendors~main-chunk-37548-c2cda3d4584ab055c403.min.js:2:3303601)
     at div
     at d (https://console-openshift-console.apps.gsphub.gsplab.local/static/vendors~main-chunk-37548-c2cda3d4584ab055c403.min.js:2:3301897)
     at div
     at div
     at w (https://console-openshift-console.apps.gsphub.gsplab.local/static/vendors~main-chunk-37548-c2cda3d4584ab055c403.min.js:2:3452729)
     at div
     at l (https://console-openshift-console.apps.gsphub.gsplab.local/static/vendors~main-chunk-37548-c2cda3d4584ab055c403.min.js:2:3681733)
     at c (https://console-openshift-console.apps.gsphub.gsplab.local/static/main-bundle-1476fa5274b883cc891a.min.js:1:875511)
     at div
     at o (https://console-openshift-console.apps.gsphub.gsplab.local/static/vendors~main-chunk-37548-c2cda3d4584ab055c403.min.js:2:3304146)
     at div
     at div
     at o (https://console-openshift-console.apps.gsphub.gsplab.local/static/vendors~main-chunk-37548-c2cda3d4584ab055c403.min.js:2:3304825)
     at h (https://console-openshift-console.apps.gsphub.gsplab.local/static/vendors~main-chunk-37548-c2cda3d4584ab055c403.min.js:2:3303601)
     at div
     at d (https://console-openshift-console.apps.gsphub.gsplab.local/static/vendors~main-chunk-37548-c2cda3d4584ab055c403.min.js:2:3301897)
     at Sn (https://console-openshift-console.apps.gsphub.gsplab.local/static/vendors~main-chunk-37548-c2cda3d4584ab055c403.min.js:2:3136530)
     at i (https://console-openshift-console.apps.gsphub.gsplab.local/static/main-bundle-1476fa5274b883cc891a.min.js:1:54467)
     at a (https://console-openshift-console.apps.gsphub.gsplab.local/static/main-bundle-1476fa5274b883cc891a.min.js:1:54036)
     at fe (https://console-openshift-console.apps.gsphub.gsplab.local/static/main-bundle-1476fa5274b883cc891a.min.js:1:946219)
     at tt (https://console-openshift-console.apps.gsphub.gsplab.local/static/vendors~main-chunk-37548-c2cda3d4584ab055c403.min.js:2:3095463)
     at fe (https://console-openshift-console.apps.gsphub.gsplab.local/static/main-bundle-1476fa5274b883cc891a.min.js:1:946219)
     at fe (https://console-openshift-console.apps.gsphub.gsplab.local/static/main-bundle-1476fa5274b883cc891a.min.js:1:946219)
     at t.QueryClientProvider (https://console-openshift-console.apps.gsphub.gsplab.local/api/plugins/mce/8168.24f24aef.js:205:1396530)
     at Zn (https://console-openshift-console.apps.gsphub.gsplab.local/api/plugins/mce/8168.24f24aef.js:205:4728798)
     at RecoilRoot (https://console-openshift-console.apps.gsphub.gsplab.local/api/plugins/mce/8168.24f24aef.js:205:4730771)
     at t.PluginDataContextProvider (https://console-openshift-console.apps.gsphub.gsplab.local/api/plugins/mce/exposed-contextProvider.e725d21f.js:1:10344)
     at fe (https://console-openshift-console.apps.gsphub.gsplab.local/static/main-bundle-1476fa5274b883cc891a.min.js:1:946219)
     at u (https://console-openshift-console.apps.gsphub.gsplab.local/api/plugins/kubevirt-plugin/exposed-modalProvider-chunk-5f1caebfdb41b6e7076c.min.js:1:423)
     at fe (https://console-openshift-console.apps.gsphub.gsplab.local/static/main-bundle-1476fa5274b883cc891a.min.js:1:946219)
     at fe (https://console-openshift-console.apps.gsphub.gsplab.local/static/main-bundle-1476fa5274b883cc891a.min.js:1:946219)
     at Suspense
     at f (https://console-openshift-console.apps.gsphub.gsplab.local/static/main-bundle-1476fa5274b883cc891a.min.js:1:112989)
     at b (https://console-openshift-console.apps.gsphub.gsplab.local/static/main-bundle-1476fa5274b883cc891a.min.js:1:111873)
     at a (https://console-openshift-console.apps.gsphub.gsplab.local/static/main-bundle-1476fa5274b883cc891a.min.js:1:23099)
     at u (https://console-openshift-console.apps.gsphub.gsplab.local/static/main-bundle-1476fa5274b883cc891a.min.js:1:25343)
     at ye (https://console-openshift-console.apps.gsphub.gsplab.local/static/main-bundle-1476fa5274b883cc891a.min.js:1:946335)
     at Se (https://console-openshift-console.apps.gsphub.gsplab.local/static/main-bundle-1476fa5274b883cc891a.min.js:1:949610)
     at _ (https://console-openshift-console.apps.gsphub.gsplab.local/static/72133-bundle-9dbb1ef6b0b5f1bf2c3e.min.js:2:51105)
     at ne (https://console-openshift-console.apps.gsphub.gsplab.local/static/72133-bundle-9dbb1ef6b0b5f1bf2c3e.min.js:2:56537)
     at _ (https://console-openshift-console.apps.gsphub.gsplab.local/static/72133-bundle-9dbb1ef6b0b5f1bf2c3e.min.js:2:51105)
     at ne (https://console-openshift-console.apps.gsphub.gsplab.local/static/72133-bundle-9dbb1ef6b0b5f1bf2c3e.min.js:2:56537)
     at te (https://console-openshift-console.apps.gsphub.gsplab.local/static/72133-bundle-9dbb1ef6b0b5f1bf2c3e.min.js:2:55950)
     at G (https://console-openshift-console.apps.gsphub.gsplab.local/static/72133-bundle-9dbb1ef6b0b5f1bf2c3e.min.js:2:43270)
     at t (https://console-openshift-console.apps.gsphub.gsplab.local/static/72133-bundle-9dbb1ef6b0b5f1bf2c3e.min.js:2:65457)
     at xe (https://console-openshift-console.apps.gsphub.gsplab.local/static/main-bundle-1476fa5274b883cc891a.min.js:1:949902)
     at g (https://console-openshift-console.apps.gsphub.gsplab.local/static/main-bundle-1476fa5274b883cc891a.min.js:1:478028)
     at c (https://console-openshift-console.apps.gsphub.gsplab.local/static/main-bundle-1476fa5274b883cc891a.min.js:1:86080)
     at e (https://console-openshift-console.apps.gsphub.gsplab.local/static/vendors~main-chunk-37548-c2cda3d4584ab055c403.min.js:2:2689533)
     at v (https://console-openshift-console.apps.gsphub.gsplab.local/static/main-bundle-1476fa5274b883cc891a.min.js:1:926590)
     at c (https://console-openshift-console.apps.gsphub.gsplab.local/static/72133-bundle-9dbb1ef6b0b5f1bf2c3e.min.js:2:16683)
     at Suspense
 
 Stack Trace: 
 TypeError: n is not iterable
     at https://console-openshift-console.apps.gsphub.gsplab.local/api/plugins/nmstate-console-plugin/exposed-Topology-chunk-e190b1827ce330a09353.min.js:1:28106
     at Array.reduce (<anonymous>)
     at https://console-openshift-console.apps.gsphub.gsplab.local/api/plugins/nmstate-console-plugin/exposed-Topology-chunk-e190b1827ce330a09353.min.js:1:28011
     at https://console-openshift-console.apps.gsphub.gsplab.local/api/plugins/nmstate-console-plugin/exposed-Topology-chunk-e190b1827ce330a09353.min.js:1:28124
     at https://console-openshift-console.apps.gsphub.gsplab.local/api/plugins/nmstate-console-plugin/exposed-Topology-chunk-e190b1827ce330a09353.min.js:1:28265
     at Array.map (<anonymous>)
     at https://console-openshift-console.apps.gsphub.gsplab.local/api/plugins/nmstate-console-plugin/exposed-Topology-chunk-e190b1827ce330a09353.min.js:1:27974
     at Pe (https://console-openshift-console.apps.gsphub.gsplab.local/api/plugins/nmstate-console-plugin/exposed-Topology-chunk-e190b1827ce330a09353.min.js:1:28311)
     at hn (https://console-openshift-console.apps.gsphub.gsplab.local/api/plugins/nmstate-console-plugin/exposed-Topology-chunk-e190b1827ce330a09353.min.js:1:52087)
     at sa (https://console-openshift-console.apps.gsphub.gsplab.local/static/vendors~main-chunk-37548-c2cda3d4584ab055c403.min.js:2:2607126)
 {code}
Status: New
#OCPBUGS-78179issue9 hours ago[openshift-4.14] ose-azure-cluster-api-controllers can't be built hermetically New
Issue 17974788: [openshift-4.13] ose-azure-cluster-api-controllers can't be built hermetically
Description: This is a clone of issue OCPBUGS-78178. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78177. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78176. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77983. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77905. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77904. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-azure-cluster-api-controllers
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-62000issue5 days agoSecurity requirements do not allow UserInfo mapping resulting in unsupported config of auth ASSIGNED
Issue 17408594: Security requirements do not allow UserInfo mapping resulting in unsupported config of auth
Description: Chase Morgan has a security requirement to disable UserInfo in their ADFS server.
 curls to the ADFS to obtain a token will fail with
 ~~~
 Error: non-200 response from UserInfo: 401, WWW-Authenticate=Bearer error="invalid_token", error_description="MSIS9921: Received invalid UserInfo request. Audience 'https://oauth-openshift.apps.wh-ngcptnt1.svr.us.jpmchase.net/oauth2callback/adfs' in the access token is not same as the identifier of the UserInfo relying party trust 'urn:microsoft:userinfo'."r
 ~~~
 
 That is with the below authconfig
 ~~~
 urls:
             authorize: https://mysso.example.com/auth/realms/master/protocol/openid-connect/auth
             token: https://mysso.example.com/auth/realms/master/protocol/openid-connect/token
             userInfo: https://mysso.example.com/auth/realms/master/protocol/openid-connect/userinfo
 ~~~
 
 Removing the userinfo from the authconfig will result in the curls working but results in an unsupported state.
 https://access.redhat.com/solutions/6960548
 https://access.redhat.com/articles/7064122
 
 They have implemented the fixes from Microsoft that was mentioned in https://issues.redhat.com/browse/RFE-3596 and https://issues.redhat.com/browse/RFE-2770
 
 How can JPMC implement the authconfig to not use userinfo with adfs and not end up unsupported?
Status: ASSIGNED
#OCPBUGS-65897issue6 weeks ago[assisted-service] go mod tidy fix ASSIGNED
Issue 17601884: [assisted-service] go mod tidy fix
Description: ART builds failing with incorrect go mod dependencies. Prefetch task failing on konflux: [https://konflux-ui.apps.kflux-ocp-p01.7ayg.p1.openshiftapps.com/ns/ocp-art-tenant/applications/openshift-4-16/pipelineruns/ose-4-16-ose-agent-installer-api-server-zwhp6/logs] (ART internal)
 
  
 {code:java}
   go: downloading golang.org/x/mod v0.17.0
   go: updates to go.mod needed; to update it:
   	go mod tidy
   2025-11-22 04:18:10,928 ERROR Failed to fetch gomod dependencies
   2025-11-22 04:18:14,114 ERROR PackageManagerError: Go execution failed: `/usr/local/bin/go mod vendor` failed with rc=1
   Error: PackageManagerError: Go execution failed: `/usr/local/bin/go mod vendor` failed with rc=1
     The cause of the failure could be:
     - something is broken in hermeto
     - something is wrong with your repository
     - communication with an external service failed (please try again)
     The output of the failing command should provide more details, please check the logs. {code}
  
 
 Builds will be paused until issue is resolved. Please feel free to close the bug as soon as the PR is merged
Status: ASSIGNED
#OCPBUGS-77206issue8 days ago[openshift-4.19] hermetic build failure - ose-ibm-vpc-block-csi-driver Verified
Issue 17926170: [openshift-4.19] hermetic build failure - ose-ibm-vpc-block-csi-driver
Description: This is a clone of issue OCPBUGS-77205. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77203. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77167. The following is the description of the original issue:  --- Description of problem:
 {code:none}
 IBM VPC Block CSI Driver build fails in hermetic mode due to non-deterministic GOPATH handling in Makefile. The build system rewrites $GOPATH causing the binary output path -o ${GOPATH}/bin/${EXE_DRIVER_NAME} to become invalid, resulting in build failures.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
 OpenShift 4.19+ (affects hermetic build conversion) {code}
 How reproducible:
 {code:none}
 Always - when building ibm-vpc-block-csi-driver in hermetic mode{code}
 Steps to Reproduce:
 {code:none}
   1. Build ibm-vpc-block-csi-driver container in hermetic mode
   2. Makefile uses absolute GOPATH-dependent output path: -o ${GOPATH}/bin/${EXE_DRIVER_NAME}
   3. Build system rewrites $GOPATH during hermetic build process     {code}
 Actual results:
 {code:none}
 Build fails because binary cannot be found at expected GOPATH location after build system modifies $GOPATH
 
 Error: building at STEP "COPY --from=builder /cachi2/output/deps/gomod/bin/ibm-vpc-block-csi-driver /bin/ibm-vpc-block-csi-driver": checking on sources under [...]: copier: stat: "/cachi2/output/deps/gomod/bin/ibm-vpc-block-csi-driver": no such file or directory{code}
 Expected results:
 {code:none}
 Build succeeds using relative path bin/${EXE_DRIVER_NAME} that doesn't depend on GOPATH manipulation, consistent with other CSI drivers like aws-ebs-csi-driver  {code}
 Additional info:
 {code:none}
 Fix implemented: Change Makefile to use relative path -o bin/${EXE_DRIVER_NAME} instead of absolute GOPATH-dependent path. This aligns with other CSI drivers and makes builds hermetic-compatible. Comment added to indicate this is a permanent carry patch for OpenShift builds.{code}
Status: Verified
#OCPBUGS-77468issue11 days ago[openshift-4.13] openshift-enterprise-hyperkube can't be built hermetically New
Issue 17940936: [openshift-4.13] openshift-enterprise-hyperkube can't be built hermetically
Description: Description of problem:
 ART can't build openshift-enterprise-hyperkube hermetically. Same applies for ose-installer-kube-apiserver-artifacts
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.
      {code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-66250issue2 months agoGCP: Install failures due to inconsistent iam apply Verified
Issue 17631044: GCP: Install failures due to inconsistent iam apply
Description: *Description of problem*. In Advanced Cluster Security we rely for OCP creation in our CI and recently observed [an increase of cluster creation failures|https://docs.google.com/document/d/1vE-kLPWooO75BXxGub6YZpM4NZolKBcOrvJULZ5NBw0/edit]. While we've been [advised to retry the failures|https://redhat-internal.slack.com/archives/C68TNFWA2/p1722873956362609] (and we do so now, see [ROX-25416|https://issues.redhat.com/browse/ROX-25416]), I'm afraid our use case is not so unique and others are affected as well.
 
 We suggest upgrading terraform and provider to the latest version (possible before license changes) in openshift-installer for 4.12+. The underlying issue is [probably already fixed upstream|https://github.com/hashicorp/terraform-provider-google/issues/17332] and released in [v5.37.0|https://github.com/hashicorp/terraform-provider-google/releases/tag/v5.37.0].
 
 Version-Release number of selected component (if applicable): TBD
 
 How reproducible: TBD
 
 Steps to Reproduce: TBD
 
 Actual results: TBD
 
 Expected results: TBD
 
 h3.  Additional info.
 
 The most common error we see in our [JIRA issues|https://issues.redhat.com/browse/ROX-25724?jql=text%20~%20%22module.master.google_service_account.master-node-sa%22] is and that is something we could find similar issues with AWS provider too eg. OCPBUGS-4213.
 {code}
 level=error msg=Error: Provider produced inconsistent result after apply .... resource was present, but now absent
 {code}
 
 Summary of errors from:
 {code}
       3 failed to create cluster: failed to apply Terraform: error(GCPComputeBackendTimeout) from Infrastructure Provider: GCP is experiencing backend service interuptions, the compute instance failed to create in reasonable time.&#34;
       3 Provider produced inconsistent result after apply\n\nWhen applying changes to\nmodule.master.google_service_account.master-node-sa[0], provider\n\&#34;provider[\\\&#34;openshift/local/google\\\&#34;]\&#34; produced an unexpected new value: Root\nresource was present, but now absent.\n\n
       6 Error waiting to create Network: Error waiting for Creating Network: timeout while waiting for state to become &#39;DONE&#39; (last state: &#39;RUNNING&#39;, timeout: 4m0s)\n\n  with module.network.google_compute_network.cluster_network[0],\n  on network/network.tf line 1, in resource \&#34;google_compute_network\&#34; \&#34;cluster_network\&#34;:\n   1: resource \&#34;google_compute_network\&#34; \&#34;cluster_network\&#34; {\n\n&#34;
       9 error applying Terraform configs: failed to apply Terraform: error(GCPComputeBackendTimeout) from Infrastructure Provider: GCP is experiencing backend service interuptions, the compute instance failed to create in reasonable time.&#34;
      14 Provider produced inconsistent result after apply\n\nWhen applying changes to module.master.google_service_account.master-node-sa,\nprovider \&#34;provider[\\\&#34;openshift/local/google\\\&#34;]\&#34; produced an unexpected new\nvalue: Root resource was present, but now absent.
      16 Provider produced inconsistent result after apply\n\nWhen applying changes to google_service_account_key.bootstrap, provider\n\&#34;provider[\\\&#34;openshift/local/google\\\&#34;]\&#34; produced an unexpected new value: Root\nresource was present, but now absent.
      18 Provider produced inconsistent result after apply\n\nWhen applying changes to module.iam.google_service_account.worker-node-sa,\nprovider \&#34;provider[\\\&#34;openshift/local/google\\\&#34;]\&#34; produced an unexpected new\nvalue: Root resource was present, but now absent.
      34 Error creating service account key: googleapi: Error 404: Service account projects/acs-san-stackroxci/serviceAccounts/XXX@acs-san-stackroxci.iam.gserviceaccount.com does not exist., notFound\n\n  with google_service_account_key.bootstrap,\n  on main.tf line 38, in resource \&#34;google_service_account_key\&#34; \&#34;bootstrap\&#34;:\n  38: resource \&#34;google_service_account_key\&#34; \&#34;bootstrap\&#34; {\n\n&#34;
      45 error applying Terraform configs: failed to apply Terraform: exit status 1\n\nError: Provider produced inconsistent result after apply\n\nWhen applying changes to\nmodule.master.google_service_account.master-node-sa[0], provider\n\&#34;provider[\\\&#34;openshift/local/google\\\&#34;]\&#34; produced an unexpected new value: Root\nresource was present, but now absent.
      59 error applying Terraform configs: failed to apply Terraform: exit status 1\n\nError: Provider produced inconsistent result after apply\n\nWhen applying changes to module.iam.google_service_account.worker-node-sa,\nprovider \&#34;provider[\\\&#34;openshift/local/google\\\&#34;]\&#34; produced an unexpected new\nvalue: Root resource was present, but now absent.
     100 Provider produced inconsistent result after apply\n\nWhen applying changes to google_service_account.bootstrap-node-sa, provider\n\&#34;provider[\\\&#34;openshift/local/google\\\&#34;]\&#34; produced an unexpected new value: Root\nresource was present, but now absent.
     103 Provider produced inconsistent result after apply\n\nWhen applying changes to module.iam.google_service_account.worker-node-sa[0],\nprovider \&#34;provider[\\\&#34;openshift/local/google\\\&#34;]\&#34; produced an unexpected new\nvalue: Root resource was present, but now absent.
     116 Provider produced inconsistent result after apply\n\nWhen applying changes to\nmodule.master.google_service_account.master-node-sa[0], provider\n\&#34;provider[\\\&#34;openshift/local/google\\\&#34;]\&#34; produced an unexpected new value: Root\nresource was present, but now absent.
 {code}
 
 The openshift installer contains a bundled terraform and google-provider
 * 4.12 
 ** Terraform 1.0.11 (November 10, 2021)
 ** https://github.com/openshift/installer/blob/release-4.12/terraform/terraform/go.mod#L5
 ** Provider: v4.5.0 (Dec 20, 2021)
 ** https://github.com/openshift/installer/blob/release-4.12/terraform/providers/google/go.mod#L5
 * 4.14+
 ** Terraform: 1.3.7 (January 04, 2023)
 ** https://github.com/openshift/installer/blob/release-4.14/terraform/terraform/go.mod#L5
 ** Provider: v4.74.0 (Jul 18, 2023)
 ** https://github.com/openshift/installer/blob/release-4.14/terraform/providers/google/go.mod#L5
 * 4.18+
 ** Does not contain google provider at all: https://github.com/openshift/installer/pull/8723
Status: Verified
#OCPBUGS-78023issue34 hours ago[openshift-4.14] ose-vsphere-cluster-api-controllers can't be built hermetically New
Issue 17969432: [openshift-4.14] ose-vsphere-cluster-api-controllers can't be built hermetically
Description: This is a clone of issue OCPBUGS-78021. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78020. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78019. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78018. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78017. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-vsphere-cluster-api-controllers hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-76586issue12 hours agoGCP: Provider produced inconsistent result after apply - google_service_account Verified
Issue 17866072: GCP: Provider produced inconsistent result after apply - google_service_account
Description: Description of problem:
 {code:none}
     The fix given for issue https://issues.redhat.com/browse/OCPBUGS-42014(pointing to same problem) was at service account key creation. but the actual problem lies at service account creation itself.  
 
 We suggest upgrading terraform and provider to the latest version in openshift-installer for 4.12+. The underlying issue is already fixed upstream and released in v6.45.0.
 
 Note: we see this problem more often in lower openshift versions 4.12, 4.13 than in 4.16.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
 Intermittent{code}
 Steps to Reproduce:
 {code:none}
     1.Install the UPI openshift cluster in GCP platform
     2.
     3.
     {code}
 Actual results:
 {code:none}
     level=error msg=Error: Provider produced inconsistent result after apply
 level=error
 level=error msg=When applying changes to
 level=error msg=module.master.google_service_account.master-node-sa[0], provider
 level=error msg="provider[\"openshift/local/google\"]" produced an unexpected new value: Root
 level=error msg=resource was present, but now absent.
 level=error
 level=error msg=This is a bug in the provider, which should be reported in the provider's own{code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: Verified
#OCPBUGS-76299issue56 minutes agoSNO ZTP deployment fails with Dell GNR-D iDRAC Firmware Version: 1.30.10.50 ASSIGNED
Issue 17846211: SNO ZTP deployment fails with Dell GNR-D  iDRAC Firmware Version: 1.30.10.50
Description: Description of problem:
 ZTP deployment failure
 Server: Dell GRNRD 
 iDRAC IP: 192.168.224.190
 iDRAC Firmware Version: 1.30.10.50
 OpenShift Version: 4.21.0
 Deployment Method: ZTP via ACM/Assisted Installer
 
 
     Version-Release number of selected component (if applicable):
 4.21
       How reproducible:
   Every time SNO deploy using ZTP 
 
 
     Steps to Reproduce:{
     1. Dell gnrd Perla with latest IDRAC version : 1.30.10.50
     2. using ZTP deploy SNO 
     3.The Malformed Redfish Attribute Error
     
 Node 8c798425-3ae4-463d-a4cd-539c1611924d failed verify step clear_job_queue with unexpected error: 
 The attribute MemorySummary/TotalSystemMemoryGiB is malformed in the resource /redfish/v1/Systems/System.Embedded.1: 
 invalid literal for int() with base 10: '256 GB'
     Actual results:
 Root Cause: iDRAC firmware returns "256 GB" (string with unit) instead of 256 (integer) for TotalSystemMemoryGiB
     
     Expected results:
 virtual media mount successful  iso 
     
     Additional info:
 
 Error Message (with idrac-redfish-virtualmedia://):
 The Value is not valid BMC address, supported protocols are redfish-virtualmedia or idrac-virtualmedia.
 
 
 When Ironic's idrac driver parses this response, it attempts to convert TotalSystemMemoryGiB to an integer:
 
 # Ironic code
 memory_gib = int(system_data['MemorySummary']['TotalSystemMemoryGiB'])
 # Raises: ValueError: invalid literal for int() with base 10: '256 GB'
 # 
 
 
  curl -k -s -u user:passowrd \
 >   https://192.168.224.190/redfish/v1/Systems/System.Embedded.1 | jq '.MemorySummary'
 {
   "Status": {
     "Health": "OK",
     "HealthRollup": "OK",
     "State": "Enabled"
   },
   "Status@Redfish.Deprecated": "Please migrate to use Status in the individual Memory resources",
   "TotalSystemMemoryGiB": 256,
   "MemoryMirroring": "System"
 }
 
 
 
 Impact on Ironic/Metal3
 
 
 # Ironic sushy library code (simplified)
 memory_gib = int(system_data['MemorySummary']['TotalSystemMemoryGiB'])
 # Raises: ValueError: invalid literal for int() with base 10: '256 GB'
Status: ASSIGNED
#OCPBUGS-73920issue6 days agoRedfish vMedia mount failure on Cisco B-Series blades during ACM provisioning New
Issue 17773649: Redfish vMedia mount failure on Cisco B-Series blades during ACM provisioning
Description: Description of problem: {code:none} Redfish Virtual Media (vMedia) mounting fails consistently on Cisco B-Series (Blade) servers when provisioned via ACM / SiteConfig Operator, whereas Cisco C-Series (Rack) servers in the same environment work successfully.
 
 When ACM attempts to map the discovery ISO, the CIMC reports a "Mount failed" error and the resource state moves to "UnavailableOffline." Logs indicate that the Redfish 'PATCH' method is not supported for this resource, and the subsequent 'InsertMedia' POST action returns a 400 Bad Request. Internal discussions suggest a potential communication gap between the CIMC and the ACM Hub through the Fabric Interconnects, but this has not been confirmed via packet capture. {code}
 
 Version-Release number of selected component (if applicable): {code:none} ACM: 2.x (via SiteConfig Operator) OCP: 4.20 (Target) Hardware: Cisco B200 M5 Firmware: 4.3(6c) {code}
 
 How reproducible: {code:none} 100% reproducible on Cisco B-Series blades managed by Fabric Interconnects. {code}
 
 Steps to Reproduce: {code:none}
 
 Define a Bare Metal cluster using ACM SiteConfig/ClusterInstance.
 
 Configure bmcAddress using 'redfish-virtualmedia+https' pointing to a Cisco B-Series CIMC.
 
 Apply the configuration and monitor the node provisioning process.
 
 Check Virtual Media status via Redfish API: /redfish/v1/Managers/CIMC/VirtualMedia/3 {code}
 
 Actual results: {code:none} CIMC reports a critical health status with "Mount failed." Sushy/Ironic logs show: "HTTP PATCH... status code: 400, error: Base.1.13.0.GeneralError: The action 'Patch Update For Resource' is not supported by the resource." followed by a failed POST to InsertMedia. {code}
 
 Expected results: {code:none} The discovery ISO should mount successfully as a Virtual DVD, allowing the blade to boot into the discovery environment. {code}
 
 Additional info: {code:none}
 
 Manual mounting via the CIMC Web UI works, indicating the ISO itself is valid.
 
 The environment uses Fabric Interconnects for B-Series management traffic, which is the primary architectural difference from the successful C-Series deployments.
 
 {code}
Status: New
#OCPBUGS-74983issue37 hours agoRedfish PATCH failure: BootSourceOverrideMode unsupported on Cisco CIMC hardware blocks discovery ASSIGNED
Issue 17829950: Redfish PATCH failure: BootSourceOverrideMode unsupported on Cisco CIMC hardware blocks discovery
Description: Description of problem: {code:none} During baremetal host provisioning via ACM 2.14 / Metal3 on Cisco hardware, the ironic-conductor (sushy) fails to set the boot mode. The Redfish API returns a 400 Bad Request because the property "BootSourceOverrideMode" is either read-only or unsupported by the Cisco CIMC Redfish implementation.
 
 Even after manually setting the hardware to UEFI and Secure Boot, Ironic attempts to patch the "BootSourceOverrideMode" to "UEFI", resulting in a fatal error that prevents the host from booting the discovery ISO.
 
 Log snippet: 2026-02-03 07:08:13.542 1 ERROR ironic.drivers.modules.redfish.management [None req-f5103489-5928-4fce-9975-360e1dede01d] Setting boot mode to uefi failed for node bec24335-d444-4460-9d62-652897ca16c9. Error: HTTP PATCH https://<BMC_IP>/redfish/v1/Systems/WZP29020ZJN returned code 400. Base.1.13.0.GeneralError: Following properties in the request body cannot be updated because they are either read only, unknown or unsupported : "BootSourceOverrideMode" {code}
 
 Version-Release number of selected component (if applicable): {code:none} OCP Hub: 4.18.14 ACM: 2.14 Hardware: Cisco CIMC {code}
 
 How reproducible: {code:none} Always reproducible on this specific Cisco hardware/firmware combination when using the Redfish driver. {code}
 
 Steps to Reproduce: {code:none}
 
 Define a BareMetalHost (BMH) with bootMode: UEFI.
 
 Attempt to provision the host using ZTP/ACM 2.14.
 
 Observe the metal3-ironic pod logs during the management/provisioning phase. {code}
 
 Actual results: {code:none} Ironic attempts to send a PATCH request with {'Boot': {'BootSourceOverrideMode': 'UEFI'}}. The BMC rejects this with a 400 GeneralError, causing Ironic to fail the operation and stop the provisioning flow. {code}
 
 Expected results: {code:none} Ironic/Sushy should verify if "BootSourceOverrideMode" is a supported/writable property before attempting a PATCH. If the host is already in the desired boot mode (manually configured), the failure to PATCH should not be fatal, or the driver should use an alternative method compatible with Cisco Redfish. {code}
 
 Additional info: {code:none}
 
 Manual verification via curl shows that UEFI is enabled on the host, but the "BootSourceOverrideMode" parameter is missing from the Redfish tree or marked as non-updatable.
 
 Must-gather attached to the support case. {code}
Status: ASSIGNED
#OCPBUGS-45804issue10 hours agoInstaller fails GetMarketplaceImage with context deadline exceeds New
Issue 16509003: Installer fails GetMarketplaceImage with context deadline exceeds
Description: Description of problem:
 
 In openshift/installer, when running GetMarketplaceImage, we get 
 {code:none}
 encountered error: failed to fetch marketplace image: could not get marketplace image: compute.VirtualMachineImagesClient#Get: Failure sending request: StatusCode=0 – Original Error: context deadline exceeded{code}
 This error comes from [https://github.com/openshift/installer/blob/2b773d8339292cb3f2ba294e3b8f186da9cf7395/pkg/asset/installconfig/azure/client.go#L367-L378]
 
 However, it doesn't have to timeout in 30 secs here, and considering the time required to the whole install, we should retry the request when it fails decent number of times.
 
 Also it's still using the deprecated SDK, which lacks retry mechanism.
 
 With the new SDK, there's a rich retry mechanism and that should retry when it gets certain errors.
 
 [https://learn.microsoft.com/en-us/azure/storage/blobs/storage-retry-policy-go]
 
 [https://azure.github.io/azure-sdk/releases/deprecated/go.html]
 
 [https://github.com/Azure/azure-sdk-for-go/blob/69f43460d64e153c91e6fe52113ad31ce6f91a06/sdk/azcore/runtime/policy_retry.go#L106]
 
  
 
 Version-Release number of selected component (if applicable):
 {code:none}
 all versions implementing GetMarketplaceImage{code}
 How reproducible:
 {code:none}
 when there's a network issue.{code}
 Steps to Reproduce:
 
 It relates to the network issue. not reproducible always.
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
 It terminates immediately.{code}
 Expected results:
 {code:none}
 It should retry some times.{code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-77912issue36 hours agoOpenShift Lightspeed "Import to console" action fails on 4.20+ ASSIGNED
Issue 17963284: OpenShift Lightspeed "Import to console" action fails on 4.20+
Description: Description of problem:
 {code:none}
 When clicking the "Import to console" action on an OpenShift Lightspeed YAML code block (generated as part of the AI response), it correctly redirects to the YAML import page, but the editor is not pre-populated with the YAML {code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
 Always{code}
 Steps to Reproduce:
 {code:none}
 1. Install OpenShift Lightspeed operator
 2. Prompt the AI with something that will cause it to respond with some YAML such as "Show me some example deployment YAML"
 3. Click the "Import to console" action on the resulting YAML code block{code}
 Actual results:
 {code:none}
 YAML not injected into editor and JavaScript error thrown{code}
 Expected results:
 {code:none}
 YAML editor should be pre-populated with the YAML code {code}
 Additional info:
 {code:none}
 Error message is: "TypeError: Cannot use 'in' operator to search for 'editor' in undefined"{code}
Status: POST
#OCPBUGS-66294issue2 months agoLCA seedgen container fails when TechPreviewNoUpgrade is enabled on seed cluster New
Issue 17633189: LCA seedgen container fails when TechPreviewNoUpgrade is enabled on seed cluster
Description: Description of problem:{code:none}
 In preperation to sanity check LCA operator deployed via OLMv1 (which currently requires the TechPreviewNoUpgrade feature gate enabled). 
 First was to enable the feature gate, but install LCA the traditional way first in order to find if the feature gate affects LCA seedgen.
 
 The result is seed gen fails due to container error on spoke cluster
 
 [core@helix80 ~]$ sudo podman logs lca_image_builder | grep -C5 -i error
 time="2025-12-02 22:47:08" level=info msg="Unauthenticated etcd server for recert is up and running"
 time="2025-12-02 22:47:08" level=info msg="Start running recert"
 time="2025-12-02 22:47:08" level=info msg="Executing nsenter with args [--target 1 --cgroup --mount --ipc --pid -- podman run --rm --network=host --privileged --replace --name recert -v /etc:/host-etc -v /etc/ssh:/ssh -v /etc/kubernetes:/kubernetes -v /var/lib/kubelet:/kubelet -v /var/tmp:/var/tmp -v /etc/machine-config-daemon:/machine-config-daemon -v /etc/pki:/pki -e RECERT_CONFIG=/var/tmp/backup/recert_config.json --authfile /var/lib/lca/ibu-seedgen-orch/auth.json registry.redhat.io/openshift4/recert-rhel9@sha256:73c87fcc1a20bfec8aef234bb2d12bd69ba24072256548f003f245e2f560618f]"
 time="2025-12-02 22:47:10" level=info msg="Killing the unauthenticated etcd server"
 time="2025-12-02 22:47:10" level=info msg="Executing nsenter with args [--target 1 --cgroup --mount --ipc --pid -- podman stop recert_etcd]"
 time="2025-12-02 22:47:10" level=error msg="failed to create seed image: failed to run once recert: failed to run recert tool container: 2025-12-02 22:47:09 - INFO - src/cluster_crypto/crypto_utils.rs:272: using openssl: OpenSSL 3.5.1 1 Jul 2025 (Library: OpenSSL 3.5.1 1 Jul 2025)\n\n2025-12-02 22:47:09 - INFO - src/cluster_crypto/crypto_utils.rs:300: OpenSSL version is compatible\nError: etcd to utf-8\n\nCaused by:\n    invalid utf-8 sequence of 1 bytes from index 0\n: exit status 1"
 time="2025-12-02 22:47:10" level=info msg="Cleaning up seed cluster"
 time="2025-12-02 22:47:10" level=info msg="Removing seed image"
 time="2025-12-02 22:47:10" level=info msg="Executing nsenter with args [--target 1 --cgroup --mount --ipc --pid -- podman rmi registry.kni-qe-84.telcoqe.eng.rdu2.dc.redhat.com:5000/ibu/seed:4.21.0-0.nightly-2025-11-22-193140]"
 time="2025-12-02 22:47:10" level=error msg="failed to remove seed image: failed to run \"podman\" in host namespace with args [rmi registry.kni-qe-84.telcoqe.eng.rdu2.dc.redhat.com:5000/ibu/seed:4.21.0-0.nightly-2025-11-22-193140]: Error: registry.kni-qe-84.telcoqe.eng.rdu2.dc.redhat.com:5000/ibu/seed:4.21.0-0.nightly-2025-11-22-193140: image not known\n: exit status 1"
 time="2025-12-02 22:47:10" level=info msg="Cleaning up systemd service units"
 time="2025-12-02 22:47:10" level=info msg="Disabling service unit installation-configuration.service"
 time="2025-12-02 22:47:10" level=info msg="Running systemctl disable [installation-configuration.service --now]"
 time="2025-12-02 22:47:10" level=info msg="Executing nsenter with args [--target 1 --cgroup --mount --ipc --pid -- systemctl disable installation-configuration.service --now]"
 time="2025-12-02 22:47:11" level=info msg="Removing service unit installation-configuration.service"
 --
 time="2025-12-02 22:47:11" level=info msg="Removing /etc/cni/multus/certs folder"
 time="2025-12-02 22:47:11" level=info msg="Removing /var/seed_data folder"
 time="2025-12-02 22:47:11" level=info msg="Restoring cluster services (i.e. kubelet.service unit)"
 time="2025-12-02 22:47:11" level=info msg="Running systemctl enable [kubelet.service --now]"
 time="2025-12-02 22:47:11" level=info msg="Executing nsenter with args [--target 1 --cgroup --mount --ipc --pid -- systemctl enable kubelet.service --now]"
 time="2025-12-02 22:47:13" level=fatal msg="Failed to restore seed cluster: encountered 1 errors during restoration"
 
 
     {code}
     Version-Release number of selected component (if applicable):{code:none}
 OCP 4.21.0-0.nightly-2025-11-22-193140
 LCA 
     {code}
     How reproducible:{code:none}
 Always
     {code}
     Steps to Reproduce:{code:none}
     1. Deploy SNO OCP 4.21 spoke for seed gen via ZTP applying RAN DU profile from RDS with the feature gate enabled with:
 
   installConfigOverrides: |
     {
       "featureSet": "TechPreviewNoUpgrade"
     }
 
     2. Run seed gen job after seed deployment and monitor progress of the seedgen CR
     3.
     {code}
     Actual results:{code:none}
 seedgen fails due to container exit 1 status
     {code}
     Expected results:{code:none}
 seedgen completes successfully.
     {code}
     Additional info:{code:none}
 This issue does not happen if the TPNU feature gate is not enabled with deployment config the same otherwise.
     {code}
Status: New
#OCPBUGS-78010issue35 hours ago[openshift-4.13] sriov-network-* images can't be built hermetically New
Issue 17969096: [openshift-4.12] sriov-network-* images can't be built hermetically
Description: This is a clone of issue OCPBUGS-78009. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78007. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build sriov-network-* images hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-74938issue30 hours agoKubelet and NetworkManager do not start automatically on any node after reboot, leaving nodes stuck in NotReady CLOSED
{code:java}
Jan 30 09:12:11 ocp-w1.ocp-prd-lv.afcucorp.local bash[3148]: Error: initializing source docker://quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2c3ae2f63061eb25dab528c0b904994ed08cbb76debc19fcb4604c2e57e6e785: pinging container registry quay.io: Get "https://quay.io/v2/": dial tcp: lookup quay.io on [::1]:53: read udp [::1]:48765->[::1]:53: read: connection refused{code}
From the logs, the runtime is failing with a “connection refused” error while trying to resolve quay.io using DNS at [::1]:53 (IPv6 loopback).
#OCPBUGS-70326issue8 days agoInspection failure due to unreachable image shows another confusing error message ASSIGNED
Issue 17727752: Inspection failure due to unreachable image shows another confusing error message
Description: Description of problem:
 
 
 
 While inspecting a Dell machine (PowerEdge R640, iDRAC9) from a dev-scripts based BMO setup, inspection fails with this error message: 
 {code:java}
 2026-01-02 14:24:24.256 1 ERROR ironic.conductor.task_manager [None req-1b5a3a1b-c989-4e24-9462-3a3f531eba68 - - - - - -] Node b26bc87d-cc12-46db-85b4-98d24b16de03 moved to provision state "inspect failed" from state "inspecting"; target provision state is "manageable": ironic.common.exception.HardwareInspectionFailure: Failed to inspect hardware. Reason: unable to start inspection: HTTP POST https://IP:443/redfish/v1/Systems/System.Embedded.1/VirtualMedia/1/Actions/VirtualMedia.InsertMedia returned code 500. Base.1.12.GeneralError: A general error has occurred. See ExtendedInfo for more information Extended information: [{'Message': 'Virtual Media is detached or Virtual Media devices are already in use.', 'MessageArgs': [], 'MessageArgs@odata.count': 0, 'MessageId': 'IDRAC.2.8.VRM0021', 'RelatedProperties': [], 'RelatedProperties@odata.count': 0, 'Resolution': 'Change the Virtual Media Attach Mode to Attach or Auto-Attach.Stop existing Virtual Media sessions and retry the operation.', 'Severity': 'Informational'}]             {code}
 There are no virtual media devices attached. the HTTP POST request to insert virtual media is as follows:
 {code:java}
 2026-01-02 14:24:21.939 1 DEBUG sushy.connector [None req-1b5a3a1b-c989-4e24-9462-3a3f531eba68 - - - - - -] HTTP request: POST https://<IP>:443/redfish/v1/Systems/System.Embedded.1/VirtualMedia/1/Actions/VirtualMedia.InsertMedia; headers: {'Content-Type': 'application/json', 'OData-Version': '4.0', 'Accept-Encoding': 'identity'}; body: {'Image': 'https://192.168.111.22:6183/redfish/boot-b26bc87d-cc12-46db-85b4 98d24b16de03.iso'}; blocking: False; timeout: 60; session arguments: {}; _op /usr/lib/python3.12/site-packages/sushy/connector.py:167  {code}
 The image URL in the request body needs to be fixed, and the error message from ironic is misleading.
 
 iDRAC firmware version was upgraded from 6.10.80.00 to 7.00.00.183, and the BIOS version was upgraded from 2.18.1 to 2.25.0, and the same issue was observed with both versions of both.
 
 
 
 Version-Release number of selected component (if applicable):
 {code:none}
     4.20{code}
 How reproducible:
 {code:none}
     Always{code}
 Steps to Reproduce:
 {code:none}
     1. Inspect a (Dell?) BMH using a dev-scripts-based BMO and ironic
     2.
     3.{code}
 Actual results:
 {code:none}
     Misleading error message{code}
 Expected results:
 {code:none}
     Better error message about image being unreachable{code}
 Additional info:
 {code:none}
     {code}
Status: ASSIGNED
#OCPBUGS-76644issue3 weeks agoABI OVE: Initial virtual device mount fails New
Issue 17872088: ABI OVE: Initial virtual device mount fails
Description: Description of problem:
 {code:none}
 There is an issue during the initial virtual mounting of the ISO. The manifest file inside the virtual mounting folder appears to be corrupt, which causes the local registry service to throw an error while attempting to pull images. Instead of displaying the appropriate failure message, the system proceeds with the next steps.    {code}
 Version-Release number of selected component (if applicable):
 {code:none}
 4.22{code}
 How reproducible:
 {code:none}
 Intermittent {code}
 Steps to Reproduce:
 {code:none}
     1. Boot the OVE ISO
     2. Observe the start-local-registry.service
     {code}
 Actual results:
 {code:none}
 Feb 12 10:27:42 localhost systemd[1]: Starting Local Registry (OCP)...
 Feb 12 10:27:51 localhost.localdomain podman[2252]: 2026-02-12 10:27:51.328585781 +0000 UTC m=+3.114220808 system refresh
 Feb 12 10:27:51 localhost.localdomain setup-local-registry.sh[2252]: Error: read /mnt/agentdata/images/registry/manifest.json: input/output error
 Feb 12 10:27:51 localhost.localdomain podman[2252]: 2026-02-12 10:27:51.353089288 +0000 UTC m=+3.138724318 image pull-error  dir:/mnt/agentdata/images/registry read /mnt/agentdata/images/registry/manifest.json: input/output error
 Feb 12 10:27:51 localhost.localdomain setup-local-registry.sh[2104]: Warning: Could not find image with digest sha256:0e889a92bb1795b586c5fb2531e7f64d8e33f98b3720b9408ece1feee1135ef3
 Feb 12 10:27:51 localhost.localdomain setup-local-registry.sh[2104]: Generating self-signed certificate for local registry
 Feb 12 10:27:51 localhost.localdomain setup-local-registry.sh[2362]: {code}
 Expected results:
 {code:none}
 The ISO should mount successfully, and the start-local-registry.service should pull images without manifest-related errors.{code}
 Additional info:
 {code:none}
 [core@master-01 ~]$ dmesg | tail -50
 [ 2085.858677] loop262: detected capacity change from 0 to 6291456
 [ 2085.861062] loop263: detected capacity change from 0 to 6291456
 [ 2085.863354] loop264: detected capacity change from 0 to 6291456
 [ 2085.865628] loop265: detected capacity change from 0 to 6291456
 [ 2085.867838] loop266: detected capacity change from 0 to 6291456
 [ 2085.870319] loop267: detected capacity change from 0 to 6291456
 [ 2085.872614] loop268: detected capacity change from 0 to 6291456
 [ 2085.875191] loop269: detected capacity change from 0 to 6291456
 [ 2085.877599] loop270: detected capacity change from 0 to 6291456
 [ 2085.879915] loop271: detected capacity change from 0 to 6291456
 [ 2085.882252] loop272: detected capacity change from 0 to 6291456
 [ 2085.884609] loop273: detected capacity change from 0 to 5096940
 [ 2085.925774] podman: attempt to access beyond end of device
                dm-0: rw=0, sector=103829356, nr_sectors = 4 limit=99468780{code}
 
 The root cause is that when we replace the ISO in virtual media, Dell iDRAC sometimes retains the old session with the previous ISO, which causes the installation to fail. This can be resolved by restarting iDRAC.
Status: New
#OCPBUGS-77165issue13 hours agoImage not known errors causing mass failures across CI since cri-o 1.35.0 ASSIGNED
Issue 17919457: Image not known errors causing mass failures across CI since cri-o 1.35.0
Description: Initially this was found as a install failure where the network operator complains that multus is still awaiting 1-2 nodes on gcp. However the situation has now evolved to where we can see the underlying issue on many tests, jobs, and platforms:
 {code:java}
 Error: reading image "92c3c9025fe657a9160372a805bfd1624fb087262f2e3db326f937c298d47d70": locating image with ID "92c3c9025fe657a9160372a805bfd1624fb087262f2e3db326f937c298d47d70": image not known
 {code}
  
 
 Comments below show job runs where this is being seen.
 
  Scanning bigquery for tests reporting this in their output:
 {code:java}
 SELECT modified_time, prowjob_build_id, test_name, success, test_id, branch, prowjob_name, failure_content
 FROM `openshift-gce-devel.ci_analysis_us.junit` 
 WHERE   success = falseAND modified_time BETWEEN DATETIME("2026-02-10") AND DATETIME("2026-02-25")  
 AND failure_content like '%image not known%'  order by modified_time asc
 LIMIT 1000 {code}
 Shows again, explosion of this problem on Feb 19. It is happening only in 4.22 jobs and a few presubmits, ruling out infrastructure or registries. (I think?) I will attach the results of the above query in json format.
 
 Sample job runs:
 
 [https://prow.ci.openshift.org/view/gs/test-platform-results/logs/periodic-ci-openshift-release-main-ci-4.22-e2e-aws-ovn-techpreview/2025590450445881344]
 
 [https://prow.ci.openshift.org/view/gs/test-platform-results/logs/periodic-ci-openshift-release-main-ci-4.22-e2e-azure-ovn-techpreview/2025540689521020928]
 
 Or see the test details report below.
 
  
 
 [Search.ci|https://search.dptools.openshift.org/?search=image+not+known&maxAge=48h&context=1&type=junit&name=&excludeName=&maxMatches=5&maxBytes=20971520&groupBy=job] can be used to find job runs impacted. It doesn't always load, but if you wait 5-10 minutes and try again it usually will.
 h2. Original Report
 
  
 
 Component Readiness has found a potential regression in the following test:
 {code:none}
 install should succeed: overall{code}
 Extreme regression detected.
 Fishers Exact probability of a regression: 100.00%.
 Test pass rate dropped from 98.81% to 69.93%.
 
 Sample (being evaluated) Release: 4.22
 Start Time: 2026-02-16T00:00:00Z
 End Time: 2026-02-23T12:00:00Z
 Success Rate: 69.93%
 Successes: 200
 Failures: 86
 Flakes: 0
 Base (historical) Release: 4.20
 Start Time: 2025-09-21T00:00:00Z
 End Time: 2025-10-21T00:00:00Z
 Success Rate: 98.81%
 Successes: 1000
 Failures: 12
 Flakes: 0
 
 View the [test details report|https://sippy-auth.dptools.openshift.org/sippy-ng/component_readiness/test_details?Architecture=amd64&FeatureSet=default&Installer=ipi&LayeredProduct=none&Network=ovn&Platform=gcp&Suite=unknown&Topology=ha&Upgrade=micro&baseEndTime=2026-02-03T23%3A59%3A59Z&baseRelease=4.21&baseStartTime=2026-01-04T00%3A00%3A00Z&capability=Other&columnGroupBy=Network%2CPlatform%2CTopology&component=Installer+%2F+openshift-installer&confidence=95&dbGroupBy=Architecture%2CFeatureSet%2CInstaller%2CLayeredProduct%2CNetwork%2CPlatform%2CSuite%2CTopology%2CUpgrade&environment=Architecture%3Aamd64+FeatureSet%3Adefault+Installer%3Aipi+LayeredProduct%3Anone+Network%3Aovn+Platform%3Agcp+Suite%3Aunknown+Topology%3Aha+Upgrade%3Amicro&flakeAsFailure=false&ignoreDisruption=true&ignoreMissing=false&includeMultiReleaseAnalysis=true&includeVariant=Architecture%3Aamd64&includeVariant=Architecture%3Aarm64&includeVariant=Architecture%3Amulti&includeVariant=CGroupMode%3Av2&includeVariant=ContainerRuntime%3Acrun&includeVariant=ContainerRuntime%3Arunc&includeVariant=FeatureSet%3Adefault&includeVariant=FeatureSet%3Atechpreview&includeVariant=Installer%3Ahypershift&includeVariant=Installer%3Aipi&includeVariant=Installer%3Aupi&includeVariant=JobTier%3Ablocking&includeVariant=JobTier%3Ainforming&includeVariant=JobTier%3Astandard&includeVariant=LayeredProduct%3Anone&includeVariant=LayeredProduct%3Avirt&includeVariant=Network%3Aovn&includeVariant=OS%3Arhcos9&includeVariant=Owner%3Aeng&includeVariant=Owner%3Aservice-delivery&includeVariant=Platform%3Aaws&includeVariant=Platform%3Aazure&includeVariant=Platform%3Agcp&includeVariant=Platform%3Ametal&includeVariant=Platform%3Arosa&includeVariant=Platform%3Avsphere&includeVariant=Topology%3Aexternal&includeVariant=Topology%3Aha&includeVariant=Topology%3Amicroshift&includeVariant=Upgrade%3Amicro&includeVariant=Upgrade%3Aminor&includeVariant=Upgrade%3Anone&minFail=3&passRateAllTests=0&passRateNewTests=95&pity=5&sampleEndTime=2026-02-23T12%3A00%3A00Z&sampleRelease=4.22&sampleStartTime=2026-02-16T00%3A00%3A00Z&testBasisRelease=4.20&testId=cluster+install%3A0cb1bb27e418491b1ffdacab58c5c8c0] for additional context.
 
 Several other gcp install variant combos are showing problems, this is just the most visible one. See the triage record link that will be added to this card in a comment shortly for the full list of all regressions, but there should be more than enough job runs to investigate in this report linked above.
 
 I think I see multiple causes here, unclear if there's one unifying underlying cause. I see problems with storage volume mounting, network operator awaiting one node, possibly more.
 
 Consider this card a bug to get gcp install rates back up from their current 70% success rate.
 
 Filed by: [~dgoodwin@redhat.com]
 
 From what I can see, almost all of these runs are also showing a failure for:
 {code:java}
 verify operator conditions network
 
 {  Operator Progressing=True (Deploying): DaemonSet "/openshift-multus/multus-additional-cni-plugins" is not available (awaiting 2 nodes)}
 {code}
 Sometimes it's just one node. Yet all nodes seem to be fine.
 
 The test seems to be showing regressed in other areas too, here's a [metal regression that looks similar|https://sippy-auth.dptools.openshift.org/sippy-ng/component_readiness/test_details?Architecture=amd64&FeatureSet=default&Installer=ipi&LayeredProduct=none&Network=ovn&Platform=metal&Suite=serial&Topology=ha&Upgrade=none&baseEndTime=2026-02-03T23%3A59%3A59Z&baseRelease=4.21&baseStartTime=2026-01-04T00%3A00%3A00Z&capability=operator-conditions&columnGroupBy=Network%2CPlatform%2CTopology&component=Networking+%2F+cluster-network-operator&confidence=95&dbGroupBy=Architecture%2CFeatureSet%2CInstaller%2CLayeredProduct%2CNetwork%2CPlatform%2CSuite%2CTopology%2CUpgrade&environment=Architecture%3Aamd64+FeatureSet%3Adefault+Installer%3Aipi+LayeredProduct%3Anone+Network%3Aovn+Platform%3Ametal+Suite%3Aserial+Topology%3Aha+Upgrade%3Anone&flakeAsFailure=false&ignoreDisruption=true&ignoreMissing=false&includeMultiReleaseAnalysis=true&includeVariant=Architecture%3Aamd64&includeVariant=Architecture%3Aarm64&includeVariant=Architecture%3Amulti&includeVariant=CGroupMode%3Av2&includeVariant=ContainerRuntime%3Acrun&includeVariant=ContainerRuntime%3Arunc&includeVariant=FeatureSet%3Adefault&includeVariant=FeatureSet%3Atechpreview&includeVariant=Installer%3Ahypershift&includeVariant=Installer%3Aipi&includeVariant=Installer%3Aupi&includeVariant=JobTier%3Ablocking&includeVariant=JobTier%3Ainforming&includeVariant=JobTier%3Astandard&includeVariant=LayeredProduct%3Anone&includeVariant=LayeredProduct%3Avirt&includeVariant=Network%3Aovn&includeVariant=OS%3Arhcos9&includeVariant=Owner%3Aeng&includeVariant=Owner%3Aservice-delivery&includeVariant=Platform%3Aaws&includeVariant=Platform%3Aazure&includeVariant=Platform%3Agcp&includeVariant=Platform%3Ametal&includeVariant=Platform%3Arosa&includeVariant=Platform%3Avsphere&includeVariant=Topology%3Aexternal&includeVariant=Topology%3Aha&includeVariant=Topology%3Amicroshift&includeVariant=Upgrade%3Amicro&includeVariant=Upgrade%3Aminor&includeVariant=Upgrade%3Anone&minFail=3&passRateAllTests=0&passRateNewTests=95&pity=5&sampleEndTime=2026-02-23T12%3A00%3A00Z&sampleRelease=4.22&sampleStartTime=2026-02-16T00%3A00%3A00Z&testId=openshift-tests%3A25ad49f2396feeec12df2da38f57bc3f].
 
 [Global test analysis|https://sippy-auth.dptools.openshift.org/sippy-ng/tests/4.22/analysis?test=verify%20operator%20conditions%20network&filters=%7B%22items%22%3A%5B%7B%22columnField%22%3A%22name%22%2C%22operatorValue%22%3A%22equals%22%2C%22value%22%3A%22verify%20operator%20conditions%20network%22%7D%5D%2C%22linkOperator%22%3A%22and%22%7D] shows a sharp decline probably around the 19th.
 
 Mostly hitting gcp but some metal in the mix as well.
Status: ASSIGNED
{code}
event happened 25 times, something is wrong: namespace/openshift-ingress node/ci-op-mw99dt7j-fb783-4fdxg-worker-eastus1-jjtbt pod/gateway-cl995-openshift-default-79dd68f786-pkzcq hmsg/f04e4a5657 - reason/Failed Error: reading image "92c3c9025fe657a9160372a805bfd1624fb087262f2e3db326f937c298d47d70": locating image with ID "92c3c9025fe657a9160372a805bfd1624fb087262f2e3db326f937c298d47d70": image not known (14:27:15Z) result=reject }
{code}
{code}
console-7b4b4768bb-66s4w hmsg/efa454141e - reason/Failed Error: reading image "a81a5024f77f881f002b28dc2d8be8adfbc4c63e3ee7981c0c169bc6f03cf46e": locating image with ID "a81a5024f77f881f002b28dc2d8be8adfbc4c63e3ee7981c0c169bc6f03cf46e": image not known (17:54:22Z) result=reject event happened 557 times, something is wrong: namespace/openshift-console node/ip-10-0-17-146.ec2.internal pod/console-7b4b4768bb-66s4w hmsg/6cba065bf0 - reason/Pulled Container image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:923d6a13397a5a3b506536c6bdce57b37aaa0bf5b95c952f59d49f76b0a00e8f" already present on machine (17:54:22Z) result=reject }
{code}
#OCPBUGS-78024issue34 hours ago[openshift-4.13] ose-vsphere-cluster-api-controllers can't be built hermetically New
Issue 17969358: [openshift-4.13] ose-vsphere-cluster-api-controllers can't be built hermetically
Description: This is a clone of issue OCPBUGS-78023. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78021. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78020. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78019. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78018. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78017. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-vsphere-cluster-api-controllers hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-77627issue7 days ago[4.18] openshift-cluster-csi-drivers vmware-vsphere-csi-driver-operator-metrics endpoint 8445 port should return 401 instead 500 ASSIGNED
Issue 17948243: [4.18] openshift-cluster-csi-drivers vmware-vsphere-csi-driver-operator-metrics endpoint 8445 port should return 401 instead 500
Description: This is a clone of issue OCPBUGS-77007. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-75912. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-74569. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-60159. The following is the description of the original issue:  --- Description of problem:
 
 see similar bug OCPBUGS-57585
 
 Stole the testing result from [https://github.com/openshift/origin/pull/30014#issuecomment-3153130188]
 
 8445 port is for openshift-cluster-csi-drivers vmware-vsphere-csi-driver-operator-metrics endpoint, did not see this port on other IAAS(example AWS,OSP)checked on vsphere, curl 8445 without auth returns 500(actually it's forbidden),
 {noformat}
 $ oc get infrastructure/cluster -o jsonpath={.spec.platformSpec.type}
 VSphere
 
 $ oc -n openshift-cluster-csi-drivers get ep
 Warning: v1 Endpoints is deprecated in v1.33+; use discovery.k8s.io/v1 EndpointSlice
 NAME                                           ENDPOINTS                                                        AGE
 vmware-vsphere-csi-driver-controller-metrics   10.129.0.17:9203,10.130.0.21:9203,10.129.0.17:9202 + 9 more...   28m
 vmware-vsphere-csi-driver-operator-metrics     10.130.0.17:8445                                                 28m
 vmware-vsphere-csi-driver-webhook-svc          10.128.0.45:8443,10.130.0.23:8443                                28m
 
 $ oc -n openshift-monitoring exec -c prometheus prometheus-k8s-0 -- curl -k -s -o /dev/null -w '%{http_code}' https://10.130.0.17:8445/metrics
 500
 {noformat}
 return code is 500, expecting 401 forbidden instead of 500.
 {code:java}
 $ oc -n openshift-monitoring exec -c prometheus prometheus-k8s-0 -- curl -k https://10.130.0.17:8445/metrics
 
 Internal Server Error: "/metrics": subjectaccessreviews.authorization.k8s.io is forbidden: User "system:serviceaccount:openshift-cluster-csi-drivers:vmware-vsphere-csi-driver-operator" cannot create resource "subjectaccessreviews" in API group "authorization.k8s.io" at the cluster scope  {code}
 curl with auth, no issue
 {noformat}
 $ token=`oc create token prometheus-k8s -n openshift-monitoring`
 $ oc -n openshift-monitoring exec -c prometheus prometheus-k8s-0 -- curl -k -H "Authorization: Bearer $token" 'https://10.130.0.17:8445/metrics' | head -n3
 # HELP aggregator_discovery_aggregation_count_total [ALPHA] Counter of number of times discovery was aggregated
 # TYPE aggregator_discovery_aggregation_count_total counter
 aggregator_discovery_aggregation_count_total 0 {noformat}
 [https://github.com/openshift/origin/pull/30014#issuecomment-3154900379]
 
  
 
 When the bug is fixed, please remove it from this code snippet to avoid regression.
 
 [https://github.com/openshift/origin/blob/4f183dd3427cffd8d97b44557caa782d65726416/test/extended/prometheus/prometheus.go#L64-L71]
Status: ASSIGNED
#OCPBUGS-78174issue9 hours ago[openshift-4.13] ose-machine-api-provider-openstack can't be built hermetically New
Issue 17974704: [openshift-4.13] ose-machine-api-provider-openstack can't be built hermetically
Description: This is a clone of issue OCPBUGS-78172. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78171. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78170. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78169. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78168. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-machine-api-provider-openstack hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-77278issue7 days agoBMH stuck in deleting state while performing IBBF Verified
Issue 17931739: BMH stuck in deleting state while performing IBBF
Description: This is a clone of issue OCPBUGS-76988. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-59133. The following is the description of the original issue:  --- Description of problem:
 {code:none}
     While attempting to perform IBBF with the siteconfig operator, the BMH for the spoke is stuck in deleting state for too long due to the old BMH being unreachable. This results in the clusterinstance timing out while waiting for the BMH resource to be removed. We should find a way to mitigate this.
 {code}
     Version-Release number of selected component (if applicable):
 {code:java}
 4.18.0-0.nightly-2025-04-02-011053 {code}
 How reproducible:
 {code:none}
     100%{code}
 Steps to Reproduce:
 {code:none}
     1. Install an IBI spoke cluster using siteconfig operator
     2. Remove old node and attempt to replace with new node through IBBF
     3.
     {code}
 Actual results:
 {code:none}
     Reinstallation fails due to timeout{code}
 Expected results:
 {code:none}
     Reinstallation succeeds{code}
 Additional info:
 {code:none}
 $ oc get clusterinstance target-0 -o json | jq -r '.status.reinstall.conditions[0]'
 {
   "lastTransitionTime": "2025-04-04T17:24:45Z",
   "message": "Encountered error executing task: Deleting rendered manifests. Error: deletion timeout exceeded for object (BareMetalHost:target-0/target-0-0): Timed out waiting to delete object (BareMetalHost:target-0/target-0-0)",
   "reason": "Failed",
   "status": "False",
   "type": "ReinstallRequestProcessed"
 }
 $ oc get bmh target-0-0 -o json | jq -r .status.errorMessage
 Failed to get power state for node 7e063860-56c7-45ca-8bfa-d8f63997bd10. Error: Redfish exception occurred. Error: Resource https://[fd2e:6f44:5dd8:4::1]:8000/redfish/v1/Systems/ac8d214a-ae10-46f5-a179-208341bcb688 not found {code}
Status: Verified
#OCPBUGS-77192issue2 weeks ago[openshift-4.20] ose-libvirt-machine-controllers can't be build hermetically New
Issue 17925856: [openshift-4.20] ose-libvirt-machine-controllers can't be build hermetically
Description: Description of problem:
 
 ART can't build ose-libvirt-machine-controllers hermetically.
 The error is the following in the prefetch-dependencies konflux task:
 {code:none}
     Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.  Please try running `go mod vendor` and committing the changes.  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
  
 
  
 
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-76610issue6 hours agoMCD Validation Discrepancy during Interrupted rpm-ostree Finalization New
Issue 17867884: MCD Validation Discrepancy during Interrupted rpm-ostree Finalization
Description: *Summary:* MCD enters Degraded loop when `rpm-ostree` finalization fails due to discrepancy between on-disk `currentconfig` and Node annotation
 
 Issue Type: Bug
 Priority: High (Severity 2)
 Component: Machine Config Operator (MCO) , RHCOS
 Labels: shift_sno, mco, rpm-ostree
 
 *Description :*
 
 In OpenShift 4.18/4.19, the Machine Config Daemon (MCD) can enter a persistent "Degraded" state if a node reboot is interrupted or if `ostree-finalize-staged.service` fails to complete (e.g., due to a timeout).
 
 During this failure mode, the MCD writes the "new" configuration to the on-disk `currentconfig` file before the kernel has successfully pivoted into it. Upon reboot, the MCD prioritizes the on-disk file over the Node's `currentConfig` annotation, leading to a validation failure because the actual running state (kernel arguments, etc.) matches the old configuration, not the "new" one recorded on disk.
 
 *Steps to Reproduce :* 
 
 1. Trigger a Configuration Change: Apply a `MachineConfig` that requires a reboot and additional kernel arguments (e.g., `audit=1`, `page_poison=1`).
 2. Simulate/Induce Finalization Failure: Interrupt the `ostree-finalize-staged.service` during the shutdown/reboot process so it exits with a 'timeout' or failure result.
 3. Boot into Old Deployment: Allow the node to reboot. Because finalization failed, the node remains on the previous RHCOS deployment/kernel version.
 4. Observe MCD Logs: Check the logs of the `machine-config-daemon` pod.
 
 *Actual Results :* 
  * Validation Error: MCD detects that the on-disk file overrides the node annotation:
 `"Disk currentConfig \"rendered-master-NEW\" overrides node's currentConfig annotation \"rendered-master-OLD\""`.
  * State Mismatch: MCD validates the system against the NEW config but finds the OLD state, resulting in a Degraded status:
 `"unexpected on-disk state validating against rendered-master-NEW: missing expected kernel arguments: [...]"`.
  * Persistent Loop: The node remains Degraded and fails to reconcile because it is stuck validating against a configuration it never successfully booted.
 
 *Expected Results :*
  * The MCD should verify that the booted deployment matches the configuration defined in the on-disk `currentconfig` before using it for validation.
  * If a discrepancy exists (i.e., the pivot failed), the MCD should fallback to the Node's `currentConfig` annotation or automatically clean up the stale on-disk file to allow for a clean reconciliation.
 
 *Environment Information :* 
  * OCP Version: 4.18 / 4.19
  * Platform: Single Node OpenShift (SNO), Disconnected, Bare Metal
  * OS: RHCOS
  * Disconnected Air Gapped and Secured Environment  
 
 *Workaround :* 
 
 1. Access the node via SSH.
 2. Backup and delete the stale on-disk config: `rm /etc/machine-config-daemon/currentconfig`.
 3. Reboot the node: `systemctl reboot`.
 4. After reboot, the MCD will correctly reconcile against the Node annotation.
Status: New
#OCPBUGS-77212issue3 days ago[openshift-4.14] hermetic build failure - ose-ibm-vpc-block-csi-driver Verified
Issue 17926182: [openshift-4.14] hermetic build failure - ose-ibm-vpc-block-csi-driver
Description: This is a clone of issue OCPBUGS-77211. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77210. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77209. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77207. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77206. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77205. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77203. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77167. The following is the description of the original issue:  --- Description of problem:
 {code:none}
 IBM VPC Block CSI Driver build fails in hermetic mode due to non-deterministic GOPATH handling in Makefile. The build system rewrites $GOPATH causing the binary output path -o ${GOPATH}/bin/${EXE_DRIVER_NAME} to become invalid, resulting in build failures.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
 OpenShift 4.19+ (affects hermetic build conversion) {code}
 How reproducible:
 {code:none}
 Always - when building ibm-vpc-block-csi-driver in hermetic mode{code}
 Steps to Reproduce:
 {code:none}
   1. Build ibm-vpc-block-csi-driver container in hermetic mode
   2. Makefile uses absolute GOPATH-dependent output path: -o ${GOPATH}/bin/${EXE_DRIVER_NAME}
   3. Build system rewrites $GOPATH during hermetic build process     {code}
 Actual results:
 {code:none}
 Build fails because binary cannot be found at expected GOPATH location after build system modifies $GOPATH
 
 Error: building at STEP "COPY --from=builder /cachi2/output/deps/gomod/bin/ibm-vpc-block-csi-driver /bin/ibm-vpc-block-csi-driver": checking on sources under [...]: copier: stat: "/cachi2/output/deps/gomod/bin/ibm-vpc-block-csi-driver": no such file or directory{code}
 Expected results:
 {code:none}
 Build succeeds using relative path bin/${EXE_DRIVER_NAME} that doesn't depend on GOPATH manipulation, consistent with other CSI drivers like aws-ebs-csi-driver  {code}
 Additional info:
 {code:none}
 Fix implemented: Change Makefile to use relative path -o bin/${EXE_DRIVER_NAME} instead of absolute GOPATH-dependent path. This aligns with other CSI drivers and makes builds hermetic-compatible. Comment added to indicate this is a permanent carry patch for OpenShift builds.{code}
Status: Verified
#OCPBUGS-55106issue5 weeks agoRHCOS 4.19 live iso x86_64 contains restrictive policy.json Verified
Issue 16833641: RHCOS 4.19 live iso x86_64 contains restrictive policy.json
Description: Description of problem:
 {code:none}
 Starting April 10th, RHCOS 4.19 live iso (from 4.19 ocp installer reference) contains a restrictive */etc/containers/policy.json* that causes OCP clusters deployed by assisted service to fail in certain scenarios.
 {code}
 Version-Release number of selected component (if applicable):
 {code:none}
 Trying to deploy 4.19.0-ec4
 
 Booted image information:
 
 NAME="Red Hat Enterprise Linux CoreOS"
 VERSION="9.6.20250321-0 (Plow)"
 ID="rhel"
 ID_LIKE="fedora"
 VERSION_ID="9.6"
 PLATFORM_ID="platform:el9"
 PRETTY_NAME="Red Hat Enterprise Linux CoreOS 9.6.20250321-0 (Plow)"
 ANSI_COLOR="0;31"
 LOGO="fedora-logo-icon"
 CPE_NAME="cpe:/o:redhat:enterprise_linux:9::baseos"
 HOME_URL="https://www.redhat.com/"
 DOCUMENTATION_URL="https://access.redhat.com/documentation/en-us/red_hat_enterprise_linux/9"
 BUG_REPORT_URL="https://issues.redhat.com/"
 REDHAT_BUGZILLA_PRODUCT="Red Hat Enterprise Linux 9"
 REDHAT_BUGZILLA_PRODUCT_VERSION=9.6
 REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux"
 REDHAT_SUPPORT_PRODUCT_VERSION="9.6"
 OSTREE_VERSION='9.6.20250321-0'
 VARIANT=CoreOS
 VARIANT_ID=coreos
 Red Hat Enterprise Linux release 9.6 (Plow)
 Red Hat Enterprise Linux release 9.6 (Plow)  {code}
 How reproducible:
 {code:none}
 100%
 
 {code}
 Steps to Reproduce:
 {code:none}
 1. Run coreos image
 2. cat /etc/containers/policy.json
 
 {code}
 Actual results:
 {code:none}
 [core@sno ~]$ cat /etc/containers/policy.json 
 {
     "default": [
         {
             "type": "insecureAcceptAnything"
         }
     ],
     "transports": {
         "docker": {
             "registry.access.redhat.com": [
                 {
                     "type": "signedBy",
                     "keyType": "GPGKeys",
                     "keyPaths": ["/etc/pki/rpm-gpg/RPM-GPG-KEY-redhat-release", "/etc/pki/rpm-gpg/RPM-GPG-KEY-redhat-beta"]                                                                                                                          
                 }
             ],
             "registry.redhat.io": [
                 {
                     "type": "signedBy",
                     "keyType": "GPGKeys",
                     "keyPaths": ["/etc/pki/rpm-gpg/RPM-GPG-KEY-redhat-release", "/etc/pki/rpm-gpg/RPM-GPG-KEY-redhat-beta"]
                 }
             ]
         },
         "docker-daemon": {
             "": [
                 {
                     "type": "insecureAcceptAnything"
                 }
             ]
         }
     }
 }  {code}
 Expected results:
 {code:none}
 {
     "default": [
         {
             "type": "insecureAcceptAnything"
         }
     ],
     "transports": {
         "docker-daemon": {
         "": [
         {
             "type": "insecureAcceptAnything"
         }
         ]
     }
     }
 }
 
 {code}
 {code:java}
 There was no change to the configuration and the deployment the day before (On April 9th completed successfully). The issue persists since then on every deployment attempt.{code}
 The issue itself is that with those records, agent image cannot be downloaded:
 {code:java}
 Apr 17 12:05:33 sno.xr11.kni-qe-9.lab.eng.rdu2.redhat.com systemd[1]: Starting agent.service...
 Apr 17 12:05:34 sno.xr11.kni-qe-9.lab.eng.rdu2.redhat.com podman[971303]: Trying to pull registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:49f93c9d88aac3a1a4bb9df2f46d52ea2581c05095a42b01993ea15594b21aa9...
 Apr 17 12:05:34 sno.xr11.kni-qe-9.lab.eng.rdu2.redhat.com podman[971303]: Error: copying system image from manifest list: Source image rejected: A signature was required, but no signature exists
 Apr 17 12:05:34 sno.xr11.kni-qe-9.lab.eng.rdu2.redhat.com podman[971303]: 2025-04-17 12:05:34.17935456 +0000 UTC m=+0.165353038 image pull-error  registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:49f93c9d88aac3a1a4bb9df2f46d52ea2581c05095a42b01993ea1559>
 Apr 17 12:05:34 sno.xr11.kni-qe-9.lab.eng.rdu2.redhat.com systemd[1]: agent.service: Control process exited, code=exited, status=125/n/a
 Apr 17 12:05:34 sno.xr11.kni-qe-9.lab.eng.rdu2.redhat.com systemd[1]: agent.service: Failed with result 'exit-code'.
 Apr 17 12:05:34 sno.xr11.kni-qe-9.lab.eng.rdu2.redhat.com systemd[1]: Failed to start agent.service.
  {code}
 Once `docker` section is removed the image is pulled and the deployment continues.
Status: Verified
Apr 17 12:05:34 sno.xr11.kni-qe-9.lab.eng.rdu2.redhat.com podman[971303]: Trying to pull registry.redhat.io/multicluster-engine/assisted-installer-agent-rhel9@sha256:49f93c9d88aac3a1a4bb9df2f46d52ea2581c05095a42b01993ea15594b21aa9...
Apr 17 12:05:34 sno.xr11.kni-qe-9.lab.eng.rdu2.redhat.com podman[971303]: Error: copying system image from manifest list: Source image rejected: A signature was required, but no signature exists
{code}
#OCPBUGS-78101issue17 hours ago[openshift-4.14] baremetal-machine-controller can't be built hermetically New
Issue 17972263: [openshift-4.14] baremetal-machine-controller can't be built hermetically
Description: This is a clone of issue OCPBUGS-78100. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78099. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78098. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build baremetal-machine-controller hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-78021issue34 hours ago[openshift-4.15] ose-vsphere-cluster-api-controllers can't be built hermetically New
Issue 17969353: [openshift-4.15] ose-vsphere-cluster-api-controllers can't be built hermetically
Description: This is a clone of issue OCPBUGS-78020. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78019. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78018. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78017. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-vsphere-cluster-api-controllers hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-77197issue2 weeks ago[openshift-4.15] ose-libvirt-machine-controllers can't be build hermetically New
Issue 17925863: [openshift-4.15] ose-libvirt-machine-controllers can't be build hermetically
Description: Description of problem:
 
 ART can't build ose-libvirt-machine-controllers hermetically.
 The error is the following in the prefetch-dependencies konflux task:
 {code:none}
     Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.  Please try running `go mod vendor` and committing the changes.  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
  
 
  
 
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-59211issue15 hours agoToo many logs related to retried oauthaccesstokens from openshift-kube-storage-version-migrator namespace POST
Issue 17134880: Too many logs related to retried oauthaccesstokens from openshift-kube-storage-version-migrator namespace
Description: Description of problem:
 
 Too many logs related to multiple retried oauthaccesstokens from openshift-kube-storage-version-migrator namespace. Example logs are mentioned below:
 
  
 {code:java}
 2025-07-07T07:53:01.002259691Z W0707 07:53:01.002152       1 core.go:223] migration of sha256~xxxxxxxxxxxx will be retried: Operation cannot be fulfilled on oauthaccesstokens.oauth.openshift.io "sha256~xxxxxxxxxxxx": StorageError: invalid object, Code: 4, Key: /openshift.io/oauth/accesstokens/sha256~xxxxxxxxxxxx, ResourceVersion: 0, AdditionalErrorMsg: Precondition failed: UID in precondition: 1c3xxxx-7xxx-4axx-9xxx-fxxxxxxxxxx0, UID in object meta: {code}
  
 
 The issue keeps re-occurring with different tokens.
Status: New
#OCPBUGS-77649issue6 days ago[openshift-4.22] hermetic build failure - monitoring-plugin Verified
Issue 17948613: [openshift-4.22] hermetic build failure - monitoring-plugin
Description: Description of problem:
 {code:none}
   nmstate-console-plugin fails to build in hermetic mode due to yarn package manager incompatibility and Cypress test framework attempting network downloads during build process.
   {code}
 Version-Release number of selected component (if applicable):
 {code:none}
   OpenShift Container Platform 4.14+
   nmstate-console-plugin component across all affected versions
   {code}
 How reproducible:
 {code:none}
   Always - occurs when converting any nmstate-console-plugin build from network_mode: open to hermetic mode
   {code}
 Steps to Reproduce:
 {code:none}
   1. Remove network_mode: open from nmstate-console-plugin.yml configuration
   2. Attempt hermetic build via Konflux build system
   3. Build fails during npm install phase when Cypress tries to download binaries
   {code}
 Actual results:
 {code:none}
   npm error Installing Cypress (version: 12.17.4)
   npm error The Cypress App could not be downloaded.
   npm error URL: https://download.cypress.io/desktop/12.17.4?platform=linux&arch=x64
   npm error Error: getaddrinfo ENOTFOUND download.cypress.io
 
   Build fails because:
   - Yarn package manager not supported by cachi2 for hermetic dependency management
   - Cypress testing framework requires network access to download.cypress.io
   - Custom yarn artifacts and dockerfile modifications conflict with hermetic requirements
   {code}
 Expected results:
 {code:none}
   nmstate-console-plugin builds successfully in hermetic mode without requiring network access during build process. Component should use npm package manager and disable Cypress binary
   downloads via CYPRESS_INSTALL_BINARY=0 environment variable.
   {code}
 Additional info:
 {code:none}
   Root cause: Component uses yarn package manager which is incompatible with cachi2 hermetic builds, plus Cypress test framework downloads binaries during build.
 
   Solution requires:
   1. Upstream yarn→npm migration (similar to networking-console-plugin)
   2. Set CYPRESS_INSTALL_BINARY=0 to disable test framework downloads
   3. Remove yarn-specific artifacts and dockerfile modifications
   4. Update ocp-build-data package manager configuration
 
   Reference upstream work: https://github.com/openshift/nmstate-console-plugin/pull/169
   Similar successful conversion: networking-console-plugin PRs #347 and #331
 
   Tracking: ART-14616 (parent task with technical analysis)
   {code}
Status: Verified
#OCPBUGS-77676issue7 days agoNNCP failing to reconfigure due to linux-bridge having extra ports after already configured CLOSED
Issue 17950683: NNCP failing to reconfigure due to linux-bridge having extra ports after already configured
Description: Description of problem:
 {code:none}
    Linux bridge was configured by nmstate to be used by cnv-bridge cni for vms. After configuration ports were added to bridge by cni, some event triggered nmstate to reconfigure/validate configuration and now it is in degraded/FailedToConfigure state due to extra ports. {code}
 Version-Release number of selected component (if applicable):
 {code:none}
     4.18{code}
 How reproducible:
 {code:none}
 Unknown     {code}
 Steps to Reproduce:
 {code:none}
     1. Create a linux bridge with nmstate with nncp 
 
 spec:
   desiredState:
     interfaces:
     - mtu: 9000
       name: bond0.33
       state: up
       type: vlan
       vlan:
         base-iface: bond0
         id: 33
     - bridge:
         options:
           stp:
             enabled: false
         port:
         - name: bond0.33
           vlan: {}
       description: Linux bridge with bond0 as a port on vlan 33
       mtu: 9000
       name: br0.33
       state: up
       type: linux-bridge
 
     2. create a net-attach-def for br0.33 
 
 spec:
   config: |
     {
       "cniVersion": "0.3.1",
       "name": "vlan-33",
       "type": "cnv-bridge",
       "bridge": "br0.33",
       "macspoofchk": true
     }
 
 
     3. After nmstate configures bridge, have cni add ports 
     4. Wait for an event that triggers nmstate to re-validate or re-configure (Not sure what triggers this nmstate to reconfigure/re-validate )
 
     {code}
 Actual results:
 {code:none}
   NNCP  vlan-33-bridge       Degraded      FailedToConfigure
   NNCE  Failing
 
 Logs:       [2026-02-17T15:56:54Z INFO  nmstate::query_apply::net_state] Retrying on: VerificationError: Verification failure: br0.33.interface.bridge.port desire '[{"name":"bond0.33","vlan":{}}]', current '[{"name":"bond0.33","stp-hairpin-mode":false,"stp-path-cost":100,"stp-priority":32,"vlan":{}},{"name":"veth8abc333d","stp-hairpin-mode":false,"stp-path-cost":2,"stp-priority":32,"vlan":{}}]'
 
 
 
 {code}
  
 
 Expected results:
 {code:none}
     For nmstate to not fail as its already configured bridge and allow CNI is adding ports. {code}
 Additional info:
 {code:none}
 Allow linux bridges to use something similar to what can be configured for ovs bridges (allow-extra-patch-ports: true) {code}
Status: CLOSED
#OCPBUGS-58262issue8 weeks agopodman Error: trying to reuse blob - failed to verify certificate: x509: certificate signed by unknown authority ASSIGNED
Issue 17106784: podman Error: trying to reuse blob - failed to verify certificate: x509: certificate signed by unknown authority
Description: Description of problem:
 {code:none}
     Can't reuse blob to push aws c2s bucket with x509 error since Dec. And on prow ci result, it failed in pull image from internal registry{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     4.18{code}
 How reproducible:
 {code:none}
     always{code}
 Steps to Reproduce:
 {code:none}
     1.Set up an aws c2s cluster with 4.18 using prow
 https://gcsweb-qe-private-deck-ci.apps.ci.l2s4.p1.openshiftapps.com/gcs/qe-private-deck/logs/periodic-ci-openshift-openshift-tests-private-release-4.18-amd64-nightly-aws-sc2s-ipi-disc-priv-fips-f2/1865304153111138304/artifacts/aws-sc2s-ipi-disc-priv-fips-f2/openshift-extended-test/build-log.txt
 
       build logs : 2024-12-07T11:25:33.257156973Z Replaced Dockerfile FROM image quay.io/openshifttest/busybox@sha256:c5439d7db88ab5423999530349d327b04279ad3161d7596d2126dfb5b02bfd1f
     2024-12-07T11:25:35.613930151Z time="2024-12-07T11:25:35Z" level=info msg="Not using native diff for overlay, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled"
     2024-12-07T11:25:35.614776370Z I1207 11:25:35.614748       1 defaults.go:112] Defaulting to storage driver "overlay" with options [mountopt=metacopy=on].
     2024-12-07T11:25:35.629453348Z Caching blobs under "/var/cache/blobs".
     2024-12-07T11:25:35.632143910Z 
     2024-12-07T11:25:35.632143910Z Pulling image image-registry.openshift-image-registry.svc:5000/e2e-test-default-image-registry-zgmf2/busybox@sha256:c5439d7db88ab5423999530349d327b04279ad3161d7596d2126dfb5b02bfd1f ...
     2024-12-07T11:25:35.636791870Z Trying to pull image-registry.openshift-image-registry.svc:5000/e2e-test-default-image-registry-zgmf2/busybox@sha256:c5439d7db88ab5423999530349d327b04279ad3161d7596d2126dfb5b02bfd1f...
     2024-12-07T11:25:36.695243084Z Warning: Pull failed, retrying in 5s ...
     2024-12-07T11:25:41.698639710Z Trying to pull image-registry.openshift-image-registry.svc:5000/e2e-test-default-image-registry-zgmf2/busybox@sha256:c5439d7db88ab5423999530349d327b04279ad3161d7596d2126dfb5b02bfd1f...
     2024-12-07T11:25:41.848354202Z Warning: Pull failed, retrying in 5s ...
     2024-12-07T11:25:46.852090834Z Trying to pull image-registry.openshift-image-registry.svc:5000/e2e-test-default-image-registry-zgmf2/busybox@sha256:c5439d7db88ab5423999530349d327b04279ad3161d7596d2126dfb5b02bfd1f...
     2024-12-07T11:25:47.406620628Z Warning: Pull failed, retrying in 5s ...
     2024-12-07T11:25:52.439067513Z error: build error: failed to pull image: After retrying 2 times, Pull image still failed due to error: copying system image from manifest list: parsing image configuration: Get "https://ci-op-5gnlt4mn-c6dd8-6xdhm-image-registry-us-isob-east-1-sxuku.s3.us-isob-east-1.sc2s.sgov.gov/docker/registry/v2/blobs/sha256/b9/b97242f89c8a29d13aea12843a08441a4bbfc33528f55b60366c1d8f6923d0d4/data?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=ASIAUW2DH7NUFSKUEAEQ%2F20241207%2Fus-isob-east-1%2Fs3%2Faws4_request&X-Amz-Date=20241207T112547Z&X-Amz-Expires=1200&X-Amz-SignedHeaders=host&X-Amz-Signature=6aba8b40b4b4f43e49074fcdae2db73c10c648c93ce9fb4d15bcd114925ef713": tls: failed to verify certificate: x509: certificate signed by unknown authority
 
     2. But when reproduce it on prow ci cluster. we could push image at first try as always, but failed to push with reuse blob.
 oc get builds
 NAME              TYPE     FROM         STATUS                               STARTED       DURATION
 test-registry-1   Docker   Dockerfile   Complete                             2 hours ago   15s
 test-registry-2   Docker   Dockerfile   Failed (PushImageToRegistryFailed)   2 hours ago   21s  
 
 oc logs -f test-registry-2-build
 error: build error: Failed to push image: trying to reuse blob sha256:530afca65e2ea04227630ae746e0c85b2bd1a179379cbf2b6501b49c4cab2ccc at destination: Head "https://ci-op-qf55lf0q-3c6ac-j65q8-image-registry-us-iso-east-1-soajkc.s3.us-iso-east-1.c2s.ic.gov/docker/registry/v2/blobs/sha256/53/530afca65e2ea04227630ae746e0c85b2bd1a179379cbf2b6501b49c4cab2ccc/data?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=ASIAUW2DH7NUKYUBPOG2%2F20241210%2Fus-iso-east-1%2Fs3%2Faws4_request&X-Amz-Date=20241210T070804Z&X-Amz-Expires=1200&X-Amz-SignedHeaders=host&X-Amz-Signature=f5a44e20c1f5548e8cbec55f0faa0677fea306ecc8f1008a298a6171b8969a79": tls: failed to verify certificate: x509: certificate signed by unknown authority   
 
 3. Can't push a existing image to s3 bucket using podman
 
  podman push image-registry.openshift-image-registry.svc:5000/wxj/test:latest
 Getting image source signatures
 Copying blob 1a8c6bfa0a12 done   |
 Copying blob 42e40303859e done   |
 Copying blob 9c6be2aa7dfd done   |
 Copying blob b80a8be6331b done   |
 Copying config cda9cd0379 done   |
 Writing manifest to image destination
 sh-5.1# podman push image-registry.openshift-image-registry.svc:5000/wxj/test:latest
 Getting image source signatures
 Copying blob 9c6be2aa7dfd done   |
 Copying blob 42e40303859e done   |
 Copying blob 1a8c6bfa0a12 done   |
 Copying blob b80a8be6331b done   |
 Copying config cda9cd0379 [--------------------------------------] 8.0b / 20.7KiB | 7.8 MiB/s
 Error: writing blob: checking whether a blob sha256:cda9cd0379eae8c992bf505926f7a9ea95f93b876cb55b15c9345b0d41a45388 exists in image-registry.openshift-image-registry.svc:5000/wxj/test: StatusCode: 403, ""     {code}
 Actual results:
 {code:none}
 Failed to pull or push reuse blob from/to image registry.{code}
 Expected results:
 {code:none}
      Shouldn't fail to pull/push image from internal registry.{code}
 Additional info:
 {code:none}
 must gather log https://gcsweb-qe-private-deck-ci.apps.ci.l2s4.p1.openshiftapps.com/gcs/qe-private-deck/logs/periodic-ci-openshift-openshift-tests-private-release-4.18-amd64-nightly-aws-sc2s-ipi-disc-priv-fips-f2/1865304153111138304/artifacts/aws-sc2s-ipi-disc-priv-fips-f2/gather-must-gather/{code}
Status: ASSIGNED
#OCPBUGS-77763issue11 hours ago[openshift-4.21] hermetic build failure - node-feature-discovery CLOSED
Issue 17953316: [openshift-4.21] hermetic build failure - node-feature-discovery
Description: This is a clone of issue OCPBUGS-77651. The following is the description of the original issue:  --- Description of problem:
 {code:none}
   Node Feature Discovery build fails in hermetic mode due to non-deterministic binary path handling in Dockerfile. The build uses "go install" command which places binaries in unpredictable GOPATH locations, and the Dockerfile copies from multiple inconsistent paths causing binary path conflicts.
   {code}
 Version-Release number of selected component (if applicable):
 {code:none}
   OpenShift Container Platform 4.19+
   node-feature-discovery component across all affected versions
   {code}
 How reproducible:
 {code:none}
   Always - when building node-feature-discovery in hermetic mode
   {code}
 Steps to Reproduce:
 {code:none}
   1. Build node-feature-discovery container in hermetic mode
   2. Dockerfile copies binaries from multiple paths: /go/bin/* and /go/node-feature-discovery/bin/*
   3. "go install" command places binaries in non-deterministic GOPATH locations during hermetic build process
   {code}
 Actual results:
 {code:none}
   Build fails because binaries are placed in unpredictable locations by "go install" and Dockerfile attempts to copy from multiple conflicting paths
 
   Error: Inconsistent binary placement causing copy failures during Docker image build process when switching from "go install" to "go build" in
   hermetic mode
   {code}
 Expected results:
 {code:none}
   Build succeeds using deterministic binary paths. Should use "go build" with explicit output paths instead of "go install" and copy binaries from single consistent location in Dockerfile.
   {code}
 Additional info:
 {code:none}
   Root cause: "go install" command places binaries in GOPATH-dependent locations that are not deterministic in hermetic builds, unlike "go build"
    which allows explicit output path control.
 
   Fix implemented: Remove redundant /go/bin/* copy command from Dockerfile and keep only the specific /go/node-feature-discovery/bin/* path. This
    ensures consistent binary placement and eliminates path conflicts during hermetic builds.
 
   Reference upstream work: https://github.com/openshift/node-feature-discovery/pull/163/changes
   Related hermetic conversion tracking: ART-14586
   {code}
Status: CLOSED
#OCPBUGS-77414issue7 days ago[openshift-4.22] hermetic build failure - networking-console-plugin ON_QA
Issue 17936763: [openshift-4.22] hermetic build failure - networking-console-plugin
Description: Description of problem:
 {code:none}
   networking-console-plugin fails to build in hermetic mode due to yarn package manager incompatibility and Cypress test framework attempting network downloads during build process.
   {code}
 Version-Release number of selected component (if applicable):
 {code:none}
   OpenShift Container Platform 4.16+ networking-console-plugin component across all affected versions
   {code}
 How reproducible:
 {code:none}
   Always - occurs when converting any networking-console-plugin build from network_mode: open to hermetic mode
   {code}
 Steps to Reproduce:
 {code:none}
   1. Remove network_mode: open from networking-console-plugin.yml configuration
   2. Attempt hermetic build via Konflux build system
   3. Build fails during npm install phase when Cypress tries to download binaries
   {code}
 Actual results:
 {code:none}
   npm error Installing Cypress (version: 12.17.4)
   npm error The Cypress App could not be downloaded.
   npm error URL: https://download.cypress.io/desktop/12.17.4?platform=linux&arch=x64
   npm error Error: getaddrinfo ENOTFOUND download.cypress.io
 
   Build fails because:
   - Yarn package manager not supported by cachi2 for hermetic dependency management
   - Cypress testing framework requires network access to download.cypress.io
   - Custom yarn artifacts and dockerfile modifications conflict with hermetic requirements
   {code}
 Expected results:
 {code:none}
   networking-console-plugin builds successfully in hermetic mode without requiring network access during build process. Component should use npm package manager and disable Cypress binary
   downloads via CYPRESS_INSTALL_BINARY=0 environment variable.
   {code}
 Additional info:
 {code:none}
   Root cause: Component uses yarn package manager which is incompatible with cachi2 hermetic builds, plus Cypress test framework downloads binaries during build.
 
   Solution requires:
   1. Upstream yarn→npm migration (similar to networking-console-plugin)
   2. Set CYPRESS_INSTALL_BINARY=0 to disable test framework downloads
   3. Remove yarn-specific artifacts and dockerfile modifications
   4. Update ocp-build-data package manager configuration
 
   Reference upstream work: https://github.com/openshift/networking-console-plugin/pull/357
 
   Tracking: ART-14633 (parent task with technical analysis)
   {code}
Status: ON_QA
#OCPBUGS-78014issue34 hours ago[openshift-4.12] sriov-cni can't be build hermetically New
Issue 17969335: [openshift-4.12] sriov-cni can't be build hermetically
Description: This is a clone of issue OCPBUGS-78013. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78011. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build sriov-cni hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-43501issue9 days agocoreos-installer iso kargs show broken when using minimal ISO ASSIGNED
Issue 16368909: coreos-installer iso kargs show broken when using minimal ISO
Description: Running the command coreos-installer iso kargs show no longer works in case of platform external.
 {code:none}
 $ coreos-installer iso kargs show agent.x86_64.iso 
 
 Writing manifest to image destination
 Error: kargs don't match at all offsets! (expected 'rw ignition.firstboot ignition.platform.id=metal console=ttyS0 'coreos.live.rootfs_url=https://rhcos.mirror.openshift.com/art/storage/prod/streams/4.17-9.4/builds/417.94.202407010929-0/x86_64/rhcos-417.94.202407010929-0-live-rootfs.x86_64.img'', but offset 112924489 has: '/images/assisted_installer_custom.img rw ignition.firstboot ignition.platform.id=metal console=ttyS0 coreos.live.rootfs_url=https://rhcos.mirror.openshift.com/art/storage/prod/streams/4.17-9.4/builds/417.94.202407010929-0/x86_64/rhcos-417.94.202407010929-0-live-rootfs.x86_64.img'){code}
 Version-Release number of selected component (if applicable):
 {code:none}
 https://github.com/openshift/installer/pull/8886{code}
Status: ASSIGNED
#OCPBUGS-70154issue6 weeks agoiDRAC 10 firmware updates succeeded but with servicing error New
Issue 17714671: iDRAC 10 firmware updates succeeded but with servicing error
Description: Description of problem:
 {code:none}
 Though iDRAC 10 firmware updates succeeded, oc describe shows servicing error. in addition, node did not return to Ready state:
 
 Normal PowerOff 52m metal3-baremetal-controller Host soft powered off
 Normal ServicingError 43m metal3-baremetal-controller Node 68a7d19d-ebc0-42b6-8599-4c36a05a0bd8 failed step {'args': {'settings': [{'component': 'bmc', 'url': 'http://fw.apps.jhajyahy-vv.qe.devcluster.openshift.com/firmimg.d10'}]}, 'interface': 'firmware', 'step': 'update', 'abortable': False, 'priority': 0}: HTTP GET https://10.26.16.132/redfish/v1/ returned code 503. Base.1.18.GeneralError: The service is temporarily unavailable. Retry in 30 seconds. Extended information: [{'MessageId': 'Base.1.18.ServiceTemporarilyUnavailable', 'MessageArgs': ['30'], 'RelatedProperties': [], 'Message': 'The service is temporarily unavailable. Retry in 30 seconds.', 'Severity': 'Critical', 'Resolution': 'Wait for the indicated retry duration and retry the operation.'}]. Also failed to tear down servicing.    {code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1. Perform FW update on IDRAC 10 server using HFC
     2. Check "oc describe bmh  -n openshift-machine-api <host>" command   output    
     3.
     {code}
 Actual results:
 {code:none}
     passed with servicing errors{code}
 Expected results:
 {code:none}
     To pass without servicing errors
  {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-77981issue39 hours ago[openshift-4.13] ose-machine-config-operator can't be built hermetically POST
Issue 17967788: [openshift-4.13] ose-machine-config-operator can't be built hermetically
Description: This is a clone of issue OCPBUGS-77980. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-machine-config-operator hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: POST
#OCPBUGS-78018issue34 hours ago[openshift-4.18] ose-vsphere-cluster-api-controllers can't be built hermetically POST
Issue 17969349: [openshift-4.18] ose-vsphere-cluster-api-controllers can't be built hermetically
Description: This is a clone of issue OCPBUGS-78017. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-vsphere-cluster-api-controllers hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: POST
#OCPBUGS-78182issue9 hours ago[openshift-4.13] ose-machine-api-provider-gcp can't be built hermetically POST
Issue 17974862: [openshift-4.13] ose-machine-api-provider-gcp can't be built hermetically
Description: This is a clone of issue OCPBUGS-78181. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-machine-api-provider-gcp hermetically
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: POST
#OCPBUGS-77213issue3 hours ago[openshift-4.13] hermetic build failure - ose-ibm-vpc-block-csi-driver ON_QA
Issue 17926440: [openshift-4.13] hermetic build failure - ose-ibm-vpc-block-csi-driver
Description: This is a clone of issue OCPBUGS-77212. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77211. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77210. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77209. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77207. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77206. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77205. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77203. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77167. The following is the description of the original issue:  --- Description of problem:
 {code:none}
 IBM VPC Block CSI Driver build fails in hermetic mode due to non-deterministic GOPATH handling in Makefile. The build system rewrites $GOPATH causing the binary output path -o ${GOPATH}/bin/${EXE_DRIVER_NAME} to become invalid, resulting in build failures.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
 OpenShift 4.19+ (affects hermetic build conversion) {code}
 How reproducible:
 {code:none}
 Always - when building ibm-vpc-block-csi-driver in hermetic mode{code}
 Steps to Reproduce:
 {code:none}
   1. Build ibm-vpc-block-csi-driver container in hermetic mode
   2. Makefile uses absolute GOPATH-dependent output path: -o ${GOPATH}/bin/${EXE_DRIVER_NAME}
   3. Build system rewrites $GOPATH during hermetic build process     {code}
 Actual results:
 {code:none}
 Build fails because binary cannot be found at expected GOPATH location after build system modifies $GOPATH
 
 Error: building at STEP "COPY --from=builder /cachi2/output/deps/gomod/bin/ibm-vpc-block-csi-driver /bin/ibm-vpc-block-csi-driver": checking on sources under [...]: copier: stat: "/cachi2/output/deps/gomod/bin/ibm-vpc-block-csi-driver": no such file or directory{code}
 Expected results:
 {code:none}
 Build succeeds using relative path bin/${EXE_DRIVER_NAME} that doesn't depend on GOPATH manipulation, consistent with other CSI drivers like aws-ebs-csi-driver  {code}
 Additional info:
 {code:none}
 Fix implemented: Change Makefile to use relative path -o bin/${EXE_DRIVER_NAME} instead of absolute GOPATH-dependent path. This aligns with other CSI drivers and makes builds hermetic-compatible. Comment added to indicate this is a permanent carry patch for OpenShift builds.{code}
Status: Verified
#OCPBUGS-77467issue11 days ago[openshift-4.14] openshift-enterprise-hyperkube can't be built hermetically New
Issue 17940935: [openshift-4.14] openshift-enterprise-hyperkube can't be built hermetically
Description: Description of problem:
 ART can't build openshift-enterprise-hyperkube hermetically. Same applies for ose-installer-kube-apiserver-artifacts
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.
      {code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-77979issue39 hours ago[openshift-4.17] ose-machine-config-operator can't be built hermetically POST
Issue 17967785: [openshift-4.17] ose-machine-config-operator can't be built hermetically
Description: This is a clone of issue OCPBUGS-77978. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-machine-config-operator hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: POST
#OCPBUGS-74133issue6 weeks ago[4.19] oauth-proxy e2e tests fail due to ambiguous container image short names on RHEL 9 POST
Issue 17781204: [4.19] oauth-proxy e2e tests fail due to ambiguous container image short names on RHEL 9
Description: This is a clone of issue OCPBUGS-73848. The following is the description of the original issue: 
 —
 The oauth-proxy e2e-component tests fail on OCP 4.19 because RHEL 9 enforces fully qualified container image names ("short name mode").
 
 The backend test image {{nginxdemos/nginx-hello:plain-text}} is rejected by Podman since it cannot determine which registry to pull from.
 h2. Version-Release number
 
 4.22
 h2. How reproducible
 
 Always
 h2. Steps to Reproduce
 
 Run the oauth-proxy e2e-component tests on OCP 4.22
 
 The test creates a pod with a backend container using image {{nginxdemos/nginx-hello:plain-text}}
 
 Observe the pod fails to start.
 h2. Actual results
 
 The backend container fails with:
 {code:java}
 ImageInspectError: short name mode is enforcing, but image name 
 nginxdemos/nginx-hello:plain-text returns ambiguous list
 {code}
 The pod never reaches Running state, and all e2e tests fail.
 h2. Expected results
 
 The e2e tests should pass. The backend container should start successfully using a fully qualified image reference.
 h2. Additional info
 
 _Root cause:_ RHEL 9 enforces "short name mode" for container images, requiring fully qualified registry paths.
 
 _Fix:_ Replace {{nginxdemos/nginx-hello:plain-text}} with {{{}registry.k8s.io/e2e-test-images/agnhost:2.53{}}}, which is:
  * Fully qualified (satisfies RHEL 9 short name policy)
  * Multi-arch (supports ARM64)
  * The standard test image adopted across OpenShift repositories ([openshift/origin PR #29672|https://github.com/openshift/origin/pull/29672])
  * Mirrored in OpenShift CI infrastructure
 
 _PR:_ [https://github.com/openshift/oauth-proxy/pull/343]
Status: POST
#OCPBUGS-73851issue7 weeks agooauth-proxy e2e tests fail due to ambiguous container image short names on RHEL 9 New
Issue 17769459: oauth-proxy e2e tests fail due to ambiguous container image short names on RHEL 9
Description: The oauth-proxy e2e-component tests fail on OCP 4.19 because RHEL 9 enforces fully qualified container image names ("short name mode"). The backend test image {{nginxdemos/nginx-hello:plain-text}} is rejected by Podman since it cannot determine which registry to pull from.  h2. Version-Release number  4.19  h2. How reproducible  Always  h2. Steps to Reproduce  h1. Run the oauth-proxy e2e-component tests on OCP 4.19 h1. The test creates a pod with a backend container using image {{nginxdemos/nginx-hello:plain-text}} h1. Observe the pod fails to start  h2. Actual results  The backend container fails with:  {code} ImageInspectError: short name mode is enforcing, but image name  nginxdemos/nginx-hello:plain-text returns ambiguous list {code}  The pod never reaches Running state, and all e2e tests fail.  h2. Expected results  The e2e tests should pass. The backend container should start successfully using a fully qualified image reference.  h2. Additional info  _Root cause:_ RHEL 9 enforces "short name mode" for container images, requiring fully qualified registry paths.  _Fix:_ Replace {{nginxdemos/nginx-hello:plain-text}} with {{registry.k8s.io/e2e-test-images/agnhost:2.53}}.  _Backport of:_ [OCPBUGS-73848|https://issues.redhat.com/browse/OCPBUGS-73848]
Status: New
#OCPBUGS-77862issue5 days ago[openshift-4.17] openshift-enterprise-console-operator can't be built hermetically New
Issue 17959746: [openshift-4.17] openshift-enterprise-console-operator can't be built hermetically
Description: This is a clone of issue OCPBUGS-77861. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-77860. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build openshift-enterprise-console-operator hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-74694issue9 hours agoPods stuck terminating on cri-o unmounting overlayfs /merged path New
Issue 17820540: Pods stuck terminating on cri-o unmounting overlayfs /merged path
Description: Description of problem:
 {code:none}
 Customer encounters pods stuck terminating with the following error. When listing mounts on the host the mount point does NOT show as mounted and no processes indicate using the path. However /proc/self/mountinfo does list the mountpoint. When we strace cri-o it's receiving -1EINVAL on the umount2() call to the relevant path.
 
 Jan 24 06:51:04 worker crio[4437]: time="2026-01-24 06:51:04.487574617Z" level=warning msg="Failed to unmount container 9c634a49e087b3b7f6c48c8bf7fd6050d328bd5e985bd48e460de5743f98d78f: replacing mount point \"/var/lib/containers/storage/overlay/804f846ce959cc4fffeb930421231dc2298146fa8721732101b9870aee1217e5/merged\": device or resource busy" id=863adbef-0903-4953-a16c-d9018a970671 name=/runtime.v1.RuntimeService/StopPodSandbox
 Jan 24 06:51:04 worker kubenswrapper[4476]: E0124 06:51:04.487776    4476 log.go:32] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to stop infra container for pod sandbox 9c634a49e087b3b7f6c48c8bf7fd6050d328bd5e985bd48e460de5743f98d78f: failed to unmount container 9c634a49e087b3b7f6c48c8bf7fd6050d328bd5e985bd48e460de5743f98d78f: replacing mount point \"/var/lib/containers/storage/overlay/804f846ce959cc4fffeb930421231dc2298146fa8721732101b9870aee1217e5/merged\": device or resource busy" podSandboxID="9c634a49e087b3b7f6c48c8bf7fd6050d328bd5e985bd48e460de5743f98d78f"
 Jan 24 06:51:04 worker kubenswrapper[4476]: E0124 06:51:04.487830    4476 kuberuntime_manager.go:1479] "Failed to stop sandbox" podSandboxID={"Type":"cri-o","ID":"9c634a49e087b3b7f6c48c8bf7fd6050d328bd5e985bd48e460de5743f98d78f"}
 Jan 24 06:51:04 worker kubenswrapper[4476]: E0124 06:51:04.487877    4476 kubelet.go:2041] "Unhandled Error" err="failed to \"KillPodSandbox\" for \"0f025850-3739-43c8-9e28-74c4bd792148\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to stop infra container for pod sandbox 9c634a49e087b3b7f6c48c8bf7fd6050d328bd5e985bd48e460de5743f98d78f: failed to unmount container 9c634a49e087b3b7f6c48c8bf7fd6050d328bd5e985bd48e460de5743f98d78f: replacing mount point \\\"/var/lib/containers/storage/overlay/804f846ce959cc4fffeb930421231dc2298146fa8721732101b9870aee1217e5/merged\\\": device or resource busy\"" logger="UnhandledError"
     {code}
 Version-Release number of selected component (if applicable):
 {code:none}
 OpenShift 4.18.24
 cri-o-0-1.31.12-3.rhaos4.18.gitdc59c78
 github.com/containers/storage v1.55.1-0.20250328180621-028e88d2378c
 kernel-0-5.14.0-427.87.1.el9_4-x86_64    {code}
 How reproducible:
 {code:none}
 Daily to weekly on two clusters in a loaded cluster with stackrox collectors and netobservability enabled. Sometimes happens in nodes that are only days old (elastic cloud environment).{code}
 Steps to Reproduce:
 {code:none}
 Unknown    {code}
 Actual results:
 {code:none}
 Pods stuck in terminating state{code}
 Expected results:
 {code:none}
 Pods terminate and cleanup    {code}
 Additional info:
 {code:none}
 This is a cluster running Stackrox 4.8.4 and Network Observability, both of which I understand to exert additional pressure on the kernel mount house keeping. However, the customer has plenty of other clusters with those components that don't encounter this problem so this may be not be valuable debugging information.
 
 Container storage has to be reset and the node rebooted to clean things up. Often they just destroy the node or a machine health check does so.
 
 
 The following kernel warnings are seen in the logs but we believe they're associated with debugging efforts rather than root cause. They happen around the time a debug toolbox container is created, additionally they do not happen in all nodes where the problem is observed. No other kernel messages are logged that seem related to this.
 
 Jan 24 07:49:34 worker kernel: overlayfs: upperdir is in-use as upperdir/workdir of another mount, accessing files from both mounts will result in undefined behavior.
 Jan 24 07:49:34 worker kernel: overlayfs: workdir is in-use as upperdir/workdir of another mount, accessing files from both mounts will result in undefined behavior.{code}
Status: New
#OCPBUGS-78105issue17 hours ago[openshift-4.14] cluster-network-operator can't be built hermetically POST
Issue 17972270: [openshift-4.14] cluster-network-operator can't be built hermetically
Description: This is a clone of issue OCPBUGS-78104. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build cluster-network-operator hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
  
 
 Version-Release number of selected component (if applicable):
 
 How reproducible:
 
 Steps to Reproduce:
 
 1.
 
 2.
 
 3.
 
 Actual results:
 
 Expected results:
 
 Additional info:
 
 Please fill in the following template while reporting a bug and provide as much relevant information as possible. Doing so will give us the best chance to find a prompt resolution.
 
 Affected Platforms:
 
 Is it an
  # internal CI failure
  # customer issue / SD
  # internal RedHat testing failure
 
 If it is an internal RedHat testing failure:
  * Please share a kubeconfig or creds to a live cluster for the assignee to debug/troubleshoot along with reproducer steps (specially if it's a telco use case like ICNI, secondary bridges or BM+kubevirt).
 
 If it is a CI failure:
  * Did it happen in different CI lanes? If so please provide links to multiple failures with the same error instance
  * Did it happen in both sdn and ovn jobs? If so please provide links to multiple failures with the same error instance
  * Did it happen in other platforms (e.g. aws, azure, gcp, baremetal etc) ? If so please provide links to multiple failures with the same error instance
  * When did the failure start happening? Please provide the UTC timestamp of the networking outage window from a sample failure run
  * If it's a connectivity issue,
  * What is the srcNode, srcIP and srcNamespace and srcPodName?
  * What is the dstNode, dstIP and dstNamespace and dstPodName?
  * What is the traffic path? (examples: pod2pod? pod2external?, pod2svc? pod2Node? etc)
 
 If it is a customer / SD issue:
  * Provide enough information in the bug description that Engineering doesn’t need to read the entire case history.
  * Don’t presume that Engineering has access to Salesforce.
  * Do presume that Engineering will access attachments through supportshell.
  * Describe what each relevant attachment is intended to demonstrate (failed pods, log errors, OVS issues, etc).
  * Referring to the attached must-gather, sosreport or other attachment, please provide the following details:
  ** If the issue is in a customer namespace then provide a namespace inspect.
  ** If it is a connectivity issue:
  *** What is the srcNode, srcNamespace, srcPodName and srcPodIP?
  *** What is the dstNode, dstNamespace, dstPodName and dstPodIP?
  *** What is the traffic path? (examples: pod2pod? pod2external?, pod2svc? pod2Node? etc)
  *** Please provide the UTC timestamp networking outage window from must-gather
  *** Please provide tcpdump pcaps taken during the outage filtered based on the above provided src/dst IPs
  ** If it is not a connectivity issue:
  *** Describe the steps taken so far to analyze the logs from networking components (cluster-network-operator, OVNK, SDN, openvswitch, ovs-configure etc) and the actual component where the issue was seen based on the attached must-gather. Please attach snippets of relevant logs around the window when problem has happened if any.
 
  * When showing the results from commands, include the entire command in the output.  
  * For OCPBUGS in which the issue has been identified, label with “sbr-triaged”
  * For OCPBUGS in which the issue has not been identified and needs Engineering help for root cause, label with “sbr-untriaged”
  * Do not set the priority, that is owned by Engineering and will be set when the bug is evaluated
  * Note: bugs that do not meet these minimum standards will be closed with label “SDN-Jira-template”
  * For guidance on using this template please see
 [OCPBUGS Template Training for Networking  components|https://source.redhat.com/groups/public/gss-openshift/cee_openshift_wiki/sbr_shift_networking#jive_content_id_Training_Materials]
Status: POST
#OCPBUGS-74134issue6 weeks ago[4.18] oauth-proxy e2e tests fail due to ambiguous container image short names on RHEL 9 New
Issue 17781205: [4.18] oauth-proxy e2e tests fail due to ambiguous container image short names on RHEL 9
Description: This is a clone of issue OCPBUGS-73848. The following is the description of the original issue: 
 —
 The oauth-proxy e2e-component tests fail on OCP 4.18 because RHEL 9 enforces fully qualified container image names ("short name mode").
 
 The backend test image {{nginxdemos/nginx-hello:plain-text}} is rejected by Podman since it cannot determine which registry to pull from.
 h2. Version-Release number
 
 4.22
 h2. How reproducible
 
 Always
 h2. Steps to Reproduce
 
 Run the oauth-proxy e2e-component tests on OCP 4.22
 
 The test creates a pod with a backend container using image {{nginxdemos/nginx-hello:plain-text}}
 
 Observe the pod fails to start.
 h2. Actual results
 
 The backend container fails with:
 {code:java}
 ImageInspectError: short name mode is enforcing, but image name 
 nginxdemos/nginx-hello:plain-text returns ambiguous list
 {code}
 The pod never reaches Running state, and all e2e tests fail.
 h2. Expected results
 
 The e2e tests should pass. The backend container should start successfully using a fully qualified image reference.
 h2. Additional info
 
 _Root cause:_ RHEL 9 enforces "short name mode" for container images, requiring fully qualified registry paths.
 
 _Fix:_ Replace {{nginxdemos/nginx-hello:plain-text}} with {{{}registry.k8s.io/e2e-test-images/agnhost:2.53{}}}, which is:
  * Fully qualified (satisfies RHEL 9 short name policy)
  * Multi-arch (supports ARM64)
  * The standard test image adopted across OpenShift repositories ([openshift/origin PR #29672|https://github.com/openshift/origin/pull/29672])
  * Mirrored in OpenShift CI infrastructure
 
 _PR:_ [https://github.com/openshift/oauth-proxy/pull/343]
Status: New
#OCPBUGS-66129issue5 weeks agoThe gather job restart loop when upload incorrect password Verified
then
echo "Error: Required Parameters have not been provided. Exiting..."
exit 1
#OCPBUGS-77948issue4 days agoTNF IPv6 Node Replacement Test misformats URL ASSIGNED
Issue 17965686: TNF IPv6 Node Replacement Test misformats URL
Description: *Description of problem:*
 {code:none}
 The IPv6 recovery test fails node replacement because the IP url used for redfish doesn't put the IPv6 address in brackets.
 {code}
 
 *Version-Release number of selected component (if applicable):*
 {code:none}
 4.22 nightlies 3/6/26
 {code}
 
 *How reproducible:*
 {code:none}
 See IPv6 recovery test
 {code}
 
 *Steps to Reproduce:*
 {code:none}
     1. Look at IPv6 recovery test in CI
     2. Check results of node replacement test
 {code}
 
 *Actual results:*
 {code:none}
 I0306 09:43:38.935142 19613 tnf_node_replacement.go:1489] BareMetalHost ostest-master-1 error message: Failed to validate power driver interface for node f5db7f08-753f-4ad9-87fa-82682bc2750d. Error: Invalid Redfish address https://fd2e:6f44:5dd8:c956::1:8000 set in driver_info/redfish_address on node f5db7f08-753f-4ad9-87fa-82682bc2750d
 {code}
 
 *Expected results:*
 {code:none}
 Test passes
 {code}
 
 Additional info:
 {code:none}
 {code}
Status: ASSIGNED
#OCPBUGS-77458issue5 days agoIn rhel 10 podman inside nodes cannot reach outside the cluster New
Issue 17940305: In rhel 10 podman inside nodes cannot reach outside the cluster
Description: Description of problem:
 {code:none}
     When in a rhel10 clutser we connect to a node, and we try to use podman inside the node, podman cannot reach the network outside the cluster.
 
     If we do it in a rhel9 node there is no problem doing that.
 
 {code}
 Version-Release number of selected component (if applicable):
 {code:none}
 4.22    {code}
 How reproducible:
 {code:none}
     Always{code}
 Steps to Reproduce:
 {code:none}
     1. Enable the rhel10 stream in a 4.22 techpreview cluster
 
  oc patch mcp worker --type=merge -p "{\"spec\":{\"osImageStream\":{\"name\":\"rhel-10\"}}}"
 
 
 
     2. Wait until the worker pool is using rhel10
 
     3. Debug into a worker node with "oc debug...."
 
     4. Create a Containerfile with this content
 
 sh-5.2# cat Containerfile 
 FROM quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4a1798a3b92a794a69d56eaf78c1521a1c4d2e52fd05057072780ec19ccabd45
 RUN curl -LO https://pkgs.tailscale.com/stable/fedora/tailscale.repo
 
 
     5. Build an image using the container file 
 
 
      sh-5.2# podman build  --authfile /var/lib/kubelet/config.json  . -t with9 
 The result is this error
 STEP 2/2: RUN curl -LO https://pkgs.tailscale.com/stable/fedora/tailscale.repo
   % Total    % Received % Xferd  Average Speed   Time    Time     Time  Current
                                  Dload  Upload   Total   Spent    Left  Speed
   0     0    0     0    0     0      0      0 --:--:--  0:00:19 --:--:--     0curl: (6) Could not resolve host: pkgs.tailscale.com
 Error: building at STEP "RUN curl -LO https://pkgs.tailscale.com/stable/fedora/tailscale.repo": while running runtime: exit status 6{code}
 Actual results:
 {code:none}
 The podman build command cannot reach the network outside the cluster    {code}
 Expected results:
 {code:none}
 The podman build command should be able to get the tailscale information and build the image{code}
 Additional info:
 {code:none}
     
 
 This is the AI analysis:
 
 Root Cause
 
   The problem is a conflict between two FORWARD chains combined with a firewall backend change between netavark versions on RHEL 9 vs RHEL 10.
 
   RHEL 9 (works) — netavark 1.17.2 with iptables-nft backend
 
   There is one FORWARD chain in table ip filter with policy DROP, and netavark injects its rules directly into it:
 
   chain FORWARD {
       policy drop;
       jump NETAVARK_FORWARD   <-- netavark adds this to allow container traffic
       <OVN rules>
   }
 
   Container traffic from 10.88.0.0/16 hits NETAVARK_FORWARD first and gets accepted. Packets are forwarded, masqueraded, and reach the internet.
 
   RHEL 10 (fails) — netavark 1.16.0 with native nftables backend
 
   There are two FORWARD chains at the same priority:
 
   1. table ip filter → chain FORWARD (from OVN/kube — policy DROP):
   chain FORWARD {
       policy drop;
       <only OVN/kube subnets accepted>
       # NO netavark rules here!
   }
   2. table inet netavark → chain FORWARD (from netavark — policy ACCEPT):
   chain FORWARD {
       policy accept;
       ct state invalid drop
       jump NETAVARK-ISOLATION-1
   }
 
   The OVN table ip filter FORWARD chain has policy DROP and only allows traffic to/from 169.254.0.1, 172.30.0.0/16, and 10.128.0.0/14. The podman bridge subnet 10.88.0.0/16 is not in that list. Even though
   netavark's own chain would accept it, the OVN/kube chain drops it first because both chains are at the same filter hook priority, and the ip family table processes before inet.
 
   So container traffic from 10.88.0.0/16 → 10.0.0.2 (DNS) gets dropped by the OVN FORWARD chain before it can reach the internet.
 
   Why --network host works
 
   With --network host, containers use the host's network stack directly — no bridge, no forwarding, no FORWARD chain involved.
 ~                                                                                                                                    
 
 
 
 
 
 
 Workaround
 
 Executing the "podman build" command with "--network host" will make the command work.{code}
Status: New
#OCPBUGS-78100issue17 hours ago[openshift-4.15] baremetal-machine-controller can't be built hermetically New
Issue 17972261: [openshift-4.15] baremetal-machine-controller can't be built hermetically
Description: This is a clone of issue OCPBUGS-78099. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78098. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build baremetal-machine-controller hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-76931issue3 weeks agoGCP: Provider produced inconsistent result after apply - google_service_account ASSIGNED
Issue 17884800: GCP: Provider produced inconsistent result after apply - google_service_account
Description: Description of problem:
 {code:none}
     The fix given for issue https://issues.redhat.com/browse/OCPBUGS-42014(pointing to same problem) was at service account key creation. but the actual problem lies at service account creation itself.  
 
 We suggest upgrading terraform and provider to the latest version in openshift-installer for 4.12+. The underlying issue is already fixed upstream and released in v6.45.0.
 
 Note: we see this problem more often in lower openshift versions 4.12, 4.13 than in 4.16.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
 Intermittent{code}
 Steps to Reproduce:
 {code:none}
     1.Install the UPI openshift cluster in GCP platform
     2.
     3.
     {code}
 Actual results:
 {code:none}
     level=error msg=Error: Provider produced inconsistent result after apply
 level=error
 level=error msg=When applying changes to
 level=error msg=module.master.google_service_account.master-node-sa[0], provider
 level=error msg="provider[\"openshift/local/google\"]" produced an unexpected new value: Root
 level=error msg=resource was present, but now absent.
 level=error
 level=error msg=This is a bug in the provider, which should be reported in the provider's own{code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: ASSIGNED
#OCPBUGS-59259issue5 weeks agoFirmware upgrade failures get cleared from the BaremetalHost making it difficult to know what failed ASSIGNED
Issue 17137858: Firmware upgrade failures get cleared from the BaremetalHost making it difficult to know what failed
Description: Description of problem:
 
 I have the following BMH definition:
 {code:none}
 apiVersion: metal3.io/v1alpha1
 kind: BareMetalHost
 metadata:
   name: dell-server
   namespace: hardware-inventory
   annotations:
 spec:
   automatedCleaningMode: disabled
   bmc:
     disableCertificateVerification: True
     address: idrac-virtualmedia://10.6.75.116/redfish/v1/Systems/System.Embedded.1
     credentialsName: dell-bmc-credentials
   bootMACAddress: B4:83:51:00:B4:88
   online: true{code}
 And I apply the following manifest to run the upgrade:
 {code:java}
 ---
 apiVersion: metal3.io/v1alpha1
 kind: HostUpdatePolicy
 metadata:
   name: dell-server
   namespace: hardware-inventory
 spec:
   firmwareSettings: onReboot
   firmwareUpdates: onReboot
 ---
 apiVersion: metal3.io/v1alpha1
 kind: HostFirmwareComponents
 metadata:
   name: dell-server
   namespace: hardware-inventory
 spec:
   updates:
     - component: bmc
       url: http://10.6.116.4:9999/iDRAC-with-Lifecycle-Controller_Firmware_R8V2F_LN64_7.20.30.00_A00.BIN
     - component: bios
       url: http://10.6.116.4:9999/BIOS_0HY8N_LN64_1.17.2.BIN {code}
 Then I apply the _reboot.metal3.io=""_ annotation to get the firmware and bios upgraded. The problem I see is that this upgrade fails, this is what I see in the BMO logs:
 {code:java}
 {"level":"info","ts":1750857141.5618482,"logger":"controllers.BareMetalHost","msg":"using PreprovisioningImage","baremetalhost":{"name":"dell-server","namespace":"hardware-inventory"},"provisioningState":"available","Image":{"ImageURL":"http://metal3-image-customization-service.openshift-machine-api.svc.cluster.local/d235e339-76f8-4709-ad13-3517db75f539","KernelURL":"","ExtraKernelParams":"","Format":"iso"}}
 {"level":"info","ts":1750857141.5847816,"logger":"provisioner.ironic","msg":"current provision state","host":"hardware-inventory~dell-server","lastError":"Firmware update failed for node c0471e4f-e9c8-4678-9893-77975de4ded2, firmware http://10.6.116.4:9999/iDRAC-with-Lifecycle-Controller_Firmware_R8V2F_LN64_7.20.30.00_A00.BIN. Error: Lifecycle Controller in use. This job will start when Lifecycle Controller is available.","current":"manageable","target":""} {code}
 Now, that's okay. The issue I see is that this error will be cleared from the BMH, this is the BMH during the upgrade:
 {code:java}
 NAME          STATE       CONSUMER   ONLINE   ERROR   AGE
 dell-server   preparing              true             46h
 dell-server   preparing              true     preparation error   46h
 dell-server   preparing              true     preparation error   46h
 dell-server   available              true                         46h {code}
  
 
  
 
 Version-Release number of selected component (if applicable):
 {code:none}
 OCP 4.18    {code}
 How reproducible:
 {code:none}
 Always{code}
 Steps to Reproduce:
 {code:none}
 Described above.{code}
 Actual results:
 {code:none}
 Firmware upgrade fails. Error gets cleared out from the BMH. User cannot see why it failed unless watching the object during the procedure.{code}
 Expected results:
 {code:none}
 Firmware upgrade fails. Error remains in the BMH until the user fixes whatever needs to be fixed.{code}
 Additional info:
 {code:none}
 Slack thread: https://redhat-internal.slack.com/archives/CFP6ST0A3/p1750857275820059{code}
Status: ASSIGNED
#OCPBUGS-77861issue5 days ago[openshift-4.19] openshift-enterprise-console-operator can't be built hermetically POST
Issue 17959742: [openshift-4.18] openshift-enterprise-console-operator can't be built hermetically
Description: This is a clone of issue OCPBUGS-77860. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build openshift-enterprise-console-operator hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: POST
#OCPBUGS-77466issue11 days ago[openshift-4.15] openshift-enterprise-hyperkube can't be built hermetically New
Issue 17940923: [openshift-4.15] openshift-enterprise-hyperkube can't be built hermetically
Description: Description of problem:
 ART can't build openshift-enterprise-hyperkube hermetically. Same applies for ose-installer-kube-apiserver-artifacts
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.
      {code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-77775issue6 days ago[openshift-4.19] ose-sriov-rdma-cni can't be built hermetically POST
Issue 17954315: [openshift-4.19] ose-sriov-rdma-cni can't be built hermetically
Description: Description of problem:
 
 ART can't build ose-sriov-rdma-cni hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: Verified
#OCPBUGS-77199issue2 weeks ago[openshift-4.13] ose-libvirt-machine-controllers can't be build hermetically New
Issue 17925866: [openshift-4.13] ose-libvirt-machine-controllers can't be build hermetically
Description: Description of problem:
 
 ART can't build ose-libvirt-machine-controllers hermetically.
 The error is the following in the prefetch-dependencies konflux task:
 {code:none}
     Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.  Please try running `go mod vendor` and committing the changes.  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
  
 
  
 
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-45017issue7 weeks agoOpenShift 4.17 IPI Private cluster installation failure due to non-empty resource group ASSIGNED
Issue 16485170: OpenShift 4.17 IPI Private cluster installation failure due to non-empty resource group
Description: Description of problem:
 
 When creating a Private IPI Cluster using an existing Azure Vnet with preconfigured Subnets, as per
 
 [https://docs.openshift.com/container-platform/4.17/installing/installing_azure/ipi/installing-azure-private.html]
 
 and
 
 [https://docs.openshift.com/container-platform/4.17/installing/installing_azure/ipi/installing-azure-vnet.html]
 
 an error is returned:
 {code:none}
 FATAL failed to fetch Cluster API Manifests: failed to generate asset "Cluster API Manifests": failed to generate Azure manifests: failed to get azure ip availability: network.VirtualNetworksClient#CheckIPAddressAvailability: Failure responding to request: StatusCode=400 -- Original Error: autorest/azure: Service returned an error. Status=400 Code="PrivateIPAddressNotInAnySubnet" Message="Private static IP address 10.0.0.100 does not belong to the range of any subnet in the virtual network /subscriptions/6e700e21-5667-435d-8f78-f421bedbe936/resourceGroups/RG-AS-OCPNP/providers/Microsoft.Network/virtualNetworks/VN-AS-OCPNP." Details=[]{code}
 This error appears during the creation of the manifests, i.e.
 {code:java}
 $ openshift-install create manifests --dir <installation_directory> {code}
 No cluster creation step was attempted yet.
 
 For the above issue,  the customer found the "fix" in the GitHub Repo[1] and it worked partially:
 
 [1] [https://github.com/openshift/installer/pull/9144]
 
 However, after cloning the git repo, adding the PR changes to that file; "pkg/asset/manifests/azure/cluster.go"
 {code:java}
 # ./openshift-install create manifests --dir installer_linux INFO Credentials loaded from file "/root/.azure/osServicePrincipal.json" INFO Consuming Install Config from target directory INFO Adding clusters... INFO Manifests created in: installer_linux/cluster-api, installer_linux/manifests and installer_linux/openshift{code}
 We could see the internal LB was assigned the correct next available IP(which previously was not happening)
 {code:java}
   networkSpec:
     apiServerLB:
       backendPool:
         name: ocpnp-pz6m2-internal
       frontendIPs:
       - name: ocpnp-pz6m2-internal-frontEnd
         privateIP: 10.89.188.4
       name: ocpnp-pz6m2-internal
       type: Internal {code}
 We used  " remotes/origin/release-4.17 " branch, from [https://github.com/openshift/installer]
 
 And then proceeded with a cluster build to test it further, but testing the cluster creation failed, this time with a new error message:
 {code:java}
 # ./openshift-install create cluster --dir installer_linux
 INFO Consuming OpenShift Install (Manifests) from target directory
 INFO Consuming Master Machines from target directory
 INFO Consuming Worker Machines from target directory
 INFO Consuming Common Manifests from target directory
 INFO Consuming Openshift Manifests from target directory
 INFO Credentials loaded from file "/root/.azure/osServicePrincipal.json"
 FATAL failed to fetch Cluster Infrastructure Variables: failed to fetch dependency of "Cluster Infrastructure Variables": failed to generate asset "Platform Provisioning Check": platform.azure.resourceGroupName: Invalid value: "RG-AS-OCPNP": resource group must be empty but it has 7 resources like /subscriptions/6e700e21-5667-435d-8f78-f421bedbe936/resourceGroups/RG-AS-OCPNP/providers/Microsoft.Network/virtualNetworks/VN-AS-OCPNP, /subscriptions/6e700e21-5667-435d-8f78-f421bedbe936/resourceGroups/RG-AS-OCPNP/providers/Microsoft.Network/networkSecurityGroups/NSG-AS-OCPNP ...{code}
 For some reason, when using an Existing Vnet / resource group etc, it's trying to validate the resource group is empty, which will never be the case.
 
 This above error appeared during v4.17 cluster installation and also v4.16.x (tried .0, .9 and .21 installer versions) wherein creating the manifests worked fine, but the error about non-empty resource group persists.
 {code:java}
 root@ocpjump:~# ./openshift-install version
 ./openshift-install 4.16.9
 built from commit deb993e72a920fb1c68a578d1b4c598071fefea4
 release image quay.io/openshift-release-dev/ocp-release@sha256:115bba6836b9feffb81ad9101791619edd5f19d333580b7f62bd6721eeda82d2
 release architecture amd64
 root@ocpjump:~# ./openshift-install create manifests --dir installer_linux
 INFO Credentials loaded from file "/root/.azure/osServicePrincipal.json"
 INFO Consuming Install Config from target directory
 INFO Manifests created in: installer_linux/manifests and installer_linux/openshift
 root@ocpjump:~# ./openshift-install create cluster --dir installer_linux
 INFO Consuming Common Manifests from target directory
 INFO Consuming Master Machines from target directory
 INFO Consuming Worker Machines from target directory
 INFO Consuming OpenShift Install (Manifests) from target directory
 INFO Consuming Openshift Manifests from target directory
 INFO Credentials loaded from file "/root/.azure/osServicePrincipal.json"
 FATAL failed to fetch Cluster Infrastructure Variables: failed to fetch dependency of "Cluster Infrastructure Variables": failed to generate asset "Platform Provisioning Check": platform.azure.resourceGroupName: Invalid value: "RG-AS-OCPNP": resource group must be empty but it has 7 resources like /subscriptions/6e700e21-5667-435d-8f78-f421bedbe936/resourceGroups/RG-AS-OCPNP/providers/Microsoft.Network/virtualNetworks/VN-AS-OCPNP, /subscriptions/6e700e21-5667-435d-8f78-f421bedbe936/resourceGroups/RG-AS-OCPNP/providers/Microsoft.Network/networkSecurityGroups/NSG-AS-OCPNP ...{code}
 Perhaps existing vnet and subnets in a private install isn't a very widely used deployment mechanism in Azure given this bug seems present for a while now. This seems peculiar as this deployment type exhibits the best security and customisation with on-premise cloud interconnects.
 
 Version-Release number of selected component (if applicable):
 {code:none}
 v4.17.Z {code}
 Actual results:
 {code:none}
     The private IPI cluster installation fails for OpenShift v4.17 on Azure{code}
 Expected results:
 {code:none}
     Private IPI OpenShift v4.17 cluster should be installed successfully on Azure{code}
Status: ASSIGNED
#OCPBUGS-78103issue17 hours ago[openshift-4.12] baremetal-machine-controller can't be built hermetically New
Issue 17972334: [openshift-4.12] baremetal-machine-controller can't be built hermetically
Description: This is a clone of issue OCPBUGS-78102. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78101. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78100. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78099. The following is the description of the original issue:  --- This is a clone of issue OCPBUGS-78098. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build baremetal-machine-controller hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-77469issue11 days ago[openshift-4.12] openshift-enterprise-hyperkube can't be built hermetically New
Issue 17940938: [openshift-4.12] openshift-enterprise-hyperkube can't be built hermetically
Description: Description of problem:
 ART can't build openshift-enterprise-hyperkube hermetically. Same applies for ose-installer-kube-apiserver-artifacts
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.
      {code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-77023issue2 weeks agoInspection error on Dell XR8620t New
Issue 17899514: Inspection error on Dell XR8620t
Description: Description of problem:
 {code:none}
     After deprovisioning cluster and deleting BMH for my xr8620t, I recreated the BMH and hit an inspection failure.
 
   errorMessage: 'Failed to inspect hardware. Reason: unable to start inspection: HTTP
     POST https://10.x.x.x/redfish/v1/Systems/System.Embedded.1/VirtualMedia/1/Actions/VirtualMedia.InsertMedia
     returned code 500. Base.1.12.GeneralError: A general error has occurred. See ExtendedInfo
     for more information Extended information: [{''Message'': ''Virtual Media is detached
     or Virtual Media devices are already in use.'', ''MessageArgs'': [], ''MessageArgs@odata.count'':
     0, ''MessageId'': ''IDRAC.2.9.VRM0021'', ''RelatedProperties'': [], ''RelatedProperties@odata.count'':
     0, ''Resolution'': ''Change the Virtual Media Attach Mode to Attach or Auto-Attach.Stop
     existing Virtual Media sessions and retry the operation.'', ''Severity'': ''Informational''}]'
   errorType: inspection error
 
 I've confirmed that there is no virtual media attached. Setting online=true has no effect, and manually powering on the node from idrac ends up with the node quickly being powered off again.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
 OCP: 4.20.0-0.nightly-2025-07-20-021531
 ACM: 2.14.0-131
 MCE: 2.9.0-102{code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: New
#OCPBUGS-77903issue4 days ago[openshift-4.18] ose-aws-cluster-api-controllers can't be built hermetically POST
Issue 17962668: [openshift-4.18] ose-aws-cluster-api-controllers can't be built hermetically
Description: This is a clone of issue OCPBUGS-77902. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-aws-cluster-api-controllers
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.
      {code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: POST
#OCPBUGS-78169issue9 hours ago[openshift-4.17] ose-machine-api-provider-openstack can't be built hermetically POST
Issue 17974764: [openshift-4.17] ose-machine-api-provider-openstack can't be built hermetically
Description: This is a clone of issue OCPBUGS-78168. The following is the description of the original issue:  --- Description of problem:
 
 ART can't build ose-machine-api-provider-openstack hermetically. 
 
 the error is the following
 {code:none}
 Error: PackageRejected: The content of the vendor directory is not consistent with go.mod. Please check the logs for more details.
  Please try running `go mod vendor` and committing the changes.
  Note that you may need to `git add --force` ignored files in the vendor/ dir.{code}
 Version-Release number of selected component (if applicable):
 {code:none}
     {code}
 How reproducible:
 {code:none}
     {code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     {code}
 Expected results:
 {code:none}
     {code}
 Additional info:
 {code:none}
     {code}
Status: POST
#OCPBUGS-60539issue5 weeks agonode looses network connection when it reboots. If the system is rebooted into emergency mode (single user) and boot is resumed, the systems boots normally New
Aug 01 04:39:59 openshift-master-0.utin-ocp-web.caas.apps.mil kubenswrapper[5286]: E0801 04:39:59.414124    5286 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/openshift-storage.cephfs.csi.ceph.com^0001-0011-openshift-storage-0000000000000001-fa12c8a7-1f30-11ed-b784-0a58ac1c0a0b podName:e4b533cc-8360-4861-86ae-8fe736c76db4 nodeName:}" failed. No retries permitted until 2025-08-01 04:42:01.414103867 +0000 UTC m=+52734.450871152 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-e5ce62d9-b588-4072-8ea9-d71a6db7aef6" (UniqueName: "kubernetes.io/csi/openshift-storage.cephfs.csi.ceph.com^0001-0011-openshift-storage-0000000000000001-fa12c8a7-1f30-11ed-b784-0a58ac1c0a0b") pod "e4b533cc-8360-4861-86ae-8fe736c76db4" (UID: "e4b533cc-8360-4861-86ae-8fe736c76db4") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name openshift-storage.cephfs.csi.ceph.com not found in the list of registered CSI drivers
>>>9 missing seconds

Found in 0.00% of runs (0.00% of failures) across 240107 total runs and 19723 jobs (30.02% failed) in 122ms - clear search | chart view - source code located on github