--- apiVersion: v1 items: - apiVersion: v1 kind: Pod metadata: annotations: k8s.ovn.org/pod-networks: '{"default":{"ip_addresses":["10.134.0.24/23"],"mac_address":"0a:58:0a:86:00:18","gateway_ips":["10.134.0.1"],"routes":[{"dest":"10.132.0.0/14","nextHop":"10.134.0.1"},{"dest":"172.31.0.0/16","nextHop":"10.134.0.1"},{"dest":"169.254.0.5/32","nextHop":"10.134.0.1"},{"dest":"100.64.0.0/16","nextHop":"10.134.0.1"}],"ip_address":"10.134.0.24/23","gateway_ip":"10.134.0.1","role":"primary"}}' k8s.v1.cni.cncf.io/network-status: |- [{ "name": "ovn-kubernetes", "interface": "eth0", "ips": [ "10.134.0.24" ], "mac": "0a:58:0a:86:00:18", "default": true, "dns": {} }] creationTimestamp: "2026-04-16T14:35:56Z" generateName: global-pull-secret-syncer- generation: 1 labels: controller-revision-hash: 66d96fcd56 hypershift.openshift.io/globalps-config-hash: 448b387c name: global-pull-secret-syncer pod-template-generation: "1" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:metadata: f:annotations: .: {} f:k8s.ovn.org/pod-networks: {} manager: ip-10-0-140-144 operation: Update subresource: status time: "2026-04-16T14:35:56Z" - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:metadata: f:generateName: {} f:labels: .: {} f:controller-revision-hash: {} f:hypershift.openshift.io/globalps-config-hash: {} f:name: {} f:pod-template-generation: {} f:ownerReferences: .: {} k:{"uid":"d9c7e270-db12-4f47-bc27-9b2b8d35622f"}: {} f:spec: f:affinity: .: {} f:nodeAffinity: .: {} f:requiredDuringSchedulingIgnoredDuringExecution: {} f:automountServiceAccountToken: {} f:containers: k:{"name":"global-pull-secret-syncer"}: .: {} f:args: {} f:command: {} f:image: {} f:imagePullPolicy: {} f:name: {} f:resources: .: {} f:requests: .: {} f:cpu: {} f:memory: {} f:securityContext: .: {} f:privileged: {} f:terminationMessagePath: {} f:terminationMessagePolicy: {} f:volumeMounts: .: {} k:{"mountPath":"/etc/original-pull-secret"}: .: {} f:mountPath: {} f:name: {} f:readOnly: {} k:{"mountPath":"/var/lib/kubelet"}: .: {} f:mountPath: {} f:name: {} k:{"mountPath":"/var/run/dbus"}: .: {} f:mountPath: {} f:name: {} f:dnsPolicy: {} f:enableServiceLinks: {} f:nodeSelector: {} f:priorityClassName: {} f:restartPolicy: {} f:schedulerName: {} f:securityContext: {} f:terminationGracePeriodSeconds: {} f:tolerations: {} f:volumes: .: {} k:{"name":"dbus"}: .: {} f:hostPath: .: {} f:path: {} f:type: {} f:name: {} k:{"name":"kubelet-config"}: .: {} f:hostPath: .: {} f:path: {} f:type: {} f:name: {} k:{"name":"original-pull-secret"}: .: {} f:name: {} f:secret: .: {} f:defaultMode: {} f:secretName: {} manager: kube-controller-manager operation: Update time: "2026-04-16T14:35:56Z" - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:metadata: f:annotations: f:k8s.v1.cni.cncf.io/network-status: {} manager: multus-daemon operation: Update subresource: status time: "2026-04-16T14:35:56Z" - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:status: f:conditions: k:{"type":"ContainersReady"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"Initialized"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"PodReadyToStartContainers"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"Ready"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} f:containerStatuses: {} f:hostIP: {} f:hostIPs: {} f:phase: {} f:podIP: {} f:podIPs: .: {} k:{"ip":"10.134.0.24"}: .: {} f:ip: {} f:startTime: {} manager: kubelet operation: Update subresource: status time: "2026-04-16T14:36:01Z" name: global-pull-secret-syncer-j5tck namespace: kube-system ownerReferences: - apiVersion: apps/v1 blockOwnerDeletion: true controller: true kind: DaemonSet name: global-pull-secret-syncer uid: d9c7e270-db12-4f47-bc27-9b2b8d35622f resourceVersion: "11406" uid: 32a0461a-84ba-4c24-a026-df53dabb8bbe spec: affinity: nodeAffinity: requiredDuringSchedulingIgnoredDuringExecution: nodeSelectorTerms: - matchFields: - key: metadata.name operator: In values: - ip-10-0-140-144.ec2.internal automountServiceAccountToken: false containers: - args: - sync-global-pullsecret command: - /usr/bin/control-plane-operator image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:46cbf85f447a33f8cec7ead60c8367442a2c09546d1ff9a9a38f0b2ac5dd3f0a imagePullPolicy: IfNotPresent name: global-pull-secret-syncer resources: requests: cpu: 5m memory: 35Mi securityContext: privileged: true terminationMessagePath: /dev/termination-log terminationMessagePolicy: FallbackToLogsOnError volumeMounts: - mountPath: /var/lib/kubelet name: kubelet-config - mountPath: /var/run/dbus name: dbus - mountPath: /etc/original-pull-secret name: original-pull-secret readOnly: true dnsPolicy: Default enableServiceLinks: true imagePullSecrets: - name: default-dockercfg-wfqmd nodeName: ip-10-0-140-144.ec2.internal nodeSelector: hypershift.openshift.io/nodepool-globalps-enabled: "true" preemptionPolicy: PreemptLowerPriority priority: 1000000000 priorityClassName: openshift-user-critical restartPolicy: Always schedulerName: default-scheduler securityContext: {} serviceAccount: default serviceAccountName: default terminationGracePeriodSeconds: 30 tolerations: - operator: Exists volumes: - hostPath: path: /var/lib/kubelet type: Directory name: kubelet-config - hostPath: path: /var/run/dbus type: Directory name: dbus - name: original-pull-secret secret: defaultMode: 420 secretName: original-pull-secret status: conditions: - lastProbeTime: null lastTransitionTime: "2026-04-16T14:36:01Z" status: "True" type: PodReadyToStartContainers - lastProbeTime: null lastTransitionTime: "2026-04-16T14:35:56Z" status: "True" type: Initialized - lastProbeTime: null lastTransitionTime: "2026-04-16T14:36:01Z" status: "True" type: Ready - lastProbeTime: null lastTransitionTime: "2026-04-16T14:36:01Z" status: "True" type: ContainersReady - lastProbeTime: null lastTransitionTime: "2026-04-16T14:35:56Z" status: "True" type: PodScheduled containerStatuses: - allocatedResources: cpu: 5m memory: 35Mi containerID: cri-o://e7e468ba1a6c1dfe3ee75e026ae696ff93d5e3280e331e6c75dd1337e5d7a584 image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:46cbf85f447a33f8cec7ead60c8367442a2c09546d1ff9a9a38f0b2ac5dd3f0a imageID: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:46cbf85f447a33f8cec7ead60c8367442a2c09546d1ff9a9a38f0b2ac5dd3f0a lastState: {} name: global-pull-secret-syncer ready: true resources: requests: cpu: 5m memory: 35Mi restartCount: 0 started: true state: running: startedAt: "2026-04-16T14:36:00Z" user: linux: gid: 0 supplementalGroups: - 0 uid: 0 volumeMounts: - mountPath: /var/lib/kubelet name: kubelet-config - mountPath: /var/run/dbus name: dbus - mountPath: /etc/original-pull-secret name: original-pull-secret readOnly: true recursiveReadOnly: Disabled hostIP: 10.0.140.144 hostIPs: - ip: 10.0.140.144 phase: Running podIP: 10.134.0.24 podIPs: - ip: 10.134.0.24 qosClass: Burstable startTime: "2026-04-16T14:35:56Z" - apiVersion: v1 kind: Pod metadata: annotations: k8s.ovn.org/pod-networks: '{"default":{"ip_addresses":["10.132.0.3/23"],"mac_address":"0a:58:0a:84:00:03","gateway_ips":["10.132.0.1"],"routes":[{"dest":"10.132.0.0/14","nextHop":"10.132.0.1"},{"dest":"172.31.0.0/16","nextHop":"10.132.0.1"},{"dest":"169.254.0.5/32","nextHop":"10.132.0.1"},{"dest":"100.64.0.0/16","nextHop":"10.132.0.1"}],"ip_address":"10.132.0.3/23","gateway_ip":"10.132.0.1","role":"primary"}}' k8s.v1.cni.cncf.io/network-status: |- [{ "name": "ovn-kubernetes", "interface": "eth0", "ips": [ "10.132.0.3" ], "mac": "0a:58:0a:84:00:03", "default": true, "dns": {} }] creationTimestamp: "2026-04-16T14:29:44Z" generateName: global-pull-secret-syncer- generation: 1 labels: controller-revision-hash: 66d96fcd56 hypershift.openshift.io/globalps-config-hash: 448b387c name: global-pull-secret-syncer pod-template-generation: "1" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:metadata: f:generateName: {} f:labels: .: {} f:controller-revision-hash: {} f:hypershift.openshift.io/globalps-config-hash: {} f:name: {} f:pod-template-generation: {} f:ownerReferences: .: {} k:{"uid":"d9c7e270-db12-4f47-bc27-9b2b8d35622f"}: {} f:spec: f:affinity: .: {} f:nodeAffinity: .: {} f:requiredDuringSchedulingIgnoredDuringExecution: {} f:automountServiceAccountToken: {} f:containers: k:{"name":"global-pull-secret-syncer"}: .: {} f:args: {} f:command: {} f:image: {} f:imagePullPolicy: {} f:name: {} f:resources: .: {} f:requests: .: {} f:cpu: {} f:memory: {} f:securityContext: .: {} f:privileged: {} f:terminationMessagePath: {} f:terminationMessagePolicy: {} f:volumeMounts: .: {} k:{"mountPath":"/etc/original-pull-secret"}: .: {} f:mountPath: {} f:name: {} f:readOnly: {} k:{"mountPath":"/var/lib/kubelet"}: .: {} f:mountPath: {} f:name: {} k:{"mountPath":"/var/run/dbus"}: .: {} f:mountPath: {} f:name: {} f:dnsPolicy: {} f:enableServiceLinks: {} f:nodeSelector: {} f:priorityClassName: {} f:restartPolicy: {} f:schedulerName: {} f:securityContext: {} f:terminationGracePeriodSeconds: {} f:tolerations: {} f:volumes: .: {} k:{"name":"dbus"}: .: {} f:hostPath: .: {} f:path: {} f:type: {} f:name: {} k:{"name":"kubelet-config"}: .: {} f:hostPath: .: {} f:path: {} f:type: {} f:name: {} k:{"name":"original-pull-secret"}: .: {} f:name: {} f:secret: .: {} f:defaultMode: {} f:secretName: {} manager: kube-controller-manager operation: Update time: "2026-04-16T14:29:44Z" - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:metadata: f:annotations: .: {} f:k8s.ovn.org/pod-networks: {} manager: ip-10-0-141-239 operation: Update subresource: status time: "2026-04-16T14:30:09Z" - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:metadata: f:annotations: f:k8s.v1.cni.cncf.io/network-status: {} manager: multus-daemon operation: Update subresource: status time: "2026-04-16T14:30:16Z" - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:status: f:conditions: k:{"type":"ContainersReady"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"Initialized"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"PodReadyToStartContainers"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"Ready"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} f:containerStatuses: {} f:hostIP: {} f:hostIPs: {} f:phase: {} f:podIP: {} f:podIPs: .: {} k:{"ip":"10.132.0.3"}: .: {} f:ip: {} f:startTime: {} manager: kubelet operation: Update subresource: status time: "2026-04-16T14:30:25Z" name: global-pull-secret-syncer-x5q8k namespace: kube-system ownerReferences: - apiVersion: apps/v1 blockOwnerDeletion: true controller: true kind: DaemonSet name: global-pull-secret-syncer uid: d9c7e270-db12-4f47-bc27-9b2b8d35622f resourceVersion: "6713" uid: fd31aa30-3e27-4c57-ae7d-843fa27b25d3 spec: affinity: nodeAffinity: requiredDuringSchedulingIgnoredDuringExecution: nodeSelectorTerms: - matchFields: - key: metadata.name operator: In values: - ip-10-0-141-239.ec2.internal automountServiceAccountToken: false containers: - args: - sync-global-pullsecret command: - /usr/bin/control-plane-operator image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:46cbf85f447a33f8cec7ead60c8367442a2c09546d1ff9a9a38f0b2ac5dd3f0a imagePullPolicy: IfNotPresent name: global-pull-secret-syncer resources: requests: cpu: 5m memory: 35Mi securityContext: privileged: true terminationMessagePath: /dev/termination-log terminationMessagePolicy: FallbackToLogsOnError volumeMounts: - mountPath: /var/lib/kubelet name: kubelet-config - mountPath: /var/run/dbus name: dbus - mountPath: /etc/original-pull-secret name: original-pull-secret readOnly: true dnsPolicy: Default enableServiceLinks: true imagePullSecrets: - name: default-dockercfg-wfqmd nodeName: ip-10-0-141-239.ec2.internal nodeSelector: hypershift.openshift.io/nodepool-globalps-enabled: "true" preemptionPolicy: PreemptLowerPriority priority: 1000000000 priorityClassName: openshift-user-critical restartPolicy: Always schedulerName: default-scheduler securityContext: {} serviceAccount: default serviceAccountName: default terminationGracePeriodSeconds: 30 tolerations: - operator: Exists volumes: - hostPath: path: /var/lib/kubelet type: Directory name: kubelet-config - hostPath: path: /var/run/dbus type: Directory name: dbus - name: original-pull-secret secret: defaultMode: 420 secretName: original-pull-secret status: conditions: - lastProbeTime: null lastTransitionTime: "2026-04-16T14:30:25Z" status: "True" type: PodReadyToStartContainers - lastProbeTime: null lastTransitionTime: "2026-04-16T14:29:44Z" status: "True" type: Initialized - lastProbeTime: null lastTransitionTime: "2026-04-16T14:30:25Z" status: "True" type: Ready - lastProbeTime: null lastTransitionTime: "2026-04-16T14:30:25Z" status: "True" type: ContainersReady - lastProbeTime: null lastTransitionTime: "2026-04-16T14:29:44Z" status: "True" type: PodScheduled containerStatuses: - allocatedResources: cpu: 5m memory: 35Mi containerID: cri-o://8adc49304cf42f2e32c14c4eae9dc6ae12d53a93b29f99c19f3f6a0f85cebf96 image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:46cbf85f447a33f8cec7ead60c8367442a2c09546d1ff9a9a38f0b2ac5dd3f0a imageID: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:46cbf85f447a33f8cec7ead60c8367442a2c09546d1ff9a9a38f0b2ac5dd3f0a lastState: {} name: global-pull-secret-syncer ready: true resources: requests: cpu: 5m memory: 35Mi restartCount: 0 started: true state: running: startedAt: "2026-04-16T14:30:25Z" user: linux: gid: 0 supplementalGroups: - 0 uid: 0 volumeMounts: - mountPath: /var/lib/kubelet name: kubelet-config - mountPath: /var/run/dbus name: dbus - mountPath: /etc/original-pull-secret name: original-pull-secret readOnly: true recursiveReadOnly: Disabled hostIP: 10.0.141.239 hostIPs: - ip: 10.0.141.239 phase: Running podIP: 10.132.0.3 podIPs: - ip: 10.132.0.3 qosClass: Burstable startTime: "2026-04-16T14:29:44Z" - apiVersion: v1 kind: Pod metadata: annotations: k8s.ovn.org/pod-networks: '{"default":{"ip_addresses":["10.133.0.5/23"],"mac_address":"0a:58:0a:85:00:05","gateway_ips":["10.133.0.1"],"routes":[{"dest":"10.132.0.0/14","nextHop":"10.133.0.1"},{"dest":"172.31.0.0/16","nextHop":"10.133.0.1"},{"dest":"169.254.0.5/32","nextHop":"10.133.0.1"},{"dest":"100.64.0.0/16","nextHop":"10.133.0.1"}],"ip_address":"10.133.0.5/23","gateway_ip":"10.133.0.1","role":"primary"}}' k8s.v1.cni.cncf.io/network-status: |- [{ "name": "ovn-kubernetes", "interface": "eth0", "ips": [ "10.133.0.5" ], "mac": "0a:58:0a:85:00:05", "default": true, "dns": {} }] creationTimestamp: "2026-04-16T14:29:54Z" generateName: global-pull-secret-syncer- generation: 1 labels: controller-revision-hash: 66d96fcd56 hypershift.openshift.io/globalps-config-hash: 448b387c name: global-pull-secret-syncer pod-template-generation: "1" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:metadata: f:generateName: {} f:labels: .: {} f:controller-revision-hash: {} f:hypershift.openshift.io/globalps-config-hash: {} f:name: {} f:pod-template-generation: {} f:ownerReferences: .: {} k:{"uid":"d9c7e270-db12-4f47-bc27-9b2b8d35622f"}: {} f:spec: f:affinity: .: {} f:nodeAffinity: .: {} f:requiredDuringSchedulingIgnoredDuringExecution: {} f:automountServiceAccountToken: {} f:containers: k:{"name":"global-pull-secret-syncer"}: .: {} f:args: {} f:command: {} f:image: {} f:imagePullPolicy: {} f:name: {} f:resources: .: {} f:requests: .: {} f:cpu: {} f:memory: {} f:securityContext: .: {} f:privileged: {} f:terminationMessagePath: {} f:terminationMessagePolicy: {} f:volumeMounts: .: {} k:{"mountPath":"/etc/original-pull-secret"}: .: {} f:mountPath: {} f:name: {} f:readOnly: {} k:{"mountPath":"/var/lib/kubelet"}: .: {} f:mountPath: {} f:name: {} k:{"mountPath":"/var/run/dbus"}: .: {} f:mountPath: {} f:name: {} f:dnsPolicy: {} f:enableServiceLinks: {} f:nodeSelector: {} f:priorityClassName: {} f:restartPolicy: {} f:schedulerName: {} f:securityContext: {} f:terminationGracePeriodSeconds: {} f:tolerations: {} f:volumes: .: {} k:{"name":"dbus"}: .: {} f:hostPath: .: {} f:path: {} f:type: {} f:name: {} k:{"name":"kubelet-config"}: .: {} f:hostPath: .: {} f:path: {} f:type: {} f:name: {} k:{"name":"original-pull-secret"}: .: {} f:name: {} f:secret: .: {} f:defaultMode: {} f:secretName: {} manager: kube-controller-manager operation: Update time: "2026-04-16T14:29:54Z" - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:metadata: f:annotations: .: {} f:k8s.ovn.org/pod-networks: {} manager: ip-10-0-128-173 operation: Update subresource: status time: "2026-04-16T14:30:11Z" - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:metadata: f:annotations: f:k8s.v1.cni.cncf.io/network-status: {} manager: multus-daemon operation: Update subresource: status time: "2026-04-16T14:30:26Z" - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:status: f:conditions: k:{"type":"ContainersReady"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"Initialized"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"PodReadyToStartContainers"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"Ready"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} f:containerStatuses: {} f:hostIP: {} f:hostIPs: {} f:phase: {} f:podIP: {} f:podIPs: .: {} k:{"ip":"10.133.0.5"}: .: {} f:ip: {} f:startTime: {} manager: kubelet operation: Update subresource: status time: "2026-04-16T14:30:31Z" name: global-pull-secret-syncer-xnzwr namespace: kube-system ownerReferences: - apiVersion: apps/v1 blockOwnerDeletion: true controller: true kind: DaemonSet name: global-pull-secret-syncer uid: d9c7e270-db12-4f47-bc27-9b2b8d35622f resourceVersion: "6810" uid: f35ee2aa-f1ac-4d97-bf10-86c8f12bc700 spec: affinity: nodeAffinity: requiredDuringSchedulingIgnoredDuringExecution: nodeSelectorTerms: - matchFields: - key: metadata.name operator: In values: - ip-10-0-128-173.ec2.internal automountServiceAccountToken: false containers: - args: - sync-global-pullsecret command: - /usr/bin/control-plane-operator image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:46cbf85f447a33f8cec7ead60c8367442a2c09546d1ff9a9a38f0b2ac5dd3f0a imagePullPolicy: IfNotPresent name: global-pull-secret-syncer resources: requests: cpu: 5m memory: 35Mi securityContext: privileged: true terminationMessagePath: /dev/termination-log terminationMessagePolicy: FallbackToLogsOnError volumeMounts: - mountPath: /var/lib/kubelet name: kubelet-config - mountPath: /var/run/dbus name: dbus - mountPath: /etc/original-pull-secret name: original-pull-secret readOnly: true dnsPolicy: Default enableServiceLinks: true imagePullSecrets: - name: default-dockercfg-wfqmd nodeName: ip-10-0-128-173.ec2.internal nodeSelector: hypershift.openshift.io/nodepool-globalps-enabled: "true" preemptionPolicy: PreemptLowerPriority priority: 1000000000 priorityClassName: openshift-user-critical restartPolicy: Always schedulerName: default-scheduler securityContext: {} serviceAccount: default serviceAccountName: default terminationGracePeriodSeconds: 30 tolerations: - operator: Exists volumes: - hostPath: path: /var/lib/kubelet type: Directory name: kubelet-config - hostPath: path: /var/run/dbus type: Directory name: dbus - name: original-pull-secret secret: defaultMode: 420 secretName: original-pull-secret status: conditions: - lastProbeTime: null lastTransitionTime: "2026-04-16T14:30:31Z" status: "True" type: PodReadyToStartContainers - lastProbeTime: null lastTransitionTime: "2026-04-16T14:29:54Z" status: "True" type: Initialized - lastProbeTime: null lastTransitionTime: "2026-04-16T14:30:31Z" status: "True" type: Ready - lastProbeTime: null lastTransitionTime: "2026-04-16T14:30:31Z" status: "True" type: ContainersReady - lastProbeTime: null lastTransitionTime: "2026-04-16T14:29:54Z" status: "True" type: PodScheduled containerStatuses: - allocatedResources: cpu: 5m memory: 35Mi containerID: cri-o://c0b53681bfb9f1b2b2a43ee545a088b31eb89e3f043297ec6d7ae4e8eb105d51 image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:46cbf85f447a33f8cec7ead60c8367442a2c09546d1ff9a9a38f0b2ac5dd3f0a imageID: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:46cbf85f447a33f8cec7ead60c8367442a2c09546d1ff9a9a38f0b2ac5dd3f0a lastState: {} name: global-pull-secret-syncer ready: true resources: requests: cpu: 5m memory: 35Mi restartCount: 0 started: true state: running: startedAt: "2026-04-16T14:30:30Z" user: linux: gid: 0 supplementalGroups: - 0 uid: 0 volumeMounts: - mountPath: /var/lib/kubelet name: kubelet-config - mountPath: /var/run/dbus name: dbus - mountPath: /etc/original-pull-secret name: original-pull-secret readOnly: true recursiveReadOnly: Disabled hostIP: 10.0.128.173 hostIPs: - ip: 10.0.128.173 phase: Running podIP: 10.133.0.5 podIPs: - ip: 10.133.0.5 qosClass: Burstable startTime: "2026-04-16T14:29:54Z" - apiVersion: v1 kind: Pod metadata: annotations: hypershift.openshift.io/release-image: quay.io/openshift-release-dev/ocp-release:4.20.18-multi creationTimestamp: "2026-04-16T14:29:42Z" generateName: konnectivity-agent- generation: 1 labels: app: konnectivity-agent controller-revision-hash: 6b84966db8 pod-template-generation: "1" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:metadata: f:annotations: .: {} f:hypershift.openshift.io/release-image: {} f:generateName: {} f:labels: .: {} f:app: {} f:controller-revision-hash: {} f:pod-template-generation: {} f:ownerReferences: .: {} k:{"uid":"628fbd4d-beee-4ea4-b1e5-90f60d081d19"}: {} f:spec: f:affinity: .: {} f:nodeAffinity: .: {} f:requiredDuringSchedulingIgnoredDuringExecution: {} f:automountServiceAccountToken: {} f:containers: k:{"name":"konnectivity-agent"}: .: {} f:args: {} f:command: {} f:env: .: {} k:{"name":"HTTP_PROXY"}: .: {} f:name: {} k:{"name":"HTTPS_PROXY"}: .: {} f:name: {} k:{"name":"NO_PROXY"}: .: {} f:name: {} f:image: {} f:imagePullPolicy: {} f:livenessProbe: .: {} f:failureThreshold: {} f:httpGet: .: {} f:path: {} f:port: {} f:scheme: {} f:periodSeconds: {} f:successThreshold: {} f:timeoutSeconds: {} f:name: {} f:readinessProbe: .: {} f:failureThreshold: {} f:httpGet: .: {} f:path: {} f:port: {} f:scheme: {} f:periodSeconds: {} f:successThreshold: {} f:timeoutSeconds: {} f:resources: .: {} f:requests: .: {} f:cpu: {} f:memory: {} f:securityContext: .: {} f:allowPrivilegeEscalation: {} f:capabilities: .: {} f:drop: {} f:readOnlyRootFilesystem: {} f:runAsNonRoot: {} f:startupProbe: .: {} f:failureThreshold: {} f:httpGet: .: {} f:path: {} f:port: {} f:scheme: {} f:periodSeconds: {} f:successThreshold: {} f:timeoutSeconds: {} f:terminationMessagePath: {} f:terminationMessagePolicy: {} f:volumeMounts: .: {} k:{"mountPath":"/etc/konnectivity/agent"}: .: {} f:mountPath: {} f:name: {} k:{"mountPath":"/etc/konnectivity/ca"}: .: {} f:mountPath: {} f:name: {} f:dnsPolicy: {} f:enableServiceLinks: {} f:hostNetwork: {} f:priorityClassName: {} f:restartPolicy: {} f:schedulerName: {} f:securityContext: .: {} f:runAsUser: {} f:terminationGracePeriodSeconds: {} f:tolerations: {} f:volumes: .: {} k:{"name":"agent-certs"}: .: {} f:name: {} f:secret: .: {} f:defaultMode: {} f:secretName: {} k:{"name":"konnectivity-ca"}: .: {} f:configMap: .: {} f:defaultMode: {} f:name: {} f:name: {} manager: kube-controller-manager operation: Update time: "2026-04-16T14:29:42Z" - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:status: f:conditions: k:{"type":"ContainersReady"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"Initialized"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"PodReadyToStartContainers"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"Ready"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} f:containerStatuses: {} f:hostIP: {} f:hostIPs: {} f:phase: {} f:podIP: {} f:podIPs: .: {} k:{"ip":"10.0.141.239"}: .: {} f:ip: {} f:startTime: {} manager: kubelet operation: Update subresource: status time: "2026-04-16T14:30:08Z" name: konnectivity-agent-2cqcm namespace: kube-system ownerReferences: - apiVersion: apps/v1 blockOwnerDeletion: true controller: true kind: DaemonSet name: konnectivity-agent uid: 628fbd4d-beee-4ea4-b1e5-90f60d081d19 resourceVersion: "6237" uid: d2ce07ef-ba7f-4f81-a82f-80f139286fa6 spec: affinity: nodeAffinity: requiredDuringSchedulingIgnoredDuringExecution: nodeSelectorTerms: - matchFields: - key: metadata.name operator: In values: - ip-10-0-141-239.ec2.internal automountServiceAccountToken: false containers: - args: - --logtostderr=true - --ca-cert - /etc/konnectivity/ca/ca.crt - --agent-cert - /etc/konnectivity/agent/tls.crt - --agent-key - /etc/konnectivity/agent/tls.key - --proxy-server-host - konnectivity-server-clusters-4c9ca0e0-6aa4-47b4-a305--20a7a192.apps.kflux-prd-es01.1ion.p1.openshiftapps.com - --proxy-server-port - "443" - --health-server-port - "2041" - --agent-identifiers=default-route=true - --keepalive-time - 30s - --probe-interval - 5s - --sync-interval - 5s - --sync-interval-cap - 30s - --v - "3" command: - /usr/bin/proxy-agent env: - name: HTTP_PROXY - name: HTTPS_PROXY - name: NO_PROXY image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:89058c002240ee79922d5ee9c66b581ce6cb78ec85b490c502dfca53b028c187 imagePullPolicy: IfNotPresent livenessProbe: failureThreshold: 6 httpGet: path: healthz port: 2041 scheme: HTTP periodSeconds: 30 successThreshold: 1 timeoutSeconds: 5 name: konnectivity-agent readinessProbe: failureThreshold: 1 httpGet: path: readyz port: 2041 scheme: HTTP periodSeconds: 30 successThreshold: 1 timeoutSeconds: 5 resources: requests: cpu: 40m memory: 50Mi securityContext: allowPrivilegeEscalation: false capabilities: drop: - ALL readOnlyRootFilesystem: true runAsNonRoot: true startupProbe: failureThreshold: 60 httpGet: path: healthz port: 2041 scheme: HTTP periodSeconds: 5 successThreshold: 1 timeoutSeconds: 5 terminationMessagePath: /dev/termination-log terminationMessagePolicy: FallbackToLogsOnError volumeMounts: - mountPath: /etc/konnectivity/agent name: agent-certs - mountPath: /etc/konnectivity/ca name: konnectivity-ca dnsPolicy: Default enableServiceLinks: true hostNetwork: true imagePullSecrets: - name: default-dockercfg-wfqmd nodeName: ip-10-0-141-239.ec2.internal preemptionPolicy: PreemptLowerPriority priority: 2000001000 priorityClassName: system-node-critical restartPolicy: Always schedulerName: default-scheduler securityContext: runAsUser: 1000 serviceAccount: default serviceAccountName: default terminationGracePeriodSeconds: 30 tolerations: - operator: Exists volumes: - name: agent-certs secret: defaultMode: 416 secretName: konnectivity-agent - configMap: defaultMode: 416 name: konnectivity-ca-bundle name: konnectivity-ca status: conditions: - lastProbeTime: null lastTransitionTime: "2026-04-16T14:30:03Z" status: "True" type: PodReadyToStartContainers - lastProbeTime: null lastTransitionTime: "2026-04-16T14:29:43Z" status: "True" type: Initialized - lastProbeTime: null lastTransitionTime: "2026-04-16T14:30:08Z" status: "True" type: Ready - lastProbeTime: null lastTransitionTime: "2026-04-16T14:30:08Z" status: "True" type: ContainersReady - lastProbeTime: null lastTransitionTime: "2026-04-16T14:29:42Z" status: "True" type: PodScheduled containerStatuses: - allocatedResources: cpu: 40m memory: 50Mi containerID: cri-o://033a54462c4a1bd6f1ae81f23abc0a075ff28e33a26c218feff97755cd815029 image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:89058c002240ee79922d5ee9c66b581ce6cb78ec85b490c502dfca53b028c187 imageID: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:59fbca1795a26a2b5f49e01c0c13b67f5bd5dac01be23ebfd19fbbd3a1a230d4 lastState: {} name: konnectivity-agent ready: true resources: requests: cpu: 40m memory: 50Mi restartCount: 0 started: true state: running: startedAt: "2026-04-16T14:30:02Z" user: linux: gid: 0 supplementalGroups: - 0 uid: 1000 volumeMounts: - mountPath: /etc/konnectivity/agent name: agent-certs - mountPath: /etc/konnectivity/ca name: konnectivity-ca hostIP: 10.0.141.239 hostIPs: - ip: 10.0.141.239 phase: Running podIP: 10.0.141.239 podIPs: - ip: 10.0.141.239 qosClass: Burstable startTime: "2026-04-16T14:29:43Z" - apiVersion: v1 kind: Pod metadata: annotations: hypershift.openshift.io/release-image: quay.io/openshift-release-dev/ocp-release:4.20.18-multi creationTimestamp: "2026-04-16T14:29:44Z" generateName: konnectivity-agent- generation: 1 labels: app: konnectivity-agent controller-revision-hash: 6b84966db8 pod-template-generation: "1" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:metadata: f:annotations: .: {} f:hypershift.openshift.io/release-image: {} f:generateName: {} f:labels: .: {} f:app: {} f:controller-revision-hash: {} f:pod-template-generation: {} f:ownerReferences: .: {} k:{"uid":"628fbd4d-beee-4ea4-b1e5-90f60d081d19"}: {} f:spec: f:affinity: .: {} f:nodeAffinity: .: {} f:requiredDuringSchedulingIgnoredDuringExecution: {} f:automountServiceAccountToken: {} f:containers: k:{"name":"konnectivity-agent"}: .: {} f:args: {} f:command: {} f:env: .: {} k:{"name":"HTTP_PROXY"}: .: {} f:name: {} k:{"name":"HTTPS_PROXY"}: .: {} f:name: {} k:{"name":"NO_PROXY"}: .: {} f:name: {} f:image: {} f:imagePullPolicy: {} f:livenessProbe: .: {} f:failureThreshold: {} f:httpGet: .: {} f:path: {} f:port: {} f:scheme: {} f:periodSeconds: {} f:successThreshold: {} f:timeoutSeconds: {} f:name: {} f:readinessProbe: .: {} f:failureThreshold: {} f:httpGet: .: {} f:path: {} f:port: {} f:scheme: {} f:periodSeconds: {} f:successThreshold: {} f:timeoutSeconds: {} f:resources: .: {} f:requests: .: {} f:cpu: {} f:memory: {} f:securityContext: .: {} f:allowPrivilegeEscalation: {} f:capabilities: .: {} f:drop: {} f:readOnlyRootFilesystem: {} f:runAsNonRoot: {} f:startupProbe: .: {} f:failureThreshold: {} f:httpGet: .: {} f:path: {} f:port: {} f:scheme: {} f:periodSeconds: {} f:successThreshold: {} f:timeoutSeconds: {} f:terminationMessagePath: {} f:terminationMessagePolicy: {} f:volumeMounts: .: {} k:{"mountPath":"/etc/konnectivity/agent"}: .: {} f:mountPath: {} f:name: {} k:{"mountPath":"/etc/konnectivity/ca"}: .: {} f:mountPath: {} f:name: {} f:dnsPolicy: {} f:enableServiceLinks: {} f:hostNetwork: {} f:priorityClassName: {} f:restartPolicy: {} f:schedulerName: {} f:securityContext: .: {} f:runAsUser: {} f:terminationGracePeriodSeconds: {} f:tolerations: {} f:volumes: .: {} k:{"name":"agent-certs"}: .: {} f:name: {} f:secret: .: {} f:defaultMode: {} f:secretName: {} k:{"name":"konnectivity-ca"}: .: {} f:configMap: .: {} f:defaultMode: {} f:name: {} f:name: {} manager: kube-controller-manager operation: Update time: "2026-04-16T14:29:44Z" - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:status: f:conditions: k:{"type":"ContainersReady"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"Initialized"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"PodReadyToStartContainers"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"Ready"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} f:containerStatuses: {} f:hostIP: {} f:hostIPs: {} f:phase: {} f:podIP: {} f:podIPs: .: {} k:{"ip":"10.0.128.173"}: .: {} f:ip: {} f:startTime: {} manager: kubelet operation: Update subresource: status time: "2026-04-16T14:30:05Z" name: konnectivity-agent-mn4h5 namespace: kube-system ownerReferences: - apiVersion: apps/v1 blockOwnerDeletion: true controller: true kind: DaemonSet name: konnectivity-agent uid: 628fbd4d-beee-4ea4-b1e5-90f60d081d19 resourceVersion: "6193" uid: 3cf0a81b-ae4c-46b8-b16c-e93bd4e87102 spec: affinity: nodeAffinity: requiredDuringSchedulingIgnoredDuringExecution: nodeSelectorTerms: - matchFields: - key: metadata.name operator: In values: - ip-10-0-128-173.ec2.internal automountServiceAccountToken: false containers: - args: - --logtostderr=true - --ca-cert - /etc/konnectivity/ca/ca.crt - --agent-cert - /etc/konnectivity/agent/tls.crt - --agent-key - /etc/konnectivity/agent/tls.key - --proxy-server-host - konnectivity-server-clusters-4c9ca0e0-6aa4-47b4-a305--20a7a192.apps.kflux-prd-es01.1ion.p1.openshiftapps.com - --proxy-server-port - "443" - --health-server-port - "2041" - --agent-identifiers=default-route=true - --keepalive-time - 30s - --probe-interval - 5s - --sync-interval - 5s - --sync-interval-cap - 30s - --v - "3" command: - /usr/bin/proxy-agent env: - name: HTTP_PROXY - name: HTTPS_PROXY - name: NO_PROXY image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:89058c002240ee79922d5ee9c66b581ce6cb78ec85b490c502dfca53b028c187 imagePullPolicy: IfNotPresent livenessProbe: failureThreshold: 6 httpGet: path: healthz port: 2041 scheme: HTTP periodSeconds: 30 successThreshold: 1 timeoutSeconds: 5 name: konnectivity-agent readinessProbe: failureThreshold: 1 httpGet: path: readyz port: 2041 scheme: HTTP periodSeconds: 30 successThreshold: 1 timeoutSeconds: 5 resources: requests: cpu: 40m memory: 50Mi securityContext: allowPrivilegeEscalation: false capabilities: drop: - ALL readOnlyRootFilesystem: true runAsNonRoot: true startupProbe: failureThreshold: 60 httpGet: path: healthz port: 2041 scheme: HTTP periodSeconds: 5 successThreshold: 1 timeoutSeconds: 5 terminationMessagePath: /dev/termination-log terminationMessagePolicy: FallbackToLogsOnError volumeMounts: - mountPath: /etc/konnectivity/agent name: agent-certs - mountPath: /etc/konnectivity/ca name: konnectivity-ca dnsPolicy: Default enableServiceLinks: true hostNetwork: true imagePullSecrets: - name: default-dockercfg-wfqmd nodeName: ip-10-0-128-173.ec2.internal preemptionPolicy: PreemptLowerPriority priority: 2000001000 priorityClassName: system-node-critical restartPolicy: Always schedulerName: default-scheduler securityContext: runAsUser: 1000 serviceAccount: default serviceAccountName: default terminationGracePeriodSeconds: 30 tolerations: - operator: Exists volumes: - name: agent-certs secret: defaultMode: 416 secretName: konnectivity-agent - configMap: defaultMode: 416 name: konnectivity-ca-bundle name: konnectivity-ca status: conditions: - lastProbeTime: null lastTransitionTime: "2026-04-16T14:30:04Z" status: "True" type: PodReadyToStartContainers - lastProbeTime: null lastTransitionTime: "2026-04-16T14:29:46Z" status: "True" type: Initialized - lastProbeTime: null lastTransitionTime: "2026-04-16T14:30:05Z" status: "True" type: Ready - lastProbeTime: null lastTransitionTime: "2026-04-16T14:30:05Z" status: "True" type: ContainersReady - lastProbeTime: null lastTransitionTime: "2026-04-16T14:29:44Z" status: "True" type: PodScheduled containerStatuses: - allocatedResources: cpu: 40m memory: 50Mi containerID: cri-o://83c712c2be85f85d54d18ffcaf13f076e8e6ef1dc9783537ed492123f84235fb image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:89058c002240ee79922d5ee9c66b581ce6cb78ec85b490c502dfca53b028c187 imageID: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:59fbca1795a26a2b5f49e01c0c13b67f5bd5dac01be23ebfd19fbbd3a1a230d4 lastState: {} name: konnectivity-agent ready: true resources: requests: cpu: 40m memory: 50Mi restartCount: 0 started: true state: running: startedAt: "2026-04-16T14:30:04Z" user: linux: gid: 0 supplementalGroups: - 0 uid: 1000 volumeMounts: - mountPath: /etc/konnectivity/agent name: agent-certs - mountPath: /etc/konnectivity/ca name: konnectivity-ca hostIP: 10.0.128.173 hostIPs: - ip: 10.0.128.173 phase: Running podIP: 10.0.128.173 podIPs: - ip: 10.0.128.173 qosClass: Burstable startTime: "2026-04-16T14:29:46Z" - apiVersion: v1 kind: Pod metadata: annotations: hypershift.openshift.io/release-image: quay.io/openshift-release-dev/ocp-release:4.20.18-multi creationTimestamp: "2026-04-16T14:29:54Z" generateName: konnectivity-agent- generation: 1 labels: app: konnectivity-agent controller-revision-hash: 6b84966db8 pod-template-generation: "1" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:metadata: f:annotations: .: {} f:hypershift.openshift.io/release-image: {} f:generateName: {} f:labels: .: {} f:app: {} f:controller-revision-hash: {} f:pod-template-generation: {} f:ownerReferences: .: {} k:{"uid":"628fbd4d-beee-4ea4-b1e5-90f60d081d19"}: {} f:spec: f:affinity: .: {} f:nodeAffinity: .: {} f:requiredDuringSchedulingIgnoredDuringExecution: {} f:automountServiceAccountToken: {} f:containers: k:{"name":"konnectivity-agent"}: .: {} f:args: {} f:command: {} f:env: .: {} k:{"name":"HTTP_PROXY"}: .: {} f:name: {} k:{"name":"HTTPS_PROXY"}: .: {} f:name: {} k:{"name":"NO_PROXY"}: .: {} f:name: {} f:image: {} f:imagePullPolicy: {} f:livenessProbe: .: {} f:failureThreshold: {} f:httpGet: .: {} f:path: {} f:port: {} f:scheme: {} f:periodSeconds: {} f:successThreshold: {} f:timeoutSeconds: {} f:name: {} f:readinessProbe: .: {} f:failureThreshold: {} f:httpGet: .: {} f:path: {} f:port: {} f:scheme: {} f:periodSeconds: {} f:successThreshold: {} f:timeoutSeconds: {} f:resources: .: {} f:requests: .: {} f:cpu: {} f:memory: {} f:securityContext: .: {} f:allowPrivilegeEscalation: {} f:capabilities: .: {} f:drop: {} f:readOnlyRootFilesystem: {} f:runAsNonRoot: {} f:startupProbe: .: {} f:failureThreshold: {} f:httpGet: .: {} f:path: {} f:port: {} f:scheme: {} f:periodSeconds: {} f:successThreshold: {} f:timeoutSeconds: {} f:terminationMessagePath: {} f:terminationMessagePolicy: {} f:volumeMounts: .: {} k:{"mountPath":"/etc/konnectivity/agent"}: .: {} f:mountPath: {} f:name: {} k:{"mountPath":"/etc/konnectivity/ca"}: .: {} f:mountPath: {} f:name: {} f:dnsPolicy: {} f:enableServiceLinks: {} f:hostNetwork: {} f:priorityClassName: {} f:restartPolicy: {} f:schedulerName: {} f:securityContext: .: {} f:runAsUser: {} f:terminationGracePeriodSeconds: {} f:tolerations: {} f:volumes: .: {} k:{"name":"agent-certs"}: .: {} f:name: {} f:secret: .: {} f:defaultMode: {} f:secretName: {} k:{"name":"konnectivity-ca"}: .: {} f:configMap: .: {} f:defaultMode: {} f:name: {} f:name: {} manager: kube-controller-manager operation: Update time: "2026-04-16T14:29:54Z" - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:status: f:conditions: k:{"type":"ContainersReady"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"Initialized"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"PodReadyToStartContainers"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"Ready"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} f:containerStatuses: {} f:hostIP: {} f:hostIPs: {} f:phase: {} f:podIP: {} f:podIPs: .: {} k:{"ip":"10.0.140.144"}: .: {} f:ip: {} f:startTime: {} manager: kubelet operation: Update subresource: status time: "2026-04-16T14:30:20Z" name: konnectivity-agent-t5h5z namespace: kube-system ownerReferences: - apiVersion: apps/v1 blockOwnerDeletion: true controller: true kind: DaemonSet name: konnectivity-agent uid: 628fbd4d-beee-4ea4-b1e5-90f60d081d19 resourceVersion: "6614" uid: 42343077-d6d1-4ad1-ba48-405f8545fbef spec: affinity: nodeAffinity: requiredDuringSchedulingIgnoredDuringExecution: nodeSelectorTerms: - matchFields: - key: metadata.name operator: In values: - ip-10-0-140-144.ec2.internal automountServiceAccountToken: false containers: - args: - --logtostderr=true - --ca-cert - /etc/konnectivity/ca/ca.crt - --agent-cert - /etc/konnectivity/agent/tls.crt - --agent-key - /etc/konnectivity/agent/tls.key - --proxy-server-host - konnectivity-server-clusters-4c9ca0e0-6aa4-47b4-a305--20a7a192.apps.kflux-prd-es01.1ion.p1.openshiftapps.com - --proxy-server-port - "443" - --health-server-port - "2041" - --agent-identifiers=default-route=true - --keepalive-time - 30s - --probe-interval - 5s - --sync-interval - 5s - --sync-interval-cap - 30s - --v - "3" command: - /usr/bin/proxy-agent env: - name: HTTP_PROXY - name: HTTPS_PROXY - name: NO_PROXY image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:89058c002240ee79922d5ee9c66b581ce6cb78ec85b490c502dfca53b028c187 imagePullPolicy: IfNotPresent livenessProbe: failureThreshold: 6 httpGet: path: healthz port: 2041 scheme: HTTP periodSeconds: 30 successThreshold: 1 timeoutSeconds: 5 name: konnectivity-agent readinessProbe: failureThreshold: 1 httpGet: path: readyz port: 2041 scheme: HTTP periodSeconds: 30 successThreshold: 1 timeoutSeconds: 5 resources: requests: cpu: 40m memory: 50Mi securityContext: allowPrivilegeEscalation: false capabilities: drop: - ALL readOnlyRootFilesystem: true runAsNonRoot: true startupProbe: failureThreshold: 60 httpGet: path: healthz port: 2041 scheme: HTTP periodSeconds: 5 successThreshold: 1 timeoutSeconds: 5 terminationMessagePath: /dev/termination-log terminationMessagePolicy: FallbackToLogsOnError volumeMounts: - mountPath: /etc/konnectivity/agent name: agent-certs - mountPath: /etc/konnectivity/ca name: konnectivity-ca dnsPolicy: Default enableServiceLinks: true hostNetwork: true imagePullSecrets: - name: default-dockercfg-wfqmd nodeName: ip-10-0-140-144.ec2.internal preemptionPolicy: PreemptLowerPriority priority: 2000001000 priorityClassName: system-node-critical restartPolicy: Always schedulerName: default-scheduler securityContext: runAsUser: 1000 serviceAccount: default serviceAccountName: default terminationGracePeriodSeconds: 30 tolerations: - operator: Exists volumes: - name: agent-certs secret: defaultMode: 416 secretName: konnectivity-agent - configMap: defaultMode: 416 name: konnectivity-ca-bundle name: konnectivity-ca status: conditions: - lastProbeTime: null lastTransitionTime: "2026-04-16T14:30:15Z" status: "True" type: PodReadyToStartContainers - lastProbeTime: null lastTransitionTime: "2026-04-16T14:29:54Z" status: "True" type: Initialized - lastProbeTime: null lastTransitionTime: "2026-04-16T14:30:20Z" status: "True" type: Ready - lastProbeTime: null lastTransitionTime: "2026-04-16T14:30:20Z" status: "True" type: ContainersReady - lastProbeTime: null lastTransitionTime: "2026-04-16T14:29:54Z" status: "True" type: PodScheduled containerStatuses: - allocatedResources: cpu: 40m memory: 50Mi containerID: cri-o://15aad2c9933e77976f9907f59232523326be65c96523400b615d4b436d7b38db image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:89058c002240ee79922d5ee9c66b581ce6cb78ec85b490c502dfca53b028c187 imageID: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:59fbca1795a26a2b5f49e01c0c13b67f5bd5dac01be23ebfd19fbbd3a1a230d4 lastState: {} name: konnectivity-agent ready: true resources: requests: cpu: 40m memory: 50Mi restartCount: 0 started: true state: running: startedAt: "2026-04-16T14:30:14Z" user: linux: gid: 0 supplementalGroups: - 0 uid: 1000 volumeMounts: - mountPath: /etc/konnectivity/agent name: agent-certs - mountPath: /etc/konnectivity/ca name: konnectivity-ca hostIP: 10.0.140.144 hostIPs: - ip: 10.0.140.144 phase: Running podIP: 10.0.140.144 podIPs: - ip: 10.0.140.144 qosClass: Burstable startTime: "2026-04-16T14:29:54Z" - apiVersion: v1 kind: Pod metadata: annotations: kubernetes.io/config.hash: a454133692b7d59775381b8452362b38 kubernetes.io/config.mirror: a454133692b7d59775381b8452362b38 kubernetes.io/config.seen: "2026-04-16T14:29:44.605439820Z" kubernetes.io/config.source: file creationTimestamp: "2026-04-16T14:29:45Z" generation: 1 labels: k8s-app: kube-apiserver-proxy managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:metadata: f:annotations: .: {} f:kubernetes.io/config.hash: {} f:kubernetes.io/config.mirror: {} f:kubernetes.io/config.seen: {} f:kubernetes.io/config.source: {} f:labels: .: {} f:k8s-app: {} f:ownerReferences: .: {} k:{"uid":"06087004-107c-4ed7-a7f2-01e53500aa43"}: {} f:spec: f:containers: k:{"name":"haproxy"}: .: {} f:command: {} f:image: {} f:imagePullPolicy: {} f:livenessProbe: .: {} f:failureThreshold: {} f:httpGet: .: {} f:host: {} f:path: {} f:port: {} f:scheme: {} f:initialDelaySeconds: {} f:periodSeconds: {} f:successThreshold: {} f:timeoutSeconds: {} f:name: {} f:ports: .: {} k:{"containerPort":6443,"protocol":"TCP"}: .: {} f:containerPort: {} f:hostPort: {} f:name: {} f:protocol: {} f:resources: .: {} f:requests: .: {} f:cpu: {} f:memory: {} f:securityContext: .: {} f:runAsUser: {} f:terminationMessagePath: {} f:terminationMessagePolicy: {} f:volumeMounts: .: {} k:{"mountPath":"/usr/local/etc/haproxy"}: .: {} f:mountPath: {} f:name: {} f:dnsPolicy: {} f:enableServiceLinks: {} f:hostNetwork: {} f:nodeName: {} f:priorityClassName: {} f:restartPolicy: {} f:schedulerName: {} f:securityContext: {} f:terminationGracePeriodSeconds: {} f:tolerations: {} f:volumes: .: {} k:{"name":"config"}: .: {} f:hostPath: .: {} f:path: {} f:type: {} f:name: {} manager: kubelet operation: Update time: "2026-04-16T14:29:45Z" - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:status: f:conditions: .: {} k:{"type":"ContainersReady"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"Initialized"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"PodReadyToStartContainers"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"PodScheduled"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"Ready"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} f:containerStatuses: {} f:hostIP: {} f:hostIPs: {} f:phase: {} f:podIP: {} f:podIPs: .: {} k:{"ip":"10.0.128.173"}: .: {} f:ip: {} f:startTime: {} manager: kubelet operation: Update subresource: status time: "2026-04-16T14:29:47Z" name: kube-apiserver-proxy-ip-10-0-128-173.ec2.internal namespace: kube-system ownerReferences: - apiVersion: v1 controller: true kind: Node name: ip-10-0-128-173.ec2.internal uid: 06087004-107c-4ed7-a7f2-01e53500aa43 resourceVersion: "5652" uid: 048cc6d7-ae55-4c3b-adef-bf5756bfd4eb spec: containers: - command: - haproxy - -f - /usr/local/etc/haproxy image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f2e4763905898d3870f64ebc9721d8d43ae2973f4ba295d48f84e36e6f72d013 imagePullPolicy: IfNotPresent livenessProbe: failureThreshold: 3 httpGet: host: 172.20.0.1 path: /version port: 6443 scheme: HTTPS initialDelaySeconds: 120 periodSeconds: 120 successThreshold: 1 timeoutSeconds: 1 name: haproxy ports: - containerPort: 6443 hostPort: 6443 name: apiserver protocol: TCP resources: requests: cpu: 13m memory: 16Mi securityContext: runAsUser: 1001 terminationMessagePath: /dev/termination-log terminationMessagePolicy: File volumeMounts: - mountPath: /usr/local/etc/haproxy name: config dnsPolicy: ClusterFirst enableServiceLinks: true hostNetwork: true nodeName: ip-10-0-128-173.ec2.internal preemptionPolicy: PreemptLowerPriority priority: 2000001000 priorityClassName: system-node-critical restartPolicy: Always schedulerName: default-scheduler securityContext: {} terminationGracePeriodSeconds: 30 tolerations: - effect: NoExecute operator: Exists - effect: NoSchedule key: node.kubernetes.io/memory-pressure operator: Exists volumes: - hostPath: path: /etc/kubernetes/apiserver-proxy-config type: "" name: config status: conditions: - lastProbeTime: null lastTransitionTime: "2026-04-16T14:29:47Z" status: "True" type: PodReadyToStartContainers - lastProbeTime: null lastTransitionTime: "2026-04-16T14:29:44Z" status: "True" type: Initialized - lastProbeTime: null lastTransitionTime: "2026-04-16T14:29:47Z" status: "True" type: Ready - lastProbeTime: null lastTransitionTime: "2026-04-16T14:29:47Z" status: "True" type: ContainersReady - lastProbeTime: null lastTransitionTime: "2026-04-16T14:29:44Z" status: "True" type: PodScheduled containerStatuses: - allocatedResources: cpu: 13m memory: 16Mi containerID: cri-o://0c94e7fff4273491fc90143854fa0bc86cfbf49b26ad254ff8e68979a576858f image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f2e4763905898d3870f64ebc9721d8d43ae2973f4ba295d48f84e36e6f72d013 imageID: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7534cffb860df8351d4e7c4872551f4238b71f161adb876a5852061455e5eb7e lastState: {} name: haproxy ready: true resources: requests: cpu: 13m memory: 16Mi restartCount: 0 started: true state: running: startedAt: "2026-04-16T14:29:47Z" user: linux: gid: 0 supplementalGroups: - 0 uid: 1001 volumeMounts: - mountPath: /usr/local/etc/haproxy name: config hostIP: 10.0.128.173 hostIPs: - ip: 10.0.128.173 phase: Running podIP: 10.0.128.173 podIPs: - ip: 10.0.128.173 qosClass: Burstable startTime: "2026-04-16T14:29:44Z" - apiVersion: v1 kind: Pod metadata: annotations: kubernetes.io/config.hash: ab72f00075c1175e59c7e696357a6702 kubernetes.io/config.mirror: ab72f00075c1175e59c7e696357a6702 kubernetes.io/config.seen: "2026-04-16T14:29:53.823939675Z" kubernetes.io/config.source: file creationTimestamp: "2026-04-16T14:29:54Z" generation: 1 labels: k8s-app: kube-apiserver-proxy managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:metadata: f:annotations: .: {} f:kubernetes.io/config.hash: {} f:kubernetes.io/config.mirror: {} f:kubernetes.io/config.seen: {} f:kubernetes.io/config.source: {} f:labels: .: {} f:k8s-app: {} f:ownerReferences: .: {} k:{"uid":"dbc9ed74-ceb3-433d-8bfb-c1237addf952"}: {} f:spec: f:containers: k:{"name":"haproxy"}: .: {} f:command: {} f:image: {} f:imagePullPolicy: {} f:livenessProbe: .: {} f:failureThreshold: {} f:httpGet: .: {} f:host: {} f:path: {} f:port: {} f:scheme: {} f:initialDelaySeconds: {} f:periodSeconds: {} f:successThreshold: {} f:timeoutSeconds: {} f:name: {} f:ports: .: {} k:{"containerPort":6443,"protocol":"TCP"}: .: {} f:containerPort: {} f:hostPort: {} f:name: {} f:protocol: {} f:resources: .: {} f:requests: .: {} f:cpu: {} f:memory: {} f:securityContext: .: {} f:runAsUser: {} f:terminationMessagePath: {} f:terminationMessagePolicy: {} f:volumeMounts: .: {} k:{"mountPath":"/usr/local/etc/haproxy"}: .: {} f:mountPath: {} f:name: {} f:dnsPolicy: {} f:enableServiceLinks: {} f:hostNetwork: {} f:nodeName: {} f:priorityClassName: {} f:restartPolicy: {} f:schedulerName: {} f:securityContext: {} f:terminationGracePeriodSeconds: {} f:tolerations: {} f:volumes: .: {} k:{"name":"config"}: .: {} f:hostPath: .: {} f:path: {} f:type: {} f:name: {} manager: kubelet operation: Update time: "2026-04-16T14:29:54Z" - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:status: f:conditions: .: {} k:{"type":"ContainersReady"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"Initialized"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"PodReadyToStartContainers"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"PodScheduled"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"Ready"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} f:containerStatuses: {} f:hostIP: {} f:hostIPs: {} f:phase: {} f:podIP: {} f:podIPs: .: {} k:{"ip":"10.0.140.144"}: .: {} f:ip: {} f:startTime: {} manager: kubelet operation: Update subresource: status time: "2026-04-16T14:30:14Z" name: kube-apiserver-proxy-ip-10-0-140-144.ec2.internal namespace: kube-system ownerReferences: - apiVersion: v1 controller: true kind: Node name: ip-10-0-140-144.ec2.internal uid: dbc9ed74-ceb3-433d-8bfb-c1237addf952 resourceVersion: "6415" uid: 7446ae76-6858-4688-8bc1-3661f097a545 spec: containers: - command: - haproxy - -f - /usr/local/etc/haproxy image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f2e4763905898d3870f64ebc9721d8d43ae2973f4ba295d48f84e36e6f72d013 imagePullPolicy: IfNotPresent livenessProbe: failureThreshold: 3 httpGet: host: 172.20.0.1 path: /version port: 6443 scheme: HTTPS initialDelaySeconds: 120 periodSeconds: 120 successThreshold: 1 timeoutSeconds: 1 name: haproxy ports: - containerPort: 6443 hostPort: 6443 name: apiserver protocol: TCP resources: requests: cpu: 13m memory: 16Mi securityContext: runAsUser: 1001 terminationMessagePath: /dev/termination-log terminationMessagePolicy: File volumeMounts: - mountPath: /usr/local/etc/haproxy name: config dnsPolicy: ClusterFirst enableServiceLinks: true hostNetwork: true nodeName: ip-10-0-140-144.ec2.internal preemptionPolicy: PreemptLowerPriority priority: 2000001000 priorityClassName: system-node-critical restartPolicy: Always schedulerName: default-scheduler securityContext: {} terminationGracePeriodSeconds: 30 tolerations: - effect: NoExecute operator: Exists - effect: NoSchedule key: node.kubernetes.io/memory-pressure operator: Exists volumes: - hostPath: path: /etc/kubernetes/apiserver-proxy-config type: "" name: config status: conditions: - lastProbeTime: null lastTransitionTime: "2026-04-16T14:30:14Z" status: "True" type: PodReadyToStartContainers - lastProbeTime: null lastTransitionTime: "2026-04-16T14:29:54Z" status: "True" type: Initialized - lastProbeTime: null lastTransitionTime: "2026-04-16T14:30:14Z" status: "True" type: Ready - lastProbeTime: null lastTransitionTime: "2026-04-16T14:30:14Z" status: "True" type: ContainersReady - lastProbeTime: null lastTransitionTime: "2026-04-16T14:29:54Z" status: "True" type: PodScheduled containerStatuses: - allocatedResources: cpu: 13m memory: 16Mi containerID: cri-o://8b3c2a07cecd780b28e4dc7e13208fc60e4999e7b760483c1b08e8e6f53baf38 image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f2e4763905898d3870f64ebc9721d8d43ae2973f4ba295d48f84e36e6f72d013 imageID: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7534cffb860df8351d4e7c4872551f4238b71f161adb876a5852061455e5eb7e lastState: {} name: haproxy ready: true resources: requests: cpu: 13m memory: 16Mi restartCount: 0 started: true state: running: startedAt: "2026-04-16T14:30:13Z" user: linux: gid: 0 supplementalGroups: - 0 uid: 1001 volumeMounts: - mountPath: /usr/local/etc/haproxy name: config hostIP: 10.0.140.144 hostIPs: - ip: 10.0.140.144 phase: Running podIP: 10.0.140.144 podIPs: - ip: 10.0.140.144 qosClass: Burstable startTime: "2026-04-16T14:29:54Z" - apiVersion: v1 kind: Pod metadata: annotations: kubernetes.io/config.hash: 93794f44aea94deab445c9557d98f20e kubernetes.io/config.mirror: 93794f44aea94deab445c9557d98f20e kubernetes.io/config.seen: "2026-04-16T14:29:42.190745801Z" kubernetes.io/config.source: file creationTimestamp: "2026-04-16T14:29:43Z" generation: 1 labels: k8s-app: kube-apiserver-proxy managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:metadata: f:annotations: .: {} f:kubernetes.io/config.hash: {} f:kubernetes.io/config.mirror: {} f:kubernetes.io/config.seen: {} f:kubernetes.io/config.source: {} f:labels: .: {} f:k8s-app: {} f:ownerReferences: .: {} k:{"uid":"8dab1a35-308c-49df-b687-fdc08ebe6058"}: {} f:spec: f:containers: k:{"name":"haproxy"}: .: {} f:command: {} f:image: {} f:imagePullPolicy: {} f:livenessProbe: .: {} f:failureThreshold: {} f:httpGet: .: {} f:host: {} f:path: {} f:port: {} f:scheme: {} f:initialDelaySeconds: {} f:periodSeconds: {} f:successThreshold: {} f:timeoutSeconds: {} f:name: {} f:ports: .: {} k:{"containerPort":6443,"protocol":"TCP"}: .: {} f:containerPort: {} f:hostPort: {} f:name: {} f:protocol: {} f:resources: .: {} f:requests: .: {} f:cpu: {} f:memory: {} f:securityContext: .: {} f:runAsUser: {} f:terminationMessagePath: {} f:terminationMessagePolicy: {} f:volumeMounts: .: {} k:{"mountPath":"/usr/local/etc/haproxy"}: .: {} f:mountPath: {} f:name: {} f:dnsPolicy: {} f:enableServiceLinks: {} f:hostNetwork: {} f:nodeName: {} f:priorityClassName: {} f:restartPolicy: {} f:schedulerName: {} f:securityContext: {} f:terminationGracePeriodSeconds: {} f:tolerations: {} f:volumes: .: {} k:{"name":"config"}: .: {} f:hostPath: .: {} f:path: {} f:type: {} f:name: {} manager: kubelet operation: Update time: "2026-04-16T14:29:43Z" - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:status: f:conditions: .: {} k:{"type":"ContainersReady"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"Initialized"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"PodReadyToStartContainers"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"PodScheduled"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} k:{"type":"Ready"}: .: {} f:lastProbeTime: {} f:lastTransitionTime: {} f:status: {} f:type: {} f:containerStatuses: {} f:hostIP: {} f:hostIPs: {} f:phase: {} f:podIP: {} f:podIPs: .: {} k:{"ip":"10.0.141.239"}: .: {} f:ip: {} f:startTime: {} manager: kubelet operation: Update subresource: status time: "2026-04-16T14:30:02Z" name: kube-apiserver-proxy-ip-10-0-141-239.ec2.internal namespace: kube-system ownerReferences: - apiVersion: v1 controller: true kind: Node name: ip-10-0-141-239.ec2.internal uid: 8dab1a35-308c-49df-b687-fdc08ebe6058 resourceVersion: "6017" uid: a7b59bc5-01e6-44ce-babe-4a18831fc64f spec: containers: - command: - haproxy - -f - /usr/local/etc/haproxy image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f2e4763905898d3870f64ebc9721d8d43ae2973f4ba295d48f84e36e6f72d013 imagePullPolicy: IfNotPresent livenessProbe: failureThreshold: 3 httpGet: host: 172.20.0.1 path: /version port: 6443 scheme: HTTPS initialDelaySeconds: 120 periodSeconds: 120 successThreshold: 1 timeoutSeconds: 1 name: haproxy ports: - containerPort: 6443 hostPort: 6443 name: apiserver protocol: TCP resources: requests: cpu: 13m memory: 16Mi securityContext: runAsUser: 1001 terminationMessagePath: /dev/termination-log terminationMessagePolicy: File volumeMounts: - mountPath: /usr/local/etc/haproxy name: config dnsPolicy: ClusterFirst enableServiceLinks: true hostNetwork: true nodeName: ip-10-0-141-239.ec2.internal preemptionPolicy: PreemptLowerPriority priority: 2000001000 priorityClassName: system-node-critical restartPolicy: Always schedulerName: default-scheduler securityContext: {} terminationGracePeriodSeconds: 30 tolerations: - effect: NoExecute operator: Exists - effect: NoSchedule key: node.kubernetes.io/memory-pressure operator: Exists volumes: - hostPath: path: /etc/kubernetes/apiserver-proxy-config type: "" name: config status: conditions: - lastProbeTime: null lastTransitionTime: "2026-04-16T14:30:02Z" status: "True" type: PodReadyToStartContainers - lastProbeTime: null lastTransitionTime: "2026-04-16T14:29:42Z" status: "True" type: Initialized - lastProbeTime: null lastTransitionTime: "2026-04-16T14:30:02Z" status: "True" type: Ready - lastProbeTime: null lastTransitionTime: "2026-04-16T14:30:02Z" status: "True" type: ContainersReady - lastProbeTime: null lastTransitionTime: "2026-04-16T14:29:42Z" status: "True" type: PodScheduled containerStatuses: - allocatedResources: cpu: 13m memory: 16Mi containerID: cri-o://ae8869a245b461e35ff3bf5dd65b41212e20c3702503e950a7bbcf1ad4aa74c7 image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f2e4763905898d3870f64ebc9721d8d43ae2973f4ba295d48f84e36e6f72d013 imageID: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7534cffb860df8351d4e7c4872551f4238b71f161adb876a5852061455e5eb7e lastState: {} name: haproxy ready: true resources: requests: cpu: 13m memory: 16Mi restartCount: 0 started: true state: running: startedAt: "2026-04-16T14:30:02Z" user: linux: gid: 0 supplementalGroups: - 0 uid: 1001 volumeMounts: - mountPath: /usr/local/etc/haproxy name: config hostIP: 10.0.141.239 hostIPs: - ip: 10.0.141.239 phase: Running podIP: 10.0.141.239 podIPs: - ip: 10.0.141.239 qosClass: Burstable startTime: "2026-04-16T14:29:42Z" kind: PodList metadata: resourceVersion: "29295"