--- apiVersion: v1 items: - apiVersion: v1 count: 1 eventTime: null firstTimestamp: "2026-04-20T14:59:18Z" involvedObject: apiVersion: operator.openshift.io/v1alpha1 kind: IstioCSR kind: Event lastTimestamp: "2026-04-20T14:59:18Z" message: controller is starting metadata: creationTimestamp: "2026-04-20T14:59:18Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:source: f:component: {} f:type: {} manager: cert-manager-operator operation: Update time: "2026-04-20T14:59:18Z" name: 222d245c-a556-4b04-85af-776f9b11e665 namespace: default resourceVersion: "11715" uid: fc80a3d7-7a3f-4a4a-8e60-a697b984bfdb reason: ControllerStarted reportingComponent: cert-manager-istio-csr-controller reportingInstance: "" source: component: cert-manager-istio-csr-controller type: Normal - apiVersion: v1 count: 1 eventTime: null firstTimestamp: "2026-04-20T14:54:14Z" involvedObject: kind: CertificateSigningRequest name: csr-79tlf kind: Event lastTimestamp: "2026-04-20T14:54:14Z" message: CSR "csr-79tlf" has been approved metadata: creationTimestamp: "2026-04-20T14:54:14Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:source: f:component: {} f:type: {} manager: ovnkube-identity operation: Update time: "2026-04-20T14:54:14Z" name: csr-79tlf.18a818603af090d5 namespace: default resourceVersion: "6587" uid: 027028e2-1cad-4510-8e52-aadd4fcbe1fd reason: CSRApproved reportingComponent: ovnkube-csr-approver-controller reportingInstance: "" source: component: ovnkube-csr-approver-controller type: Normal - apiVersion: v1 count: 1 eventTime: null firstTimestamp: "2026-04-20T14:54:02Z" involvedObject: kind: CertificateSigningRequest name: csr-cpj92 kind: Event lastTimestamp: "2026-04-20T14:54:02Z" message: CSR "csr-cpj92" has been approved metadata: creationTimestamp: "2026-04-20T14:54:02Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:source: f:component: {} f:type: {} manager: ovnkube-identity operation: Update time: "2026-04-20T14:54:02Z" name: csr-cpj92.18a8185d6c7d8c7b namespace: default resourceVersion: "6306" uid: 4bdcc455-9d23-431b-95d6-53d21377a241 reason: CSRApproved reportingComponent: ovnkube-csr-approver-controller reportingInstance: "" source: component: ovnkube-csr-approver-controller type: Normal - apiVersion: v1 count: 1 eventTime: null firstTimestamp: "2026-04-20T14:54:20Z" involvedObject: kind: CertificateSigningRequest name: csr-f6crl kind: Event lastTimestamp: "2026-04-20T14:54:20Z" message: CSR "csr-f6crl" has been approved metadata: creationTimestamp: "2026-04-20T14:54:20Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:source: f:component: {} f:type: {} manager: ovnkube-identity operation: Update time: "2026-04-20T14:54:20Z" name: csr-f6crl.18a81861adba5cba namespace: default resourceVersion: "6680" uid: fb04d454-6613-4d11-9eeb-ed0951a311db reason: CSRApproved reportingComponent: ovnkube-csr-approver-controller reportingInstance: "" source: component: ovnkube-csr-approver-controller type: Normal - apiVersion: v1 count: 1 eventTime: null firstTimestamp: "2026-04-20T14:53:55Z" involvedObject: kind: CertificateSigningRequest name: csr-js5p8 kind: Event lastTimestamp: "2026-04-20T14:53:55Z" message: CSR "csr-js5p8" has been approved metadata: creationTimestamp: "2026-04-20T14:53:55Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:source: f:component: {} f:type: {} manager: ovnkube-identity operation: Update time: "2026-04-20T14:53:55Z" name: csr-js5p8.18a8185bb559e4c8 namespace: default resourceVersion: "6155" uid: 7ee516db-9027-411a-9b63-b35ab4a2c1f6 reason: CSRApproved reportingComponent: ovnkube-csr-approver-controller reportingInstance: "" source: component: ovnkube-csr-approver-controller type: Normal - apiVersion: v1 count: 1 eventTime: null firstTimestamp: "2026-04-20T14:53:59Z" involvedObject: kind: CertificateSigningRequest name: csr-lxwbj kind: Event lastTimestamp: "2026-04-20T14:53:59Z" message: CSR "csr-lxwbj" has been approved metadata: creationTimestamp: "2026-04-20T14:53:59Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:source: f:component: {} f:type: {} manager: ovnkube-identity operation: Update time: "2026-04-20T14:53:59Z" name: csr-lxwbj.18a8185cb86a7279 namespace: default resourceVersion: "6237" uid: e4c7e401-024b-4292-a097-58c728c3258e reason: CSRApproved reportingComponent: ovnkube-csr-approver-controller reportingInstance: "" source: component: ovnkube-csr-approver-controller type: Normal - apiVersion: v1 count: 1 eventTime: null firstTimestamp: "2026-04-20T14:53:53Z" involvedObject: kind: CertificateSigningRequest name: csr-wlbpx kind: Event lastTimestamp: "2026-04-20T14:53:53Z" message: CSR "csr-wlbpx" has been approved metadata: creationTimestamp: "2026-04-20T14:53:53Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:source: f:component: {} f:type: {} manager: ovnkube-identity operation: Update time: "2026-04-20T14:53:53Z" name: csr-wlbpx.18a8185b40ff6664 namespace: default resourceVersion: "6107" uid: 2ad60666-1a0e-4887-9149-c5322e080d11 reason: CSRApproved reportingComponent: ovnkube-csr-approver-controller reportingInstance: "" source: component: ovnkube-csr-approver-controller type: Normal - apiVersion: v1 count: 11 eventTime: null firstTimestamp: "2026-04-20T15:00:08Z" involvedObject: apiVersion: services.platform.opendatahub.io/v1alpha1 kind: GatewayConfig name: default-gateway resourceVersion: "13934" uid: c7d1550a-ae09-4649-a161-8fb8ef02a912 kind: Event lastTimestamp: "2026-04-20T15:00:16Z" message: 'failed to lookup object openshift-ingress/data-science-tls-rule: no matches for kind "DestinationRule" in version "networking.istio.io/v1"' metadata: creationTimestamp: "2026-04-20T15:00:08Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:source: f:component: {} f:type: {} manager: manager operation: Update time: "2026-04-20T15:00:16Z" name: default-gateway.18a818b2b938636c namespace: default resourceVersion: "14143" uid: 0ab91250-9ecc-4a3f-ab45-498ed6643ce4 reason: ProvisioningError reportingComponent: gatewayconfig reportingInstance: "" source: component: gatewayconfig type: Warning - apiVersion: v1 count: 1 eventTime: null firstTimestamp: "2026-04-20T15:00:43Z" involvedObject: apiVersion: services.platform.opendatahub.io/v1alpha1 kind: GatewayConfig name: default-gateway resourceVersion: "15669" uid: c7d1550a-ae09-4649-a161-8fb8ef02a912 kind: Event lastTimestamp: "2026-04-20T15:00:43Z" message: 'failed to detect ingress mode: failed to update GatewayConfig with detected mode: Operation cannot be fulfilled on gatewayconfigs.services.platform.opendatahub.io "default-gateway": the object has been modified; please apply your changes to the latest version and try again' metadata: creationTimestamp: "2026-04-20T15:00:43Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:source: f:component: {} f:type: {} manager: manager operation: Update time: "2026-04-20T15:00:43Z" name: default-gateway.18a818baab41e6dd namespace: default resourceVersion: "15670" uid: 6ae3c989-98f0-4d8f-85e4-792e315dde32 reason: ProvisioningError reportingComponent: gatewayconfig reportingInstance: "" source: component: gatewayconfig type: Warning - apiVersion: v1 count: 13 eventTime: null firstTimestamp: "2026-04-20T15:00:18Z" involvedObject: apiVersion: components.platform.opendatahub.io/v1alpha1 kind: Kserve name: default-kserve resourceVersion: "14267" uid: 7ebc5c5a-06f4-404f-ac2a-c32e9c5fd98c kind: Event lastTimestamp: "2026-04-20T15:00:47Z" message: "failure deploying resource {map[apiVersion:serving.kserve.io/v1alpha1 kind:LLMInferenceServiceConfig metadata:map[annotations:map[internal.config.kubernetes.io/previousKinds:LLMInferenceServiceConfig internal.config.kubernetes.io/previousNames:kserve-config-llm-decode-template internal.config.kubernetes.io/previousNamespaces:opendatahub platform.opendatahub.io/instance.generation:1 platform.opendatahub.io/instance.name:default-kserve platform.opendatahub.io/instance.uid:7ebc5c5a-06f4-404f-ac2a-c32e9c5fd98c platform.opendatahub.io/type:Open Data Hub platform.opendatahub.io/version:3.4.0-ea.1 serving.kserve.io/well-known-config:true] labels:map[app.kubernetes.io/part-of:kserve app.opendatahub.io/kserve:true platform.opendatahub.io/part-of:kserve] name:v3-4-0-ea-1-kserve-config-llm-decode-template namespace:opendatahub ownerReferences:[map[apiVersion:components.platform.opendatahub.io/v1alpha1 blockOwnerDeletion:%!s(bool=true) controller:%!s(bool=true) kind:Kserve name:default-kserve uid:7ebc5c5a-06f4-404f-ac2a-c32e9c5fd98c]]] spec:map[template:map[containers:[map[args:[if [ \"$KSERVE_INFER_ROCE\" = \"true\" ]; then\n echo \"Trying to infer RoCE configs ... \"\n grep -H . /sys/class/infiniband/*/ports/*/gids/* 2>/dev/null\n grep -H . /sys/class/infiniband/*/ports/*/gid_attrs/types/* 2>/dev/null\n\n cat /proc/driver/nvidia/params\n\n \ KSERVE_INFER_IB_GID_INDEX_GREP=${KSERVE_INFER_IB_GID_INDEX_GREP:-\"RoCE v2\"}\n\n \ echo \"[Infer RoCE] Discovering active HCAs ...\"\n active_hcas=()\n # Loop through all mlx5 devices found in sysfs\n for hca_dir in /sys/class/infiniband/mlx5_*; do\n # Ensure it's a directory before proceeding\n if [ -d \"$hca_dir\" ]; then\n hca_name=$(basename \"$hca_dir\")\n port_state_file=\"$hca_dir/ports/1/state\" # Assume port 1\n type_file=\"$hca_dir/ports/1/gid_attrs/types/*\"\n\n \ echo \"[Infer RoCE] Check if the port state file ${port_state_file} exists and contains 'ACTIVE'\"\n if [ -f \"$port_state_file\" ] && grep -q \"ACTIVE\" \"$port_state_file\" && grep -q \"${KSERVE_INFER_IB_GID_INDEX_GREP}\" ${type_file} 2>/dev/null; then\n echo \"[Infer RoCE] Found active HCA: $hca_name\"\n active_hcas+=(\"$hca_name\")\n else\n \ echo \"[Infer RoCE] Skipping inactive or down HCA: $hca_name\"\n \ fi\n fi\n done\n\n ucx_hcas=()\n for hca in \"${active_hcas[@]}\"; do\n ucx_hcas+=(\"${hca}:1\")\n done\n\n # Check if we found any active HCAs\n \ if [ ${#active_hcas[@]} -gt 0 ]; then\n # Join the array elements with a comma\n hcas=$(IFS=,; echo \"${active_hcas[*]}\")\n echo \"[Infer RoCE] Setting active HCAs: ${hcas}\"\n export NCCL_IB_HCA=${NCCL_IB_HCA:-${hcas}}\n \ export NVSHMEM_HCA_LIST=${NVSHMEM_HCA_LIST:-${ucx_hcas}}\n export UCX_NET_DEVICES=${UCX_NET_DEVICES:-${ucx_hcas}}\n\n \ echo \"[Infer RoCE] NCCL_IB_HCA=${NCCL_IB_HCA}\"\n echo \"[Infer RoCE] NVSHMEM_HCA_LIST=${NVSHMEM_HCA_LIST}\"\n else\n echo \"[Infer RoCE] WARNING: No active RoCE HCAs found. NCCL_IB_HCA will not be set.\"\n fi\n\n if [ ${#active_hcas[@]} -gt 0 ]; then\n echo \"[Infer RoCE] Finding GID_INDEX for each active HCA (SR-IOV compatible)...\"\n\n # For SR-IOV environments, find the most common IPv4 RoCE v2 GID index across all HCAs\n declare -A gid_index_count\n declare -A hca_gid_index\n\n for hca_name in \"${active_hcas[@]}\"; do\n echo \"[Infer RoCE] Processing HCA: ${hca_name}\"\n\n # Find all RoCE v2 IPv4 GIDs for this HCA and count by index\n for tpath in /sys/class/infiniband/${hca_name}/ports/1/gid_attrs/types/*; do\n if grep -q \"${KSERVE_INFER_IB_GID_INDEX_GREP}\" \"$tpath\" 2>/dev/null; then\n idx=$(basename \"$tpath\")\n gid_file=\"/sys/class/infiniband/${hca_name}/ports/1/gids/${idx}\"\n \ # Check for IPv4 GID (contains ffff:)\n if [ -f \"$gid_file\" ] && grep -q \"ffff:\" \"$gid_file\"; then\n gid_value=$(cat \"$gid_file\" 2>/dev/null || echo \"\")\n echo \"[Infer RoCE] Found IPv4 RoCE v2 GID for ${hca_name}: index=${idx}, gid=${gid_value}\"\n hca_gid_index[\"${hca_name}\"]=\"${idx}\"\n \ gid_index_count[\"${idx}\"]=$((${gid_index_count[\"${idx}\"]} + 1))\n break # Use first found IPv4 GID per HCA\n fi\n \ fi\n done\n done\n\n # Find the most common GID index (most likely to be consistent across nodes)\n best_gid_index=\"\"\n \ max_count=0\n for idx in \"${!gid_index_count[@]}\"; do\n count=${gid_index_count[\"${idx}\"]}\n \ echo \"[Infer RoCE] GID_INDEX ${idx} found on ${count} HCAs\"\n if [ $count -gt $max_count ]; then\n max_count=$count\n best_gid_index=\"$idx\"\n \ fi\n done\n\n # Use deterministic fallback if counts are equal - prefer lower index number \n if [ ${#gid_index_count[@]} -gt 1 ]; then\n \ echo \"[Infer RoCE] Multiple GID indices found, selecting most common: ${best_gid_index}\"\n # If there's a tie, prefer index 3 as it's most common in SR-IOV setups\n if [ -n \"${gid_index_count['3']}\" ] && [ \"${gid_index_count['3']}\" -eq \"$max_count\" ]; then\n best_gid_index=\"3\"\n \ echo \"[Infer RoCE] Using deterministic fallback: GID_INDEX=3 (SR-IOV standard)\"\n fi\n fi\n\n # Check if GID_INDEX is already set via environment variables\n if [ -n \"${NCCL_IB_GID_INDEX}\" ]; then\n echo \"[Infer RoCE] Using pre-configured NCCL_IB_GID_INDEX=${NCCL_IB_GID_INDEX} from environment\"\n export NVSHMEM_IB_GID_INDEX=${NVSHMEM_IB_GID_INDEX:-$NCCL_IB_GID_INDEX}\n \ export UCX_IB_GID_INDEX=${UCX_IB_GID_INDEX:-$NCCL_IB_GID_INDEX}\n echo \"[Infer RoCE] Using hardcoded GID_INDEX=${NCCL_IB_GID_INDEX} for NCCL, NVSHMEM, and UCX\"\n elif [ -n \"$best_gid_index\" ]; then\n echo \"[Infer RoCE] Selected GID_INDEX: ${best_gid_index} (found on ${max_count} HCAs)\"\n\n \ export NCCL_IB_GID_INDEX=${NCCL_IB_GID_INDEX:-$best_gid_index}\n export NVSHMEM_IB_GID_INDEX=${NVSHMEM_IB_GID_INDEX:-$best_gid_index}\n export UCX_IB_GID_INDEX=${UCX_IB_GID_INDEX:-$best_gid_index}\n\n echo \"[Infer RoCE] Exported GID_INDEX=${best_gid_index} for NCCL, NVSHMEM, and UCX\"\n else\n \ echo \"[Infer RoCE] ERROR: No valid IPv4 ${KSERVE_INFER_IB_GID_INDEX_GREP} GID_INDEX found on any HCA.\"\n fi\n else\n echo \"[Infer RoCE] No active HCAs found, skipping GID_INDEX inference.\"\n fi\nfi\n\neval \"vllm serve /mnt/models \\\n --served-model-name \"{{ .Spec.Model.Name }}\" \\\n --port 8001 \\\n ${VLLM_ADDITIONAL_ARGS} \\\n --enable-ssl-refresh \\\n --ssl-certfile /var/run/kserve/tls/tls.crt \\\n --ssl-keyfile /var/run/kserve/tls/tls.key\"] command:[/bin/bash -c] env:[map[name:HOME value:/home] map[name:VLLM_LOGGING_LEVEL value:INFO] map[name:HF_HUB_CACHE value:/models]] image:registry.redhat.io/rhaiis/vllm-cuda-rhel9@sha256:fc68d623d1bfc36c8cb2fe4a71f19c8578cfb420ce8ce07b20a02c1ee0be0cf3 imagePullPolicy:IfNotPresent livenessProbe:map[failureThreshold:%!s(int64=3) httpGet:map[path:/health port:%!s(int64=8001) scheme:HTTPS] initialDelaySeconds:%!s(int64=120) periodSeconds:%!s(int64=10) timeoutSeconds:%!s(int64=10)] name:main ports:[map[containerPort:%!s(int64=8001) protocol:TCP]] readinessProbe:map[failureThreshold:%!s(int64=60) httpGet:map[path:/health port:%!s(int64=8001) scheme:HTTPS] initialDelaySeconds:%!s(int64=10) periodSeconds:%!s(int64=10) timeoutSeconds:%!s(int64=5)] securityContext:map[allowPrivilegeEscalation:%!s(bool=false) capabilities:map[drop:[ALL]] readOnlyRootFilesystem:%!s(bool=false) runAsNonRoot:%!s(bool=true) seccompProfile:map[type:RuntimeDefault]] terminationMessagePath:/dev/termination-log terminationMessagePolicy:FallbackToLogsOnError volumeMounts:[map[mountPath:/home name:home] map[mountPath:/dev/shm name:dshm] map[mountPath:/models name:model-cache] map[mountPath:/var/run/kserve/tls name:tls-certs readOnly:%!s(bool=true)]]]] initContainers:[map[args:[--port=8000 --vllm-port=8001 --connector=nixlv2 --secure-proxy=true --cert-path=/var/run/kserve/tls --decoder-use-tls=true --prefiller-use-tls=true --enable-ssrf-protection=true --pool-group=inference.networking.x-k8s.io] env:[map[name:INFERENCE_POOL_NAMESPACE valueFrom:map[fieldRef:map[fieldPath:metadata.namespace]]] map[name:SSL_CERT_DIR value:/var/run/kserve/tls:/var/run/secrets/kubernetes.io/serviceaccount:/etc/pki/tls/certs]] image:quay.io/opendatahub/llm-d-routing-sidecar:release-v0.4 imagePullPolicy:IfNotPresent livenessProbe:map[failureThreshold:%!s(int64=3) httpGet:map[path:/health port:%!s(int64=8000) scheme:HTTPS] initialDelaySeconds:%!s(int64=10) periodSeconds:%!s(int64=10) timeoutSeconds:%!s(int64=10)] name:llm-d-routing-sidecar ports:[map[containerPort:%!s(int64=8000) protocol:TCP]] readinessProbe:map[failureThreshold:%!s(int64=10) httpGet:map[path:/health port:%!s(int64=8000) scheme:HTTPS] initialDelaySeconds:%!s(int64=10) periodSeconds:%!s(int64=10) timeoutSeconds:%!s(int64=5)] resources:map[] restartPolicy:Always securityContext:map[allowPrivilegeEscalation:%!s(bool=false) capabilities:map[drop:[ALL]] readOnlyRootFilesystem:%!s(bool=false) runAsNonRoot:%!s(bool=true)] terminationMessagePath:/dev/termination-log terminationMessagePolicy:FallbackToLogsOnError volumeMounts:[map[mountPath:/var/run/kserve/tls name:tls-certs readOnly:%!s(bool=true)]]]] terminationGracePeriodSeconds:%!s(int64=30) volumes:[map[emptyDir:map[] name:home] map[emptyDir:map[medium:Memory sizeLimit:1Gi] name:dshm] map[emptyDir:map[] name:model-cache] map[name:tls-certs secret:map[secretName:{{ ChildName .ObjectMeta.Name `-kserve-self-signed-certs` }}]]]]]]}: apply failed serving.kserve.io/v1alpha1, Kind=LLMInferenceServiceConfig: unable to patch serving.kserve.io/v1alpha1, Kind=LLMInferenceServiceConfig opendatahub/v3-4-0-ea-1-kserve-config-llm-decode-template: Internal error occurred: failed calling webhook \"llminferenceserviceconfig.kserve-webhook-server.validator\": failed to call webhook: Post \"https://kserve-webhook-server-service.opendatahub.svc:443/validate-serving-kserve-io-v1alpha1-llminferenceserviceconfig?timeout=10s\": no endpoints available for service \"kserve-webhook-server-service\"" metadata: creationTimestamp: "2026-04-20T15:00:18Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:source: f:component: {} f:type: {} manager: manager operation: Update time: "2026-04-20T15:00:47Z" name: default-kserve.18a818b501b6d0cc namespace: default resourceVersion: "15778" uid: c5676f73-0436-4395-b027-0dcff82ba92b reason: ProvisioningError reportingComponent: kserve reportingInstance: "" source: component: kserve type: Warning - apiVersion: v1 count: 30 eventTime: null firstTimestamp: "2026-04-20T15:00:07Z" involvedObject: apiVersion: services.platform.opendatahub.io/v1alpha1 kind: Monitoring name: default-monitoring resourceVersion: "13674" uid: f6fcca61-23a0-4322-8aa2-f53d7a050019 kind: Event lastTimestamp: "2026-04-20T15:11:30Z" message: 'error fetching list of deployments: unable to list: opendatahub-monitoring because of unknown namespace for the cache' metadata: creationTimestamp: "2026-04-20T15:00:07Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:source: f:component: {} f:type: {} manager: manager operation: Update time: "2026-04-20T15:11:30Z" name: default-monitoring.18a818b275174546 namespace: default resourceVersion: "29563" uid: 57f6cbd7-7381-40d4-94cc-1125461ee6d1 reason: ProvisioningError reportingComponent: monitoring reportingInstance: "" source: component: monitoring type: Warning - apiVersion: v1 count: 6 eventTime: null firstTimestamp: "2026-04-20T14:53:28Z" involvedObject: kind: Node name: ip-10-0-129-82.ec2.internal uid: ip-10-0-129-82.ec2.internal kind: Event lastTimestamp: "2026-04-20T14:53:28Z" message: 'Node ip-10-0-129-82.ec2.internal status is now: NodeHasSufficientMemory' metadata: creationTimestamp: "2026-04-20T14:53:28Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:reportingInstance: {} f:source: f:component: {} f:host: {} f:type: {} manager: kubelet operation: Update time: "2026-04-20T14:53:29Z" name: ip-10-0-129-82.ec2.internal.18a818559105079d namespace: default resourceVersion: "5267" uid: d35857df-9fc6-46de-83cd-de65d7110fdd reason: NodeHasSufficientMemory reportingComponent: kubelet reportingInstance: ip-10-0-129-82.ec2.internal source: component: kubelet host: ip-10-0-129-82.ec2.internal type: Normal - apiVersion: v1 count: 6 eventTime: null firstTimestamp: "2026-04-20T14:53:28Z" involvedObject: kind: Node name: ip-10-0-129-82.ec2.internal uid: ip-10-0-129-82.ec2.internal kind: Event lastTimestamp: "2026-04-20T14:53:28Z" message: 'Node ip-10-0-129-82.ec2.internal status is now: NodeHasNoDiskPressure' metadata: creationTimestamp: "2026-04-20T14:53:28Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:reportingInstance: {} f:source: f:component: {} f:host: {} f:type: {} manager: kubelet operation: Update time: "2026-04-20T14:53:29Z" name: ip-10-0-129-82.ec2.internal.18a81855910559be namespace: default resourceVersion: "5271" uid: 96d32eb2-2666-4ba8-bc22-f2cef37c9587 reason: NodeHasNoDiskPressure reportingComponent: kubelet reportingInstance: ip-10-0-129-82.ec2.internal source: component: kubelet host: ip-10-0-129-82.ec2.internal type: Normal - apiVersion: v1 count: 6 eventTime: null firstTimestamp: "2026-04-20T14:53:28Z" involvedObject: kind: Node name: ip-10-0-129-82.ec2.internal uid: ip-10-0-129-82.ec2.internal kind: Event lastTimestamp: "2026-04-20T14:53:28Z" message: 'Node ip-10-0-129-82.ec2.internal status is now: NodeHasSufficientPID' metadata: creationTimestamp: "2026-04-20T14:53:28Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:reportingInstance: {} f:source: f:component: {} f:host: {} f:type: {} manager: kubelet operation: Update time: "2026-04-20T14:53:29Z" name: ip-10-0-129-82.ec2.internal.18a8185591058310 namespace: default resourceVersion: "5277" uid: dbfc43f9-f339-4ca3-bc50-8ecbe1cb35c0 reason: NodeHasSufficientPID reportingComponent: kubelet reportingInstance: ip-10-0-129-82.ec2.internal source: component: kubelet host: ip-10-0-129-82.ec2.internal type: Normal - apiVersion: v1 count: 1 eventTime: null firstTimestamp: "2026-04-20T14:53:28Z" involvedObject: kind: Node name: ip-10-0-129-82.ec2.internal uid: ip-10-0-129-82.ec2.internal kind: Event lastTimestamp: "2026-04-20T14:53:28Z" message: Updated Node Allocatable limit across pods metadata: creationTimestamp: "2026-04-20T14:53:28Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:reportingInstance: {} f:source: f:component: {} f:host: {} f:type: {} manager: kubelet operation: Update time: "2026-04-20T14:53:28Z" name: ip-10-0-129-82.ec2.internal.18a818559366f56f namespace: default resourceVersion: "5239" uid: 3c54a857-67f9-448a-80f4-755f1d10aea1 reason: NodeAllocatableEnforced reportingComponent: kubelet reportingInstance: ip-10-0-129-82.ec2.internal source: component: kubelet host: ip-10-0-129-82.ec2.internal type: Normal - apiVersion: v1 count: 1 eventTime: null firstTimestamp: "2026-04-20T14:53:29Z" involvedObject: apiVersion: v1 kind: Node name: ip-10-0-129-82.ec2.internal resourceVersion: "5240" uid: 0275a6d0-e4d1-4077-aade-3e2372c0f00e kind: Event lastTimestamp: "2026-04-20T14:53:29Z" message: Node synced successfully metadata: creationTimestamp: "2026-04-20T14:53:29Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:source: f:component: {} f:type: {} manager: aws-cloud-controller-manager operation: Update time: "2026-04-20T14:53:29Z" name: ip-10-0-129-82.ec2.internal.18a81855aa887e82 namespace: default resourceVersion: "5321" uid: e8e15887-57ff-46d3-a7ed-c6c4f9c9499d reason: Synced reportingComponent: cloud-node-controller reportingInstance: "" source: component: cloud-node-controller type: Normal - apiVersion: v1 count: 1 eventTime: null firstTimestamp: "2026-04-20T14:53:31Z" involvedObject: apiVersion: v1 kind: Node name: ip-10-0-129-82.ec2.internal uid: 0275a6d0-e4d1-4077-aade-3e2372c0f00e kind: Event lastTimestamp: "2026-04-20T14:53:31Z" message: 'Node ip-10-0-129-82.ec2.internal event: Registered Node ip-10-0-129-82.ec2.internal in Controller' metadata: creationTimestamp: "2026-04-20T14:53:31Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:source: f:component: {} f:type: {} manager: kube-controller-manager operation: Update time: "2026-04-20T14:53:31Z" name: ip-10-0-129-82.ec2.internal.18a8185627f98a48 namespace: default resourceVersion: "5485" uid: c53bb1fa-a104-43af-bf8d-86c57c5a1130 reason: RegisteredNode reportingComponent: node-controller reportingInstance: "" source: component: node-controller type: Normal - apiVersion: v1 count: 1 eventTime: null firstTimestamp: "2026-04-20T14:54:01Z" involvedObject: kind: Node name: ip-10-0-129-82.ec2.internal uid: ip-10-0-129-82.ec2.internal kind: Event lastTimestamp: "2026-04-20T14:54:01Z" message: 'Node ip-10-0-129-82.ec2.internal status is now: NodeReady' metadata: creationTimestamp: "2026-04-20T14:54:01Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:reportingInstance: {} f:source: f:component: {} f:host: {} f:type: {} manager: kubelet operation: Update time: "2026-04-20T14:54:01Z" name: ip-10-0-129-82.ec2.internal.18a8185d16d8c0f0 namespace: default resourceVersion: "6256" uid: c67d0c86-876e-4320-9c52-d8a2cf97f911 reason: NodeReady reportingComponent: kubelet reportingInstance: ip-10-0-129-82.ec2.internal source: component: kubelet host: ip-10-0-129-82.ec2.internal type: Normal - apiVersion: v1 count: 1 eventTime: null firstTimestamp: "2026-04-20T14:56:25Z" involvedObject: apiVersion: v1 kind: Node name: ip-10-0-129-82.ec2.internal uid: 0275a6d0-e4d1-4077-aade-3e2372c0f00e kind: Event lastTimestamp: "2026-04-20T14:56:25Z" message: 'Node ip-10-0-129-82.ec2.internal event: Registered Node ip-10-0-129-82.ec2.internal in Controller' metadata: creationTimestamp: "2026-04-20T14:56:25Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:source: f:component: {} f:type: {} manager: kube-controller-manager operation: Update time: "2026-04-20T14:56:25Z" name: ip-10-0-129-82.ec2.internal.18a8187ea041e39b namespace: default resourceVersion: "8025" uid: 86a7b3fa-33ab-491e-a314-d68c66fcd030 reason: RegisteredNode reportingComponent: node-controller reportingInstance: "" source: component: node-controller type: Normal - apiVersion: v1 count: 6 eventTime: null firstTimestamp: "2026-04-20T14:53:48Z" involvedObject: kind: Node name: ip-10-0-140-93.ec2.internal uid: ip-10-0-140-93.ec2.internal kind: Event lastTimestamp: "2026-04-20T14:53:48Z" message: 'Node ip-10-0-140-93.ec2.internal status is now: NodeHasSufficientMemory' metadata: creationTimestamp: "2026-04-20T14:53:48Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:reportingInstance: {} f:source: f:component: {} f:host: {} f:type: {} manager: kubelet operation: Update time: "2026-04-20T14:53:49Z" name: ip-10-0-140-93.ec2.internal.18a8185a372d92df namespace: default resourceVersion: "5890" uid: 45d3ad40-ebef-4db0-900e-8b3ec54c9ce1 reason: NodeHasSufficientMemory reportingComponent: kubelet reportingInstance: ip-10-0-140-93.ec2.internal source: component: kubelet host: ip-10-0-140-93.ec2.internal type: Normal - apiVersion: v1 count: 6 eventTime: null firstTimestamp: "2026-04-20T14:53:48Z" involvedObject: kind: Node name: ip-10-0-140-93.ec2.internal uid: ip-10-0-140-93.ec2.internal kind: Event lastTimestamp: "2026-04-20T14:53:48Z" message: 'Node ip-10-0-140-93.ec2.internal status is now: NodeHasNoDiskPressure' metadata: creationTimestamp: "2026-04-20T14:53:48Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:reportingInstance: {} f:source: f:component: {} f:host: {} f:type: {} manager: kubelet operation: Update time: "2026-04-20T14:53:49Z" name: ip-10-0-140-93.ec2.internal.18a8185a372dd801 namespace: default resourceVersion: "5893" uid: f35afb08-c211-4d27-8878-6ce61c0a9b3b reason: NodeHasNoDiskPressure reportingComponent: kubelet reportingInstance: ip-10-0-140-93.ec2.internal source: component: kubelet host: ip-10-0-140-93.ec2.internal type: Normal - apiVersion: v1 count: 6 eventTime: null firstTimestamp: "2026-04-20T14:53:48Z" involvedObject: kind: Node name: ip-10-0-140-93.ec2.internal uid: ip-10-0-140-93.ec2.internal kind: Event lastTimestamp: "2026-04-20T14:53:48Z" message: 'Node ip-10-0-140-93.ec2.internal status is now: NodeHasSufficientPID' metadata: creationTimestamp: "2026-04-20T14:53:48Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:reportingInstance: {} f:source: f:component: {} f:host: {} f:type: {} manager: kubelet operation: Update time: "2026-04-20T14:53:49Z" name: ip-10-0-140-93.ec2.internal.18a8185a372dfc8a namespace: default resourceVersion: "5898" uid: 98f4c98d-ff1b-4cf6-9900-8f6486c2e66f reason: NodeHasSufficientPID reportingComponent: kubelet reportingInstance: ip-10-0-140-93.ec2.internal source: component: kubelet host: ip-10-0-140-93.ec2.internal type: Normal - apiVersion: v1 count: 1 eventTime: null firstTimestamp: "2026-04-20T14:53:48Z" involvedObject: kind: Node name: ip-10-0-140-93.ec2.internal uid: ip-10-0-140-93.ec2.internal kind: Event lastTimestamp: "2026-04-20T14:53:48Z" message: Updated Node Allocatable limit across pods metadata: creationTimestamp: "2026-04-20T14:53:48Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:reportingInstance: {} f:source: f:component: {} f:host: {} f:type: {} manager: kubelet operation: Update time: "2026-04-20T14:53:48Z" name: ip-10-0-140-93.ec2.internal.18a8185a39d118e2 namespace: default resourceVersion: "5789" uid: 1e92aaea-3cd3-4b7e-b3f8-6092c2fac885 reason: NodeAllocatableEnforced reportingComponent: kubelet reportingInstance: ip-10-0-140-93.ec2.internal source: component: kubelet host: ip-10-0-140-93.ec2.internal type: Normal - apiVersion: v1 count: 1 eventTime: null firstTimestamp: "2026-04-20T14:53:49Z" involvedObject: apiVersion: v1 kind: Node name: ip-10-0-140-93.ec2.internal resourceVersion: "5793" uid: efb3ff41-3553-45ef-b471-56e057f1df08 kind: Event lastTimestamp: "2026-04-20T14:53:49Z" message: Node synced successfully metadata: creationTimestamp: "2026-04-20T14:53:49Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:source: f:component: {} f:type: {} manager: aws-cloud-controller-manager operation: Update time: "2026-04-20T14:53:49Z" name: ip-10-0-140-93.ec2.internal.18a8185a4ea204b0 namespace: default resourceVersion: "5910" uid: 5b6ed330-7fc5-4414-9c80-7259e5c33bc8 reason: Synced reportingComponent: cloud-node-controller reportingInstance: "" source: component: cloud-node-controller type: Normal - apiVersion: v1 count: 1 eventTime: null firstTimestamp: "2026-04-20T14:53:51Z" involvedObject: apiVersion: v1 kind: Node name: ip-10-0-140-93.ec2.internal uid: efb3ff41-3553-45ef-b471-56e057f1df08 kind: Event lastTimestamp: "2026-04-20T14:53:51Z" message: 'Node ip-10-0-140-93.ec2.internal event: Registered Node ip-10-0-140-93.ec2.internal in Controller' metadata: creationTimestamp: "2026-04-20T14:53:51Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:source: f:component: {} f:type: {} manager: kube-controller-manager operation: Update time: "2026-04-20T14:53:51Z" name: ip-10-0-140-93.ec2.internal.18a8185ad049ace0 namespace: default resourceVersion: "6038" uid: 1e94a10a-e90f-4846-b6ee-27e0991575c7 reason: RegisteredNode reportingComponent: node-controller reportingInstance: "" source: component: node-controller type: Normal - apiVersion: v1 count: 1 eventTime: null firstTimestamp: "2026-04-20T14:54:22Z" involvedObject: kind: Node name: ip-10-0-140-93.ec2.internal uid: ip-10-0-140-93.ec2.internal kind: Event lastTimestamp: "2026-04-20T14:54:22Z" message: 'Node ip-10-0-140-93.ec2.internal status is now: NodeReady' metadata: creationTimestamp: "2026-04-20T14:54:22Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:reportingInstance: {} f:source: f:component: {} f:host: {} f:type: {} manager: kubelet operation: Update time: "2026-04-20T14:54:22Z" name: ip-10-0-140-93.ec2.internal.18a81861fe5c0c92 namespace: default resourceVersion: "6698" uid: 68ca8f01-c640-409c-9a23-3e3887cd2e28 reason: NodeReady reportingComponent: kubelet reportingInstance: ip-10-0-140-93.ec2.internal source: component: kubelet host: ip-10-0-140-93.ec2.internal type: Normal - apiVersion: v1 count: 1 eventTime: null firstTimestamp: "2026-04-20T14:56:25Z" involvedObject: apiVersion: v1 kind: Node name: ip-10-0-140-93.ec2.internal uid: efb3ff41-3553-45ef-b471-56e057f1df08 kind: Event lastTimestamp: "2026-04-20T14:56:25Z" message: 'Node ip-10-0-140-93.ec2.internal event: Registered Node ip-10-0-140-93.ec2.internal in Controller' metadata: creationTimestamp: "2026-04-20T14:56:25Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:source: f:component: {} f:type: {} manager: kube-controller-manager operation: Update time: "2026-04-20T14:56:25Z" name: ip-10-0-140-93.ec2.internal.18a8187ea0425c96 namespace: default resourceVersion: "8029" uid: f92e6d2a-2cb0-497a-98c1-103254de37ff reason: RegisteredNode reportingComponent: node-controller reportingInstance: "" source: component: node-controller type: Normal - apiVersion: v1 count: 1 eventTime: null firstTimestamp: "2026-04-20T14:53:30Z" involvedObject: kind: Node name: ip-10-0-142-255.ec2.internal uid: ip-10-0-142-255.ec2.internal kind: Event lastTimestamp: "2026-04-20T14:53:30Z" message: Starting kubelet. metadata: creationTimestamp: "2026-04-20T14:53:30Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:reportingInstance: {} f:source: f:component: {} f:host: {} f:type: {} manager: kubelet operation: Update time: "2026-04-20T14:53:30Z" name: ip-10-0-142-255.ec2.internal.18a81855fcf2e186 namespace: default resourceVersion: "5367" uid: ab7ec92c-ce7d-4a4b-a863-496bd48d59d1 reason: Starting reportingComponent: kubelet reportingInstance: ip-10-0-142-255.ec2.internal source: component: kubelet host: ip-10-0-142-255.ec2.internal type: Normal - apiVersion: v1 count: 6 eventTime: null firstTimestamp: "2026-04-20T14:53:30Z" involvedObject: kind: Node name: ip-10-0-142-255.ec2.internal uid: ip-10-0-142-255.ec2.internal kind: Event lastTimestamp: "2026-04-20T14:53:30Z" message: 'Node ip-10-0-142-255.ec2.internal status is now: NodeHasSufficientMemory' metadata: creationTimestamp: "2026-04-20T14:53:30Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:reportingInstance: {} f:source: f:component: {} f:host: {} f:type: {} manager: kubelet operation: Update time: "2026-04-20T14:53:30Z" name: ip-10-0-142-255.ec2.internal.18a81855febffb49 namespace: default resourceVersion: "5462" uid: 601155f7-b4b3-4e47-9320-b77eda4202dd reason: NodeHasSufficientMemory reportingComponent: kubelet reportingInstance: ip-10-0-142-255.ec2.internal source: component: kubelet host: ip-10-0-142-255.ec2.internal type: Normal - apiVersion: v1 count: 6 eventTime: null firstTimestamp: "2026-04-20T14:53:30Z" involvedObject: kind: Node name: ip-10-0-142-255.ec2.internal uid: ip-10-0-142-255.ec2.internal kind: Event lastTimestamp: "2026-04-20T14:53:30Z" message: 'Node ip-10-0-142-255.ec2.internal status is now: NodeHasNoDiskPressure' metadata: creationTimestamp: "2026-04-20T14:53:30Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:reportingInstance: {} f:source: f:component: {} f:host: {} f:type: {} manager: kubelet operation: Update time: "2026-04-20T14:53:31Z" name: ip-10-0-142-255.ec2.internal.18a81855fec041ca namespace: default resourceVersion: "5470" uid: 5f4ca21e-7760-43b9-9575-98a157fa5f79 reason: NodeHasNoDiskPressure reportingComponent: kubelet reportingInstance: ip-10-0-142-255.ec2.internal source: component: kubelet host: ip-10-0-142-255.ec2.internal type: Normal - apiVersion: v1 count: 6 eventTime: null firstTimestamp: "2026-04-20T14:53:30Z" involvedObject: kind: Node name: ip-10-0-142-255.ec2.internal uid: ip-10-0-142-255.ec2.internal kind: Event lastTimestamp: "2026-04-20T14:53:30Z" message: 'Node ip-10-0-142-255.ec2.internal status is now: NodeHasSufficientPID' metadata: creationTimestamp: "2026-04-20T14:53:30Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:reportingInstance: {} f:source: f:component: {} f:host: {} f:type: {} manager: kubelet operation: Update time: "2026-04-20T14:53:31Z" name: ip-10-0-142-255.ec2.internal.18a81855fec06661 namespace: default resourceVersion: "5474" uid: 06e958c6-e96b-4b7f-bea3-ca9f2e2d6814 reason: NodeHasSufficientPID reportingComponent: kubelet reportingInstance: ip-10-0-142-255.ec2.internal source: component: kubelet host: ip-10-0-142-255.ec2.internal type: Normal - apiVersion: v1 count: 1 eventTime: null firstTimestamp: "2026-04-20T14:53:30Z" involvedObject: kind: Node name: ip-10-0-142-255.ec2.internal uid: ip-10-0-142-255.ec2.internal kind: Event lastTimestamp: "2026-04-20T14:53:30Z" message: Updated Node Allocatable limit across pods metadata: creationTimestamp: "2026-04-20T14:53:30Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:reportingInstance: {} f:source: f:component: {} f:host: {} f:type: {} manager: kubelet operation: Update time: "2026-04-20T14:53:30Z" name: ip-10-0-142-255.ec2.internal.18a818560151724e namespace: default resourceVersion: "5371" uid: 53f6ccab-42ee-4c7c-b031-6a31cda8d700 reason: NodeAllocatableEnforced reportingComponent: kubelet reportingInstance: ip-10-0-142-255.ec2.internal source: component: kubelet host: ip-10-0-142-255.ec2.internal type: Normal - apiVersion: v1 count: 1 eventTime: null firstTimestamp: "2026-04-20T14:53:31Z" involvedObject: apiVersion: v1 kind: Node name: ip-10-0-142-255.ec2.internal resourceVersion: "5376" uid: 25c93488-b1ca-44d3-ae26-d013a68d729f kind: Event lastTimestamp: "2026-04-20T14:53:31Z" message: Node synced successfully metadata: creationTimestamp: "2026-04-20T14:53:31Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:source: f:component: {} f:type: {} manager: aws-cloud-controller-manager operation: Update time: "2026-04-20T14:53:31Z" name: ip-10-0-142-255.ec2.internal.18a818561720a981 namespace: default resourceVersion: "5476" uid: dc94acdc-562b-4951-9957-9ff0d3b16d96 reason: Synced reportingComponent: cloud-node-controller reportingInstance: "" source: component: cloud-node-controller type: Normal - apiVersion: v1 count: 1 eventTime: null firstTimestamp: "2026-04-20T14:53:31Z" involvedObject: apiVersion: v1 kind: Node name: ip-10-0-142-255.ec2.internal uid: 25c93488-b1ca-44d3-ae26-d013a68d729f kind: Event lastTimestamp: "2026-04-20T14:53:31Z" message: 'Node ip-10-0-142-255.ec2.internal event: Registered Node ip-10-0-142-255.ec2.internal in Controller' metadata: creationTimestamp: "2026-04-20T14:53:31Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:source: f:component: {} f:type: {} manager: kube-controller-manager operation: Update time: "2026-04-20T14:53:31Z" name: ip-10-0-142-255.ec2.internal.18a8185627f80ec7 namespace: default resourceVersion: "5484" uid: 67a396ba-2145-4804-86bb-2af737028eb9 reason: RegisteredNode reportingComponent: node-controller reportingInstance: "" source: component: node-controller type: Normal - apiVersion: v1 count: 1 eventTime: null firstTimestamp: "2026-04-20T14:54:04Z" involvedObject: kind: Node name: ip-10-0-142-255.ec2.internal uid: ip-10-0-142-255.ec2.internal kind: Event lastTimestamp: "2026-04-20T14:54:04Z" message: 'Node ip-10-0-142-255.ec2.internal status is now: NodeReady' metadata: creationTimestamp: "2026-04-20T14:54:04Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:reportingInstance: {} f:source: f:component: {} f:host: {} f:type: {} manager: kubelet operation: Update time: "2026-04-20T14:54:04Z" name: ip-10-0-142-255.ec2.internal.18a8185dc4ab5c89 namespace: default resourceVersion: "6330" uid: a7328140-08fd-4983-86cc-bdab77cdc4a2 reason: NodeReady reportingComponent: kubelet reportingInstance: ip-10-0-142-255.ec2.internal source: component: kubelet host: ip-10-0-142-255.ec2.internal type: Normal - apiVersion: v1 count: 1 eventTime: null firstTimestamp: "2026-04-20T14:56:25Z" involvedObject: apiVersion: v1 kind: Node name: ip-10-0-142-255.ec2.internal uid: 25c93488-b1ca-44d3-ae26-d013a68d729f kind: Event lastTimestamp: "2026-04-20T14:56:25Z" message: 'Node ip-10-0-142-255.ec2.internal event: Registered Node ip-10-0-142-255.ec2.internal in Controller' metadata: creationTimestamp: "2026-04-20T14:56:25Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:source: f:component: {} f:type: {} manager: kube-controller-manager operation: Update time: "2026-04-20T14:56:25Z" name: ip-10-0-142-255.ec2.internal.18a8187ea0426d1c namespace: default resourceVersion: "8040" uid: 53c47ab0-9cd2-4d79-b5cb-9c081b048083 reason: RegisteredNode reportingComponent: node-controller reportingInstance: "" source: component: node-controller type: Normal - apiVersion: v1 count: 9 eventTime: null firstTimestamp: "2026-04-20T15:01:20Z" involvedObject: apiVersion: v1 kind: Namespace name: kuadrant-system resourceVersion: "16032" uid: b634b0ad-89d2-4d0a-a832-a3ffc66dd2aa kind: Event lastTimestamp: "2026-04-20T15:01:23Z" message: 'error using catalogsource kuadrant-system/kuadrant-operator-catalog: failed to list bundles: rpc error: code = Unavailable desc = connection error: desc = "error reading server preface: http2: frame too large"' metadata: creationTimestamp: "2026-04-20T15:01:20Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:count: {} f:firstTimestamp: {} f:involvedObject: {} f:lastTimestamp: {} f:message: {} f:reason: {} f:reportingComponent: {} f:source: f:component: {} f:type: {} manager: catalog operation: Update time: "2026-04-20T15:01:25Z" name: kuadrant-system.18a818c36764e04e namespace: default resourceVersion: "16275" uid: 984016bc-ca38-4b6b-a694-fb16a58e72d4 reason: ResolutionFailed reportingComponent: operator-lifecycle-manager reportingInstance: "" source: component: operator-lifecycle-manager type: Warning - apiVersion: v1 eventTime: null firstTimestamp: null involvedObject: apiVersion: v1 kind: Namespace name: kube-system namespace: default kind: Event lastTimestamp: null message: Received signal to terminate, becoming unready, but keeping serving metadata: creationTimestamp: "2026-04-20T14:56:05Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:involvedObject: {} f:message: {} f:reason: {} f:source: f:component: {} f:host: {} f:type: {} manager: openshift-apiserver operation: Update time: "2026-04-20T14:56:05Z" name: kube-system.18a8187a1fa20f88 namespace: default resourceVersion: "7629" uid: 503ea171-a7ef-4c02-924f-b9f27dc1a19d reason: TerminationStart reportingComponent: "" reportingInstance: "" source: component: apiserver host: openshift-apiserver-6675f54895-jjckv type: Normal - apiVersion: v1 eventTime: null firstTimestamp: null involvedObject: apiVersion: v1 kind: Namespace name: kube-system namespace: default kind: Event lastTimestamp: null message: The minimal shutdown duration of 15s finished metadata: creationTimestamp: "2026-04-20T14:56:20Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:involvedObject: {} f:message: {} f:reason: {} f:source: f:component: {} f:host: {} f:type: {} manager: openshift-apiserver operation: Update time: "2026-04-20T14:56:20Z" name: kube-system.18a8187d9e20a86a namespace: default resourceVersion: "7909" uid: 1c53727f-cb12-4483-aa9e-ae22e5bdb30f reason: TerminationMinimalShutdownDurationFinished reportingComponent: "" reportingInstance: "" source: component: apiserver host: openshift-apiserver-6675f54895-jjckv type: Normal - apiVersion: v1 eventTime: null firstTimestamp: null involvedObject: apiVersion: v1 kind: Namespace name: kube-system namespace: default kind: Event lastTimestamp: null message: Server has stopped listening metadata: creationTimestamp: "2026-04-20T14:56:20Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:involvedObject: {} f:message: {} f:reason: {} f:source: f:component: {} f:host: {} f:type: {} manager: openshift-apiserver operation: Update time: "2026-04-20T14:56:20Z" name: kube-system.18a8187d9e530b77 namespace: default resourceVersion: "7910" uid: 6286625f-009b-4bc2-acd5-55175ded3e8e reason: TerminationStoppedServing reportingComponent: "" reportingInstance: "" source: component: apiserver host: openshift-apiserver-6675f54895-jjckv type: Normal - apiVersion: v1 eventTime: null firstTimestamp: null involvedObject: apiVersion: v1 kind: Namespace name: kube-system namespace: default kind: Event lastTimestamp: null message: All pre-shutdown hooks have been finished metadata: creationTimestamp: "2026-04-20T14:56:20Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:involvedObject: {} f:message: {} f:reason: {} f:source: f:component: {} f:host: {} f:type: {} manager: openshift-apiserver operation: Update time: "2026-04-20T14:56:20Z" name: kube-system.18a8187d9ea98c06 namespace: default resourceVersion: "7911" uid: 4a96afec-6fac-4814-b750-a9a955efbc27 reason: TerminationPreShutdownHooksFinished reportingComponent: "" reportingInstance: "" source: component: apiserver host: openshift-apiserver-6675f54895-jjckv type: Normal - apiVersion: v1 eventTime: null firstTimestamp: null involvedObject: apiVersion: v1 kind: Namespace name: kube-system namespace: default kind: Event lastTimestamp: null message: All pending requests processed metadata: creationTimestamp: "2026-04-20T14:57:20Z" managedFields: - apiVersion: v1 fieldsType: FieldsV1 fieldsV1: f:involvedObject: {} f:message: {} f:reason: {} f:source: f:component: {} f:host: {} f:type: {} manager: openshift-apiserver operation: Update time: "2026-04-20T14:57:20Z" name: kube-system.18a8188b969f8cc0 namespace: default resourceVersion: "10101" uid: cedac90a-5fdb-43c7-b5ef-83ff5fc1cf51 reason: TerminationGracefulTerminationFinished reportingComponent: "" reportingInstance: "" source: component: apiserver host: openshift-apiserver-6675f54895-jjckv type: Normal - apiVersion: v1 count: 1 eventTime: null firstTimestamp: "2026-04-20T14:49:08Z" involvedObject: apiVersion: v1 kind: Namespace name: openshift-kube-apiserver namespace: default kind: Event lastTimestamp: "2026-04-20T14:49:08Z" message: readyz=true metadata: creationTimestamp: "2026-04-20T14:49:08Z" name: openshift-kube-apiserver.18a81818e5e304b6 namespace: default resourceVersion: "274" uid: 28a4ffa4-ffd2-4c43-a86f-cf1c94221aa0 reason: KubeAPIReadyz reportingComponent: "" reportingInstance: "" source: component: apiserver host: kube-apiserver-546c4b559c-zvm2b type: Warning kind: EventList metadata: resourceVersion: "45734"