Apr 22 17:32:13.115984 ip-10-0-138-57 systemd[1]: kubelet.service: Failed to load environment files: No such file or directory Apr 22 17:32:13.115998 ip-10-0-138-57 systemd[1]: kubelet.service: Failed to run 'start-pre' task: No such file or directory Apr 22 17:32:13.116007 ip-10-0-138-57 systemd[1]: kubelet.service: Failed with result 'resources'. Apr 22 17:32:13.116335 ip-10-0-138-57 systemd[1]: Failed to start Kubernetes Kubelet. Apr 22 17:32:23.241284 ip-10-0-138-57 systemd[1]: kubelet.service: Failed to schedule restart job: Unit crio.service not found. Apr 22 17:32:23.241306 ip-10-0-138-57 systemd[1]: kubelet.service: Failed with result 'resources'. -- Boot 8bf8db5aa5a04767b101688c093bacc5 -- Apr 22 17:34:51.991211 ip-10-0-138-57 systemd[1]: Starting Kubernetes Kubelet... Apr 22 17:34:52.408833 ip-10-0-138-57 kubenswrapper[2581]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 22 17:34:52.408833 ip-10-0-138-57 kubenswrapper[2581]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Apr 22 17:34:52.408833 ip-10-0-138-57 kubenswrapper[2581]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 22 17:34:52.408833 ip-10-0-138-57 kubenswrapper[2581]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Apr 22 17:34:52.408833 ip-10-0-138-57 kubenswrapper[2581]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 22 17:34:52.410233 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.410146 2581 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 22 17:34:52.413059 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413040 2581 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 22 17:34:52.413059 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413058 2581 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 22 17:34:52.413128 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413061 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 22 17:34:52.413128 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413065 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 22 17:34:52.413128 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413067 2581 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 22 17:34:52.413128 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413071 2581 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 22 17:34:52.413128 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413073 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 22 17:34:52.413128 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413076 2581 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 22 17:34:52.413128 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413079 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 22 17:34:52.413128 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413082 2581 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 22 17:34:52.413128 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413084 2581 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 22 17:34:52.413128 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413087 2581 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 22 17:34:52.413128 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413090 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 22 17:34:52.413128 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413093 2581 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 22 17:34:52.413128 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413101 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 22 17:34:52.413128 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413104 2581 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 22 17:34:52.413128 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413107 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 22 17:34:52.413128 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413109 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 22 17:34:52.413128 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413112 2581 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 22 17:34:52.413128 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413114 2581 feature_gate.go:328] unrecognized feature gate: Example2 Apr 22 17:34:52.413128 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413117 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 22 17:34:52.413606 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413119 2581 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 22 17:34:52.413606 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413121 2581 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 22 17:34:52.413606 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413124 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 22 17:34:52.413606 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413127 2581 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 22 17:34:52.413606 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413129 2581 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 22 17:34:52.413606 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413132 2581 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 22 17:34:52.413606 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413136 2581 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 22 17:34:52.413606 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413139 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 22 17:34:52.413606 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413142 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 22 17:34:52.413606 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413145 2581 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 22 17:34:52.413606 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413147 2581 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 22 17:34:52.413606 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413149 2581 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 22 17:34:52.413606 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413152 2581 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 22 17:34:52.413606 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413154 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 22 17:34:52.413606 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413157 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 22 17:34:52.413606 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413159 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 22 17:34:52.413606 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413163 2581 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 22 17:34:52.413606 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413167 2581 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 22 17:34:52.413606 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413170 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 22 17:34:52.413606 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413172 2581 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 22 17:34:52.414108 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413174 2581 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 22 17:34:52.414108 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413177 2581 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 22 17:34:52.414108 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413179 2581 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 22 17:34:52.414108 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413181 2581 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 22 17:34:52.414108 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413184 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 22 17:34:52.414108 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413188 2581 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 22 17:34:52.414108 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413190 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 22 17:34:52.414108 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413193 2581 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 22 17:34:52.414108 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413195 2581 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 22 17:34:52.414108 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413198 2581 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 22 17:34:52.414108 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413200 2581 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 22 17:34:52.414108 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413203 2581 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 22 17:34:52.414108 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413205 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 22 17:34:52.414108 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413209 2581 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 22 17:34:52.414108 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413212 2581 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 22 17:34:52.414108 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413214 2581 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 22 17:34:52.414108 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413217 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 22 17:34:52.414108 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413219 2581 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 22 17:34:52.414108 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413222 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 22 17:34:52.414108 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413225 2581 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 22 17:34:52.414588 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413227 2581 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 22 17:34:52.414588 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413230 2581 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 22 17:34:52.414588 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413232 2581 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 22 17:34:52.414588 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413235 2581 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 22 17:34:52.414588 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413237 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 22 17:34:52.414588 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413240 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 22 17:34:52.414588 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413242 2581 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 22 17:34:52.414588 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413244 2581 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 22 17:34:52.414588 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413247 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 22 17:34:52.414588 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413249 2581 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 22 17:34:52.414588 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413251 2581 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 22 17:34:52.414588 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413254 2581 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 22 17:34:52.414588 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413257 2581 feature_gate.go:328] unrecognized feature gate: Example Apr 22 17:34:52.414588 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413259 2581 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 22 17:34:52.414588 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413263 2581 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 22 17:34:52.414588 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413266 2581 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 22 17:34:52.414588 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413268 2581 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 22 17:34:52.414588 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413271 2581 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 22 17:34:52.414588 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413273 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 22 17:34:52.414588 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413277 2581 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 22 17:34:52.415179 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413281 2581 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 22 17:34:52.415179 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413284 2581 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 22 17:34:52.415179 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413287 2581 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 22 17:34:52.415179 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413289 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 22 17:34:52.415179 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413292 2581 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 22 17:34:52.415179 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413695 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 22 17:34:52.415179 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413700 2581 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 22 17:34:52.415179 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413703 2581 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 22 17:34:52.415179 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413705 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 22 17:34:52.415179 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413710 2581 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 22 17:34:52.415179 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413713 2581 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 22 17:34:52.415179 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413716 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 22 17:34:52.415179 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413719 2581 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 22 17:34:52.415179 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413722 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 22 17:34:52.415179 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413725 2581 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 22 17:34:52.415179 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413727 2581 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 22 17:34:52.415179 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413730 2581 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 22 17:34:52.415179 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413732 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 22 17:34:52.415179 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413735 2581 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 22 17:34:52.415179 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413737 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 22 17:34:52.415670 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413740 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 22 17:34:52.415670 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413742 2581 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 22 17:34:52.415670 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413745 2581 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 22 17:34:52.415670 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413747 2581 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 22 17:34:52.415670 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413750 2581 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 22 17:34:52.415670 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413752 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 22 17:34:52.415670 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413755 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 22 17:34:52.415670 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413757 2581 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 22 17:34:52.415670 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413760 2581 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 22 17:34:52.415670 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413763 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 22 17:34:52.415670 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413765 2581 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 22 17:34:52.415670 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413767 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 22 17:34:52.415670 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413770 2581 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 22 17:34:52.415670 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413772 2581 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 22 17:34:52.415670 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413774 2581 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 22 17:34:52.415670 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413777 2581 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 22 17:34:52.415670 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413779 2581 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 22 17:34:52.415670 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413787 2581 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 22 17:34:52.415670 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413790 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 22 17:34:52.415670 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413792 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 22 17:34:52.416189 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413795 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 22 17:34:52.416189 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413797 2581 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 22 17:34:52.416189 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413799 2581 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 22 17:34:52.416189 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413802 2581 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 22 17:34:52.416189 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413806 2581 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 22 17:34:52.416189 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413809 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 22 17:34:52.416189 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413812 2581 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 22 17:34:52.416189 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413815 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 22 17:34:52.416189 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413817 2581 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 22 17:34:52.416189 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413820 2581 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 22 17:34:52.416189 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413822 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 22 17:34:52.416189 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413825 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 22 17:34:52.416189 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413827 2581 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 22 17:34:52.416189 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413829 2581 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 22 17:34:52.416189 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413833 2581 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 22 17:34:52.416189 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413835 2581 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 22 17:34:52.416189 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413838 2581 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 22 17:34:52.416189 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413840 2581 feature_gate.go:328] unrecognized feature gate: Example2 Apr 22 17:34:52.416189 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413843 2581 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 22 17:34:52.416661 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413845 2581 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 22 17:34:52.416661 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413848 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 22 17:34:52.416661 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413850 2581 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 22 17:34:52.416661 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413852 2581 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 22 17:34:52.416661 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413855 2581 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 22 17:34:52.416661 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413857 2581 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 22 17:34:52.416661 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413860 2581 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 22 17:34:52.416661 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413863 2581 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 22 17:34:52.416661 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413865 2581 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 22 17:34:52.416661 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413868 2581 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 22 17:34:52.416661 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413888 2581 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 22 17:34:52.416661 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413891 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 22 17:34:52.416661 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413894 2581 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 22 17:34:52.416661 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413896 2581 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 22 17:34:52.416661 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413899 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 22 17:34:52.416661 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413901 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 22 17:34:52.416661 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413904 2581 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 22 17:34:52.416661 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413906 2581 feature_gate.go:328] unrecognized feature gate: Example Apr 22 17:34:52.416661 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413909 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 22 17:34:52.416661 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413912 2581 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 22 17:34:52.417163 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413915 2581 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 22 17:34:52.417163 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413917 2581 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 22 17:34:52.417163 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413920 2581 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 22 17:34:52.417163 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413922 2581 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 22 17:34:52.417163 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413925 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 22 17:34:52.417163 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413927 2581 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 22 17:34:52.417163 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413929 2581 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 22 17:34:52.417163 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413932 2581 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 22 17:34:52.417163 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413934 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 22 17:34:52.417163 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413937 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 22 17:34:52.417163 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413940 2581 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 22 17:34:52.417163 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.413943 2581 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 22 17:34:52.417163 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415318 2581 flags.go:64] FLAG: --address="0.0.0.0" Apr 22 17:34:52.417163 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415329 2581 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Apr 22 17:34:52.417163 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415335 2581 flags.go:64] FLAG: --anonymous-auth="true" Apr 22 17:34:52.417163 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415340 2581 flags.go:64] FLAG: --application-metrics-count-limit="100" Apr 22 17:34:52.417163 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415345 2581 flags.go:64] FLAG: --authentication-token-webhook="false" Apr 22 17:34:52.417163 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415348 2581 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Apr 22 17:34:52.417163 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415353 2581 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Apr 22 17:34:52.417163 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415358 2581 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Apr 22 17:34:52.417656 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415361 2581 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Apr 22 17:34:52.417656 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415364 2581 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Apr 22 17:34:52.417656 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415369 2581 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Apr 22 17:34:52.417656 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415372 2581 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Apr 22 17:34:52.417656 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415375 2581 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Apr 22 17:34:52.417656 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415378 2581 flags.go:64] FLAG: --cgroup-root="" Apr 22 17:34:52.417656 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415381 2581 flags.go:64] FLAG: --cgroups-per-qos="true" Apr 22 17:34:52.417656 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415384 2581 flags.go:64] FLAG: --client-ca-file="" Apr 22 17:34:52.417656 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415387 2581 flags.go:64] FLAG: --cloud-config="" Apr 22 17:34:52.417656 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415389 2581 flags.go:64] FLAG: --cloud-provider="external" Apr 22 17:34:52.417656 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415392 2581 flags.go:64] FLAG: --cluster-dns="[]" Apr 22 17:34:52.417656 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415399 2581 flags.go:64] FLAG: --cluster-domain="" Apr 22 17:34:52.417656 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415402 2581 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Apr 22 17:34:52.417656 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415405 2581 flags.go:64] FLAG: --config-dir="" Apr 22 17:34:52.417656 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415408 2581 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Apr 22 17:34:52.417656 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415411 2581 flags.go:64] FLAG: --container-log-max-files="5" Apr 22 17:34:52.417656 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415416 2581 flags.go:64] FLAG: --container-log-max-size="10Mi" Apr 22 17:34:52.417656 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415418 2581 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Apr 22 17:34:52.417656 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415422 2581 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Apr 22 17:34:52.417656 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415425 2581 flags.go:64] FLAG: --containerd-namespace="k8s.io" Apr 22 17:34:52.417656 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415427 2581 flags.go:64] FLAG: --contention-profiling="false" Apr 22 17:34:52.417656 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415430 2581 flags.go:64] FLAG: --cpu-cfs-quota="true" Apr 22 17:34:52.417656 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415433 2581 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Apr 22 17:34:52.417656 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415437 2581 flags.go:64] FLAG: --cpu-manager-policy="none" Apr 22 17:34:52.417656 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415440 2581 flags.go:64] FLAG: --cpu-manager-policy-options="" Apr 22 17:34:52.418253 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415444 2581 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Apr 22 17:34:52.418253 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415447 2581 flags.go:64] FLAG: --enable-controller-attach-detach="true" Apr 22 17:34:52.418253 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415450 2581 flags.go:64] FLAG: --enable-debugging-handlers="true" Apr 22 17:34:52.418253 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415453 2581 flags.go:64] FLAG: --enable-load-reader="false" Apr 22 17:34:52.418253 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415455 2581 flags.go:64] FLAG: --enable-server="true" Apr 22 17:34:52.418253 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415458 2581 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Apr 22 17:34:52.418253 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415462 2581 flags.go:64] FLAG: --event-burst="100" Apr 22 17:34:52.418253 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415465 2581 flags.go:64] FLAG: --event-qps="50" Apr 22 17:34:52.418253 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415468 2581 flags.go:64] FLAG: --event-storage-age-limit="default=0" Apr 22 17:34:52.418253 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415472 2581 flags.go:64] FLAG: --event-storage-event-limit="default=0" Apr 22 17:34:52.418253 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415476 2581 flags.go:64] FLAG: --eviction-hard="" Apr 22 17:34:52.418253 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415480 2581 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Apr 22 17:34:52.418253 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415482 2581 flags.go:64] FLAG: --eviction-minimum-reclaim="" Apr 22 17:34:52.418253 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415486 2581 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Apr 22 17:34:52.418253 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415489 2581 flags.go:64] FLAG: --eviction-soft="" Apr 22 17:34:52.418253 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415492 2581 flags.go:64] FLAG: --eviction-soft-grace-period="" Apr 22 17:34:52.418253 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415495 2581 flags.go:64] FLAG: --exit-on-lock-contention="false" Apr 22 17:34:52.418253 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415498 2581 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Apr 22 17:34:52.418253 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415501 2581 flags.go:64] FLAG: --experimental-mounter-path="" Apr 22 17:34:52.418253 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415504 2581 flags.go:64] FLAG: --fail-cgroupv1="false" Apr 22 17:34:52.418253 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415507 2581 flags.go:64] FLAG: --fail-swap-on="true" Apr 22 17:34:52.418253 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415510 2581 flags.go:64] FLAG: --feature-gates="" Apr 22 17:34:52.418253 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415514 2581 flags.go:64] FLAG: --file-check-frequency="20s" Apr 22 17:34:52.418253 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415517 2581 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Apr 22 17:34:52.418253 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415520 2581 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Apr 22 17:34:52.418852 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415523 2581 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Apr 22 17:34:52.418852 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415526 2581 flags.go:64] FLAG: --healthz-port="10248" Apr 22 17:34:52.418852 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415529 2581 flags.go:64] FLAG: --help="false" Apr 22 17:34:52.418852 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415532 2581 flags.go:64] FLAG: --hostname-override="ip-10-0-138-57.ec2.internal" Apr 22 17:34:52.418852 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415535 2581 flags.go:64] FLAG: --housekeeping-interval="10s" Apr 22 17:34:52.418852 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415538 2581 flags.go:64] FLAG: --http-check-frequency="20s" Apr 22 17:34:52.418852 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415542 2581 flags.go:64] FLAG: --image-credential-provider-bin-dir="/usr/libexec/kubelet-image-credential-provider-plugins" Apr 22 17:34:52.418852 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415546 2581 flags.go:64] FLAG: --image-credential-provider-config="/etc/kubernetes/credential-providers/ecr-credential-provider.yaml" Apr 22 17:34:52.418852 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415549 2581 flags.go:64] FLAG: --image-gc-high-threshold="85" Apr 22 17:34:52.418852 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415552 2581 flags.go:64] FLAG: --image-gc-low-threshold="80" Apr 22 17:34:52.418852 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415555 2581 flags.go:64] FLAG: --image-service-endpoint="" Apr 22 17:34:52.418852 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415558 2581 flags.go:64] FLAG: --kernel-memcg-notification="false" Apr 22 17:34:52.418852 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415561 2581 flags.go:64] FLAG: --kube-api-burst="100" Apr 22 17:34:52.418852 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415563 2581 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Apr 22 17:34:52.418852 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415566 2581 flags.go:64] FLAG: --kube-api-qps="50" Apr 22 17:34:52.418852 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415569 2581 flags.go:64] FLAG: --kube-reserved="" Apr 22 17:34:52.418852 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415573 2581 flags.go:64] FLAG: --kube-reserved-cgroup="" Apr 22 17:34:52.418852 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415576 2581 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Apr 22 17:34:52.418852 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415578 2581 flags.go:64] FLAG: --kubelet-cgroups="" Apr 22 17:34:52.418852 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415582 2581 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Apr 22 17:34:52.418852 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415585 2581 flags.go:64] FLAG: --lock-file="" Apr 22 17:34:52.418852 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415588 2581 flags.go:64] FLAG: --log-cadvisor-usage="false" Apr 22 17:34:52.418852 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415591 2581 flags.go:64] FLAG: --log-flush-frequency="5s" Apr 22 17:34:52.418852 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415594 2581 flags.go:64] FLAG: --log-json-info-buffer-size="0" Apr 22 17:34:52.419485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415599 2581 flags.go:64] FLAG: --log-json-split-stream="false" Apr 22 17:34:52.419485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415602 2581 flags.go:64] FLAG: --log-text-info-buffer-size="0" Apr 22 17:34:52.419485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415605 2581 flags.go:64] FLAG: --log-text-split-stream="false" Apr 22 17:34:52.419485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415608 2581 flags.go:64] FLAG: --logging-format="text" Apr 22 17:34:52.419485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415611 2581 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Apr 22 17:34:52.419485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415615 2581 flags.go:64] FLAG: --make-iptables-util-chains="true" Apr 22 17:34:52.419485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415617 2581 flags.go:64] FLAG: --manifest-url="" Apr 22 17:34:52.419485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415620 2581 flags.go:64] FLAG: --manifest-url-header="" Apr 22 17:34:52.419485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415625 2581 flags.go:64] FLAG: --max-housekeeping-interval="15s" Apr 22 17:34:52.419485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415628 2581 flags.go:64] FLAG: --max-open-files="1000000" Apr 22 17:34:52.419485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415632 2581 flags.go:64] FLAG: --max-pods="110" Apr 22 17:34:52.419485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415635 2581 flags.go:64] FLAG: --maximum-dead-containers="-1" Apr 22 17:34:52.419485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415638 2581 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Apr 22 17:34:52.419485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415641 2581 flags.go:64] FLAG: --memory-manager-policy="None" Apr 22 17:34:52.419485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415644 2581 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Apr 22 17:34:52.419485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415647 2581 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Apr 22 17:34:52.419485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415650 2581 flags.go:64] FLAG: --node-ip="0.0.0.0" Apr 22 17:34:52.419485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415654 2581 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/worker=,node.openshift.io/os_id=rhel" Apr 22 17:34:52.419485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415661 2581 flags.go:64] FLAG: --node-status-max-images="50" Apr 22 17:34:52.419485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415664 2581 flags.go:64] FLAG: --node-status-update-frequency="10s" Apr 22 17:34:52.419485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415668 2581 flags.go:64] FLAG: --oom-score-adj="-999" Apr 22 17:34:52.419485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415671 2581 flags.go:64] FLAG: --pod-cidr="" Apr 22 17:34:52.419485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415674 2581 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c8cfe89231412ff3ee8cb6207fa0be33cad0f08e88c9c0f1e9f7e8c6f14d6715" Apr 22 17:34:52.420062 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415680 2581 flags.go:64] FLAG: --pod-manifest-path="" Apr 22 17:34:52.420062 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415684 2581 flags.go:64] FLAG: --pod-max-pids="-1" Apr 22 17:34:52.420062 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415688 2581 flags.go:64] FLAG: --pods-per-core="0" Apr 22 17:34:52.420062 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415691 2581 flags.go:64] FLAG: --port="10250" Apr 22 17:34:52.420062 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415694 2581 flags.go:64] FLAG: --protect-kernel-defaults="false" Apr 22 17:34:52.420062 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415697 2581 flags.go:64] FLAG: --provider-id="aws:///us-east-1a/i-0d73bd9cfcb70aed5" Apr 22 17:34:52.420062 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415700 2581 flags.go:64] FLAG: --qos-reserved="" Apr 22 17:34:52.420062 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415703 2581 flags.go:64] FLAG: --read-only-port="10255" Apr 22 17:34:52.420062 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415706 2581 flags.go:64] FLAG: --register-node="true" Apr 22 17:34:52.420062 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415709 2581 flags.go:64] FLAG: --register-schedulable="true" Apr 22 17:34:52.420062 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415712 2581 flags.go:64] FLAG: --register-with-taints="" Apr 22 17:34:52.420062 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415716 2581 flags.go:64] FLAG: --registry-burst="10" Apr 22 17:34:52.420062 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415718 2581 flags.go:64] FLAG: --registry-qps="5" Apr 22 17:34:52.420062 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415721 2581 flags.go:64] FLAG: --reserved-cpus="" Apr 22 17:34:52.420062 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415724 2581 flags.go:64] FLAG: --reserved-memory="" Apr 22 17:34:52.420062 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415728 2581 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Apr 22 17:34:52.420062 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415731 2581 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Apr 22 17:34:52.420062 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415734 2581 flags.go:64] FLAG: --rotate-certificates="false" Apr 22 17:34:52.420062 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415737 2581 flags.go:64] FLAG: --rotate-server-certificates="false" Apr 22 17:34:52.420062 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415739 2581 flags.go:64] FLAG: --runonce="false" Apr 22 17:34:52.420062 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415743 2581 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Apr 22 17:34:52.420062 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415746 2581 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Apr 22 17:34:52.420062 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415749 2581 flags.go:64] FLAG: --seccomp-default="false" Apr 22 17:34:52.420062 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415752 2581 flags.go:64] FLAG: --serialize-image-pulls="true" Apr 22 17:34:52.420062 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415754 2581 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Apr 22 17:34:52.420062 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415758 2581 flags.go:64] FLAG: --storage-driver-db="cadvisor" Apr 22 17:34:52.420730 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415761 2581 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Apr 22 17:34:52.420730 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415764 2581 flags.go:64] FLAG: --storage-driver-password="root" Apr 22 17:34:52.420730 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415767 2581 flags.go:64] FLAG: --storage-driver-secure="false" Apr 22 17:34:52.420730 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415770 2581 flags.go:64] FLAG: --storage-driver-table="stats" Apr 22 17:34:52.420730 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415773 2581 flags.go:64] FLAG: --storage-driver-user="root" Apr 22 17:34:52.420730 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415775 2581 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Apr 22 17:34:52.420730 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415778 2581 flags.go:64] FLAG: --sync-frequency="1m0s" Apr 22 17:34:52.420730 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415782 2581 flags.go:64] FLAG: --system-cgroups="" Apr 22 17:34:52.420730 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415789 2581 flags.go:64] FLAG: --system-reserved="cpu=500m,ephemeral-storage=1Gi,memory=1Gi" Apr 22 17:34:52.420730 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415794 2581 flags.go:64] FLAG: --system-reserved-cgroup="" Apr 22 17:34:52.420730 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415797 2581 flags.go:64] FLAG: --tls-cert-file="" Apr 22 17:34:52.420730 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415800 2581 flags.go:64] FLAG: --tls-cipher-suites="[]" Apr 22 17:34:52.420730 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415804 2581 flags.go:64] FLAG: --tls-min-version="" Apr 22 17:34:52.420730 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415807 2581 flags.go:64] FLAG: --tls-private-key-file="" Apr 22 17:34:52.420730 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415810 2581 flags.go:64] FLAG: --topology-manager-policy="none" Apr 22 17:34:52.420730 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415813 2581 flags.go:64] FLAG: --topology-manager-policy-options="" Apr 22 17:34:52.420730 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415816 2581 flags.go:64] FLAG: --topology-manager-scope="container" Apr 22 17:34:52.420730 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415819 2581 flags.go:64] FLAG: --v="2" Apr 22 17:34:52.420730 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415824 2581 flags.go:64] FLAG: --version="false" Apr 22 17:34:52.420730 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415828 2581 flags.go:64] FLAG: --vmodule="" Apr 22 17:34:52.420730 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415832 2581 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Apr 22 17:34:52.420730 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.415836 2581 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Apr 22 17:34:52.420730 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.415955 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 22 17:34:52.420730 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.415960 2581 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 22 17:34:52.420730 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.415963 2581 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 22 17:34:52.421337 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.415966 2581 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 22 17:34:52.421337 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.415969 2581 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 22 17:34:52.421337 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.415972 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 22 17:34:52.421337 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.415975 2581 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 22 17:34:52.421337 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.415978 2581 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 22 17:34:52.421337 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.415980 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 22 17:34:52.421337 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.415983 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 22 17:34:52.421337 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.415986 2581 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 22 17:34:52.421337 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.415989 2581 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 22 17:34:52.421337 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.415991 2581 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 22 17:34:52.421337 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.415994 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 22 17:34:52.421337 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.415996 2581 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 22 17:34:52.421337 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.415999 2581 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 22 17:34:52.421337 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416001 2581 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 22 17:34:52.421337 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416004 2581 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 22 17:34:52.421337 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416008 2581 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 22 17:34:52.421337 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416011 2581 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 22 17:34:52.421337 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416013 2581 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 22 17:34:52.421337 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416016 2581 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 22 17:34:52.421817 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416018 2581 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 22 17:34:52.421817 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416020 2581 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 22 17:34:52.421817 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416023 2581 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 22 17:34:52.421817 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416025 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 22 17:34:52.421817 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416028 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 22 17:34:52.421817 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416030 2581 feature_gate.go:328] unrecognized feature gate: Example2 Apr 22 17:34:52.421817 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416033 2581 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 22 17:34:52.421817 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416035 2581 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 22 17:34:52.421817 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416038 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 22 17:34:52.421817 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416040 2581 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 22 17:34:52.421817 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416043 2581 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 22 17:34:52.421817 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416046 2581 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 22 17:34:52.421817 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416048 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 22 17:34:52.421817 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416051 2581 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 22 17:34:52.421817 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416053 2581 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 22 17:34:52.421817 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416055 2581 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 22 17:34:52.421817 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416058 2581 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 22 17:34:52.421817 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416061 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 22 17:34:52.421817 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416065 2581 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 22 17:34:52.422297 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416069 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 22 17:34:52.422297 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416072 2581 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 22 17:34:52.422297 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416075 2581 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 22 17:34:52.422297 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416078 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 22 17:34:52.422297 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416081 2581 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 22 17:34:52.422297 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416084 2581 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 22 17:34:52.422297 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416086 2581 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 22 17:34:52.422297 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416089 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 22 17:34:52.422297 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416092 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 22 17:34:52.422297 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416096 2581 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 22 17:34:52.422297 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416099 2581 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 22 17:34:52.422297 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416102 2581 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 22 17:34:52.422297 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416104 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 22 17:34:52.422297 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416107 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 22 17:34:52.422297 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416109 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 22 17:34:52.422297 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416112 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 22 17:34:52.422297 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416115 2581 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 22 17:34:52.422297 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416119 2581 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 22 17:34:52.422297 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416122 2581 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 22 17:34:52.422297 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416124 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 22 17:34:52.422786 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416127 2581 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 22 17:34:52.422786 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416129 2581 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 22 17:34:52.422786 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416132 2581 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 22 17:34:52.422786 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416135 2581 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 22 17:34:52.422786 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416137 2581 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 22 17:34:52.422786 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416140 2581 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 22 17:34:52.422786 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416142 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 22 17:34:52.422786 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416145 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 22 17:34:52.422786 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416148 2581 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 22 17:34:52.422786 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416152 2581 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 22 17:34:52.422786 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416155 2581 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 22 17:34:52.422786 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416157 2581 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 22 17:34:52.422786 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416161 2581 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 22 17:34:52.422786 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416164 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 22 17:34:52.422786 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416166 2581 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 22 17:34:52.422786 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416169 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 22 17:34:52.422786 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416172 2581 feature_gate.go:328] unrecognized feature gate: Example Apr 22 17:34:52.422786 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416174 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 22 17:34:52.422786 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416177 2581 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 22 17:34:52.422786 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416179 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 22 17:34:52.423290 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416182 2581 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 22 17:34:52.423290 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416186 2581 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 22 17:34:52.423290 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416188 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 22 17:34:52.423290 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416191 2581 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 22 17:34:52.423290 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.416194 2581 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 22 17:34:52.423290 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.416753 2581 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 22 17:34:52.423290 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.423280 2581 server.go:530] "Kubelet version" kubeletVersion="v1.33.9" Apr 22 17:34:52.423475 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.423299 2581 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 22 17:34:52.423475 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423365 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 22 17:34:52.423475 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423371 2581 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 22 17:34:52.423475 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423375 2581 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 22 17:34:52.423475 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423378 2581 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 22 17:34:52.423475 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423381 2581 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 22 17:34:52.423475 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423384 2581 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 22 17:34:52.423475 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423387 2581 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 22 17:34:52.423475 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423390 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 22 17:34:52.423475 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423392 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 22 17:34:52.423475 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423395 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 22 17:34:52.423475 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423398 2581 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 22 17:34:52.423475 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423400 2581 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 22 17:34:52.423475 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423403 2581 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 22 17:34:52.423475 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423406 2581 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 22 17:34:52.423475 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423409 2581 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 22 17:34:52.423475 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423412 2581 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 22 17:34:52.423475 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423414 2581 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 22 17:34:52.423475 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423417 2581 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 22 17:34:52.423475 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423419 2581 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 22 17:34:52.424012 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423422 2581 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 22 17:34:52.424012 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423424 2581 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 22 17:34:52.424012 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423427 2581 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 22 17:34:52.424012 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423429 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 22 17:34:52.424012 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423432 2581 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 22 17:34:52.424012 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423434 2581 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 22 17:34:52.424012 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423437 2581 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 22 17:34:52.424012 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423439 2581 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 22 17:34:52.424012 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423442 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 22 17:34:52.424012 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423445 2581 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 22 17:34:52.424012 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423448 2581 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 22 17:34:52.424012 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423450 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 22 17:34:52.424012 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423454 2581 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 22 17:34:52.424012 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423458 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 22 17:34:52.424012 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423460 2581 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 22 17:34:52.424012 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423463 2581 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 22 17:34:52.424012 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423466 2581 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 22 17:34:52.424012 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423468 2581 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 22 17:34:52.424012 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423471 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 22 17:34:52.424488 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423473 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 22 17:34:52.424488 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423475 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 22 17:34:52.424488 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423478 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 22 17:34:52.424488 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423481 2581 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 22 17:34:52.424488 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423484 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 22 17:34:52.424488 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423486 2581 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 22 17:34:52.424488 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423489 2581 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 22 17:34:52.424488 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423491 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 22 17:34:52.424488 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423494 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 22 17:34:52.424488 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423497 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 22 17:34:52.424488 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423499 2581 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 22 17:34:52.424488 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423502 2581 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 22 17:34:52.424488 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423504 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 22 17:34:52.424488 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423507 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 22 17:34:52.424488 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423509 2581 feature_gate.go:328] unrecognized feature gate: Example Apr 22 17:34:52.424488 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423511 2581 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 22 17:34:52.424488 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423514 2581 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 22 17:34:52.424488 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423516 2581 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 22 17:34:52.424488 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423519 2581 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 22 17:34:52.424958 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423523 2581 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 22 17:34:52.424958 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423527 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 22 17:34:52.424958 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423530 2581 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 22 17:34:52.424958 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423533 2581 feature_gate.go:328] unrecognized feature gate: Example2 Apr 22 17:34:52.424958 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423536 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 22 17:34:52.424958 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423538 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 22 17:34:52.424958 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423541 2581 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 22 17:34:52.424958 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423545 2581 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 22 17:34:52.424958 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423548 2581 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 22 17:34:52.424958 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423550 2581 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 22 17:34:52.424958 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423552 2581 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 22 17:34:52.424958 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423555 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 22 17:34:52.424958 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423557 2581 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 22 17:34:52.424958 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423561 2581 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 22 17:34:52.424958 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423565 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 22 17:34:52.424958 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423568 2581 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 22 17:34:52.424958 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423571 2581 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 22 17:34:52.424958 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423573 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 22 17:34:52.424958 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423576 2581 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 22 17:34:52.425415 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423578 2581 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 22 17:34:52.425415 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423581 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 22 17:34:52.425415 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423583 2581 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 22 17:34:52.425415 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423586 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 22 17:34:52.425415 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423588 2581 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 22 17:34:52.425415 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423591 2581 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 22 17:34:52.425415 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423593 2581 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 22 17:34:52.425415 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423596 2581 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 22 17:34:52.425415 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423598 2581 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 22 17:34:52.425415 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423601 2581 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 22 17:34:52.425415 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.423606 2581 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 22 17:34:52.425415 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423710 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 22 17:34:52.425415 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423715 2581 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 22 17:34:52.425415 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423719 2581 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 22 17:34:52.425415 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423721 2581 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 22 17:34:52.425783 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423724 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 22 17:34:52.425783 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423727 2581 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 22 17:34:52.425783 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423730 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 22 17:34:52.425783 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423732 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 22 17:34:52.425783 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423734 2581 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 22 17:34:52.425783 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423737 2581 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 22 17:34:52.425783 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423740 2581 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 22 17:34:52.425783 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423743 2581 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 22 17:34:52.425783 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423746 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 22 17:34:52.425783 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423749 2581 feature_gate.go:328] unrecognized feature gate: Example2 Apr 22 17:34:52.425783 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423751 2581 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 22 17:34:52.425783 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423754 2581 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 22 17:34:52.425783 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423757 2581 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 22 17:34:52.425783 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423759 2581 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 22 17:34:52.425783 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423762 2581 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 22 17:34:52.425783 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423764 2581 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 22 17:34:52.425783 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423767 2581 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 22 17:34:52.425783 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423769 2581 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 22 17:34:52.425783 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423772 2581 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 22 17:34:52.425783 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423774 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 22 17:34:52.426277 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423777 2581 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 22 17:34:52.426277 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423779 2581 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 22 17:34:52.426277 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423782 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 22 17:34:52.426277 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423784 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 22 17:34:52.426277 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423787 2581 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 22 17:34:52.426277 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423789 2581 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 22 17:34:52.426277 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423792 2581 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 22 17:34:52.426277 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423794 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 22 17:34:52.426277 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423796 2581 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 22 17:34:52.426277 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423799 2581 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 22 17:34:52.426277 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423801 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 22 17:34:52.426277 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423804 2581 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 22 17:34:52.426277 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423807 2581 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 22 17:34:52.426277 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423809 2581 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 22 17:34:52.426277 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423812 2581 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 22 17:34:52.426277 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423814 2581 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 22 17:34:52.426277 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423816 2581 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 22 17:34:52.426277 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423819 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 22 17:34:52.426277 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423822 2581 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 22 17:34:52.426277 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423825 2581 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 22 17:34:52.426753 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423827 2581 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 22 17:34:52.426753 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423830 2581 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 22 17:34:52.426753 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423832 2581 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 22 17:34:52.426753 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423835 2581 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 22 17:34:52.426753 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423838 2581 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 22 17:34:52.426753 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423841 2581 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 22 17:34:52.426753 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423844 2581 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 22 17:34:52.426753 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423846 2581 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 22 17:34:52.426753 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423848 2581 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 22 17:34:52.426753 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423851 2581 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 22 17:34:52.426753 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423853 2581 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 22 17:34:52.426753 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423855 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 22 17:34:52.426753 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423858 2581 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 22 17:34:52.426753 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423860 2581 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 22 17:34:52.426753 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423863 2581 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 22 17:34:52.426753 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423865 2581 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 22 17:34:52.426753 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423867 2581 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 22 17:34:52.426753 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423870 2581 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 22 17:34:52.426753 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423872 2581 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 22 17:34:52.427260 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423889 2581 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 22 17:34:52.427260 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423892 2581 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 22 17:34:52.427260 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423894 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 22 17:34:52.427260 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423897 2581 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 22 17:34:52.427260 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423899 2581 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 22 17:34:52.427260 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423902 2581 feature_gate.go:328] unrecognized feature gate: Example Apr 22 17:34:52.427260 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423905 2581 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 22 17:34:52.427260 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423907 2581 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 22 17:34:52.427260 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423910 2581 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 22 17:34:52.427260 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423913 2581 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 22 17:34:52.427260 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423917 2581 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 22 17:34:52.427260 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423920 2581 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 22 17:34:52.427260 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423923 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 22 17:34:52.427260 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423926 2581 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 22 17:34:52.427260 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423929 2581 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 22 17:34:52.427260 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423931 2581 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 22 17:34:52.427260 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423933 2581 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 22 17:34:52.427260 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423936 2581 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 22 17:34:52.427260 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423939 2581 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 22 17:34:52.427260 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423941 2581 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 22 17:34:52.427747 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423944 2581 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 22 17:34:52.427747 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423946 2581 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 22 17:34:52.427747 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:52.423948 2581 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 22 17:34:52.427747 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.423954 2581 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 22 17:34:52.427747 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.424635 2581 server.go:962] "Client rotation is on, will bootstrap in background" Apr 22 17:34:52.427747 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.426597 2581 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Apr 22 17:34:52.427747 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.427681 2581 server.go:1019] "Starting client certificate rotation" Apr 22 17:34:52.427945 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.427775 2581 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kube-apiserver-client-kubelet" Apr 22 17:34:52.428597 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.428584 2581 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kube-apiserver-client-kubelet" Apr 22 17:34:52.452170 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.452142 2581 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Apr 22 17:34:52.458809 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.458788 2581 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Apr 22 17:34:52.475536 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.475508 2581 log.go:25] "Validated CRI v1 runtime API" Apr 22 17:34:52.481227 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.481206 2581 log.go:25] "Validated CRI v1 image API" Apr 22 17:34:52.482469 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.482442 2581 server.go:1452] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Apr 22 17:34:52.488508 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.488482 2581 fs.go:135] Filesystem UUIDs: map[7B77-95E7:/dev/nvme0n1p2 8dc26d71-616e-4f9b-9e0e-1c9429623e26:/dev/nvme0n1p3 e55a2278-c080-4325-85db-2ee0ca2812f5:/dev/nvme0n1p4] Apr 22 17:34:52.488592 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.488505 2581 fs.go:136] Filesystem partitions: map[/dev/nvme0n1p3:{mountpoint:/boot major:259 minor:3 fsType:ext4 blockSize:0} /dev/nvme0n1p4:{mountpoint:/var major:259 minor:4 fsType:xfs blockSize:0} /dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:31 fsType:tmpfs blockSize:0} composefs_0-33:{mountpoint:/ major:0 minor:33 fsType:overlay blockSize:0}] Apr 22 17:34:52.494598 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.494473 2581 manager.go:217] Machine: {Timestamp:2026-04-22 17:34:52.49235404 +0000 UTC m=+0.385049755 CPUVendorID:GenuineIntel NumCores:8 NumPhysicalCores:4 NumSockets:1 CpuFrequency:3096975 MemoryCapacity:32812175360 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:ec2735bbc9bcbbaa82dd881a10b99f98 SystemUUID:ec2735bb-c9bc-bbaa-82dd-881a10b99f98 BootID:8bf8db5a-a5a0-4767-b101-688c093bacc5 Filesystems:[{Device:composefs_0-33 DeviceMajor:0 DeviceMinor:33 Capacity:6103040 Type:vfs Inodes:18446744073709551615 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:31 Capacity:16406089728 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/nvme0n1p3 DeviceMajor:259 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16406085632 Type:vfs Inodes:4005392 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6562435072 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/nvme0n1p4 DeviceMajor:259 DeviceMinor:4 Capacity:128243970048 Type:vfs Inodes:62651840 HasInodes:true}] DiskMap:map[259:0:{Name:nvme0n1 Major:259 Minor:0 Size:128849018880 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:02:eb:f3:d7:43:0d Speed:0 Mtu:9001} {Name:ens5 MacAddress:02:eb:f3:d7:43:0d Speed:0 Mtu:9001} {Name:ovs-system MacAddress:5a:43:45:16:8d:c4 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:32812175360 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0 4] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:1 Threads:[1 5] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:2 Threads:[2 6] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:3 Threads:[3 7] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:}] Caches:[{Id:0 Size:34603008 Type:Unified Level:3}] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Apr 22 17:34:52.494598 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.494585 2581 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Apr 22 17:34:52.494721 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.494677 2581 manager.go:233] Version: {KernelVersion:5.14.0-570.107.1.el9_6.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 9.6.20260414-0 (Plow) DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Apr 22 17:34:52.495921 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.495867 2581 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 22 17:34:52.496100 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.495924 2581 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-10-0-138-57.ec2.internal","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"500m","ephemeral-storage":"1Gi","memory":"1Gi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 22 17:34:52.496178 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.496112 2581 topology_manager.go:138] "Creating topology manager with none policy" Apr 22 17:34:52.496178 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.496124 2581 container_manager_linux.go:306] "Creating device plugin manager" Apr 22 17:34:52.496178 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.496143 2581 manager.go:141] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Apr 22 17:34:52.497094 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.497075 2581 reflector.go:430] "Caches populated" logger="kubernetes.io/kube-apiserver-client-kubelet" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Apr 22 17:34:52.497608 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.497593 2581 server.go:72] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Apr 22 17:34:52.498774 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.498761 2581 state_mem.go:36] "Initialized new in-memory state store" Apr 22 17:34:52.498931 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.498920 2581 server.go:1267] "Using root directory" path="/var/lib/kubelet" Apr 22 17:34:52.501078 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.501064 2581 kubelet.go:491] "Attempting to sync node with API server" Apr 22 17:34:52.501139 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.501084 2581 kubelet.go:386] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 22 17:34:52.501139 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.501104 2581 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Apr 22 17:34:52.501139 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.501121 2581 kubelet.go:397] "Adding apiserver pod source" Apr 22 17:34:52.501139 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.501135 2581 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 22 17:34:52.502292 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.502278 2581 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Apr 22 17:34:52.502368 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.502301 2581 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Apr 22 17:34:52.505421 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.505403 2581 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="cri-o" version="1.33.10-2.rhaos4.20.gita4d0894.el9" apiVersion="v1" Apr 22 17:34:52.506784 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.506769 2581 kubelet.go:953] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Apr 22 17:34:52.508406 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.508394 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Apr 22 17:34:52.508457 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.508411 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Apr 22 17:34:52.508457 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.508417 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Apr 22 17:34:52.508457 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.508424 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Apr 22 17:34:52.508457 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.508429 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Apr 22 17:34:52.508457 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.508435 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/secret" Apr 22 17:34:52.508457 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.508441 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Apr 22 17:34:52.508457 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.508446 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Apr 22 17:34:52.508457 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.508453 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/fc" Apr 22 17:34:52.508457 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.508462 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Apr 22 17:34:52.508686 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.508476 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/projected" Apr 22 17:34:52.508686 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.508485 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Apr 22 17:34:52.509382 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.509372 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/csi" Apr 22 17:34:52.509421 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.509383 2581 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/image" Apr 22 17:34:52.512931 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.512916 2581 watchdog_linux.go:99] "Systemd watchdog is not enabled" Apr 22 17:34:52.512997 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.512957 2581 server.go:1295] "Started kubelet" Apr 22 17:34:52.513069 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.513043 2581 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Apr 22 17:34:52.513108 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.513047 2581 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 22 17:34:52.513108 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.513101 2581 server_v1.go:47] "podresources" method="list" useActivePods=true Apr 22 17:34:52.513800 ip-10-0-138-57 systemd[1]: Started Kubernetes Kubelet. Apr 22 17:34:52.514841 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.514821 2581 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 22 17:34:52.515260 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:52.515235 2581 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: nodes \"ip-10-0-138-57.ec2.internal\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 22 17:34:52.515439 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.515416 2581 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "ip-10-0-138-57.ec2.internal" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Apr 22 17:34:52.515750 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:52.515723 2581 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 22 17:34:52.517069 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.517048 2581 server.go:317] "Adding debug handlers to kubelet server" Apr 22 17:34:52.521284 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.521263 2581 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 22 17:34:52.521370 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.521278 2581 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kubelet-serving" Apr 22 17:34:52.522485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.522008 2581 volume_manager.go:295] "The desired_state_of_world populator starts" Apr 22 17:34:52.522485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.522040 2581 volume_manager.go:297] "Starting Kubelet Volume Manager" Apr 22 17:34:52.522485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.522222 2581 reconstruct.go:97] "Volume reconstruction finished" Apr 22 17:34:52.522485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.522231 2581 reconciler.go:26] "Reconciler: start to sync state" Apr 22 17:34:52.522485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.522234 2581 factory.go:55] Registering systemd factory Apr 22 17:34:52.522485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.522254 2581 factory.go:223] Registration of the systemd container factory successfully Apr 22 17:34:52.522485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.522311 2581 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Apr 22 17:34:52.522485 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:52.522329 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-138-57.ec2.internal\" not found" Apr 22 17:34:52.522485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.522465 2581 factory.go:153] Registering CRI-O factory Apr 22 17:34:52.522485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.522478 2581 factory.go:223] Registration of the crio container factory successfully Apr 22 17:34:52.523026 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.522536 2581 factory.go:221] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Apr 22 17:34:52.523026 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.522562 2581 factory.go:103] Registering Raw factory Apr 22 17:34:52.523026 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.522579 2581 manager.go:1196] Started watching for new ooms in manager Apr 22 17:34:52.523271 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.523073 2581 manager.go:319] Starting recovery of all containers Apr 22 17:34:52.524132 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:52.523194 2581 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{ip-10-0-138-57.ec2.internal.18a8be4d5efb054b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ip-10-0-138-57.ec2.internal,UID:ip-10-0-138-57.ec2.internal,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ip-10-0-138-57.ec2.internal,},FirstTimestamp:2026-04-22 17:34:52.512929099 +0000 UTC m=+0.405624814,LastTimestamp:2026-04-22 17:34:52.512929099 +0000 UTC m=+0.405624814,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ip-10-0-138-57.ec2.internal,}" Apr 22 17:34:52.526910 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:52.526863 2581 kubelet.go:1618] "Image garbage collection failed once. Stats initialization may not have completed yet" err="failed to get imageFs info: unable to find data in memory cache" Apr 22 17:34:52.530665 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.530489 2581 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-6tvpk" Apr 22 17:34:52.532329 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.532310 2581 manager.go:324] Recovery completed Apr 22 17:34:52.533737 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:52.533721 2581 watcher.go:152] Failed to watch directory "/sys/fs/cgroup/system.slice/systemd-update-utmp-runlevel.service": inotify_add_watch /sys/fs/cgroup/system.slice/systemd-update-utmp-runlevel.service: no such file or directory Apr 22 17:34:52.537032 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.537018 2581 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 22 17:34:52.538776 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:52.538754 2581 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"ip-10-0-138-57.ec2.internal\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="200ms" Apr 22 17:34:52.539851 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.539837 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-138-57.ec2.internal" event="NodeHasSufficientMemory" Apr 22 17:34:52.539941 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.539864 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-138-57.ec2.internal" event="NodeHasNoDiskPressure" Apr 22 17:34:52.539941 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.539897 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-138-57.ec2.internal" event="NodeHasSufficientPID" Apr 22 17:34:52.540490 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.540473 2581 cpu_manager.go:222] "Starting CPU manager" policy="none" Apr 22 17:34:52.540490 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.540490 2581 cpu_manager.go:223] "Reconciling" reconcilePeriod="10s" Apr 22 17:34:52.540589 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.540508 2581 state_mem.go:36] "Initialized new in-memory state store" Apr 22 17:34:52.540804 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.540790 2581 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-6tvpk" Apr 22 17:34:52.543014 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.543000 2581 policy_none.go:49] "None policy: Start" Apr 22 17:34:52.543083 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.543017 2581 memory_manager.go:186] "Starting memorymanager" policy="None" Apr 22 17:34:52.543083 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.543027 2581 state_mem.go:35] "Initializing new in-memory state store" Apr 22 17:34:52.548505 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.548488 2581 reflector.go:430] "Caches populated" type="*v1.CSIDriver" reflector="k8s.io/client-go/informers/factory.go:160" Apr 22 17:34:52.582657 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.582631 2581 manager.go:341] "Starting Device Plugin manager" Apr 22 17:34:52.592117 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:52.582675 2581 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Apr 22 17:34:52.592117 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.582689 2581 server.go:85] "Starting device plugin registration server" Apr 22 17:34:52.592117 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.583049 2581 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 22 17:34:52.592117 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.583062 2581 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 22 17:34:52.592117 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.583187 2581 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Apr 22 17:34:52.592117 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.583272 2581 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Apr 22 17:34:52.592117 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.583280 2581 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 22 17:34:52.592117 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:52.583838 2581 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="non-existent label \"crio-containers\"" Apr 22 17:34:52.592117 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:52.583944 2581 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ip-10-0-138-57.ec2.internal\" not found" Apr 22 17:34:52.632930 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.632863 2581 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Apr 22 17:34:52.634236 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.634213 2581 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Apr 22 17:34:52.634348 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.634255 2581 status_manager.go:230] "Starting to sync pod status with apiserver" Apr 22 17:34:52.634348 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.634281 2581 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 22 17:34:52.634348 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.634291 2581 kubelet.go:2451] "Starting kubelet main sync loop" Apr 22 17:34:52.634348 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:52.634334 2581 kubelet.go:2475] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Apr 22 17:34:52.636697 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.636671 2581 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Apr 22 17:34:52.683646 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.683558 2581 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 22 17:34:52.687430 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.687394 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-138-57.ec2.internal" event="NodeHasSufficientMemory" Apr 22 17:34:52.687430 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.687432 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-138-57.ec2.internal" event="NodeHasNoDiskPressure" Apr 22 17:34:52.687622 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.687444 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-138-57.ec2.internal" event="NodeHasSufficientPID" Apr 22 17:34:52.687622 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.687470 2581 kubelet_node_status.go:78] "Attempting to register node" node="ip-10-0-138-57.ec2.internal" Apr 22 17:34:52.693367 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.693346 2581 kubelet_node_status.go:81] "Successfully registered node" node="ip-10-0-138-57.ec2.internal" Apr 22 17:34:52.693477 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:52.693371 2581 kubelet_node_status.go:597] "Error updating node status, will retry" err="error getting node \"ip-10-0-138-57.ec2.internal\": node \"ip-10-0-138-57.ec2.internal\" not found" Apr 22 17:34:52.709169 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:52.709141 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-138-57.ec2.internal\" not found" Apr 22 17:34:52.735178 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.735147 2581 kubelet.go:2537] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-138-57.ec2.internal","kube-system/kube-apiserver-proxy-ip-10-0-138-57.ec2.internal"] Apr 22 17:34:52.735320 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.735241 2581 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 22 17:34:52.736234 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.736218 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-138-57.ec2.internal" event="NodeHasSufficientMemory" Apr 22 17:34:52.736306 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.736250 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-138-57.ec2.internal" event="NodeHasNoDiskPressure" Apr 22 17:34:52.736306 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.736264 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-138-57.ec2.internal" event="NodeHasSufficientPID" Apr 22 17:34:52.737385 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.737373 2581 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 22 17:34:52.737526 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.737511 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-138-57.ec2.internal" Apr 22 17:34:52.737596 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.737548 2581 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 22 17:34:52.738106 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.738090 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-138-57.ec2.internal" event="NodeHasSufficientMemory" Apr 22 17:34:52.738188 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.738119 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-138-57.ec2.internal" event="NodeHasNoDiskPressure" Apr 22 17:34:52.738188 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.738135 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-138-57.ec2.internal" event="NodeHasSufficientPID" Apr 22 17:34:52.738188 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.738093 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-138-57.ec2.internal" event="NodeHasSufficientMemory" Apr 22 17:34:52.738188 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.738177 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-138-57.ec2.internal" event="NodeHasNoDiskPressure" Apr 22 17:34:52.738188 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.738189 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-138-57.ec2.internal" event="NodeHasSufficientPID" Apr 22 17:34:52.739325 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.739310 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/kube-apiserver-proxy-ip-10-0-138-57.ec2.internal" Apr 22 17:34:52.739406 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.739336 2581 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 22 17:34:52.740307 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.740289 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-138-57.ec2.internal" event="NodeHasSufficientMemory" Apr 22 17:34:52.740384 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.740324 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-138-57.ec2.internal" event="NodeHasNoDiskPressure" Apr 22 17:34:52.740384 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.740339 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-138-57.ec2.internal" event="NodeHasSufficientPID" Apr 22 17:34:52.763802 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:52.763768 2581 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-10-0-138-57.ec2.internal\" not found" node="ip-10-0-138-57.ec2.internal" Apr 22 17:34:52.768321 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:52.768301 2581 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-10-0-138-57.ec2.internal\" not found" node="ip-10-0-138-57.ec2.internal" Apr 22 17:34:52.809668 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:52.809635 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-138-57.ec2.internal\" not found" Apr 22 17:34:52.824660 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.824630 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b835e279c462c30381fc4ead00ea7937-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-138-57.ec2.internal\" (UID: \"b835e279c462c30381fc4ead00ea7937\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-138-57.ec2.internal" Apr 22 17:34:52.824764 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.824663 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/8df1f5cf5626886b4ad3778d0dba9ac9-config\") pod \"kube-apiserver-proxy-ip-10-0-138-57.ec2.internal\" (UID: \"8df1f5cf5626886b4ad3778d0dba9ac9\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-138-57.ec2.internal" Apr 22 17:34:52.824764 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.824681 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/b835e279c462c30381fc4ead00ea7937-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-138-57.ec2.internal\" (UID: \"b835e279c462c30381fc4ead00ea7937\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-138-57.ec2.internal" Apr 22 17:34:52.910254 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:52.910222 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-138-57.ec2.internal\" not found" Apr 22 17:34:52.925820 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.925789 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/b835e279c462c30381fc4ead00ea7937-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-138-57.ec2.internal\" (UID: \"b835e279c462c30381fc4ead00ea7937\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-138-57.ec2.internal" Apr 22 17:34:52.925951 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.925825 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b835e279c462c30381fc4ead00ea7937-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-138-57.ec2.internal\" (UID: \"b835e279c462c30381fc4ead00ea7937\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-138-57.ec2.internal" Apr 22 17:34:52.925951 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.925843 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/8df1f5cf5626886b4ad3778d0dba9ac9-config\") pod \"kube-apiserver-proxy-ip-10-0-138-57.ec2.internal\" (UID: \"8df1f5cf5626886b4ad3778d0dba9ac9\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-138-57.ec2.internal" Apr 22 17:34:52.925951 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.925897 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/host-path/8df1f5cf5626886b4ad3778d0dba9ac9-config\") pod \"kube-apiserver-proxy-ip-10-0-138-57.ec2.internal\" (UID: \"8df1f5cf5626886b4ad3778d0dba9ac9\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-138-57.ec2.internal" Apr 22 17:34:52.925951 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.925925 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b835e279c462c30381fc4ead00ea7937-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-138-57.ec2.internal\" (UID: \"b835e279c462c30381fc4ead00ea7937\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-138-57.ec2.internal" Apr 22 17:34:52.925951 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:52.925923 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/b835e279c462c30381fc4ead00ea7937-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-138-57.ec2.internal\" (UID: \"b835e279c462c30381fc4ead00ea7937\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-138-57.ec2.internal" Apr 22 17:34:53.010522 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:53.010441 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-138-57.ec2.internal\" not found" Apr 22 17:34:53.066964 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:53.066932 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-138-57.ec2.internal" Apr 22 17:34:53.071429 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:53.071409 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/kube-apiserver-proxy-ip-10-0-138-57.ec2.internal" Apr 22 17:34:53.110848 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:53.110819 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-138-57.ec2.internal\" not found" Apr 22 17:34:53.211436 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:53.211391 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-138-57.ec2.internal\" not found" Apr 22 17:34:53.312018 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:53.311923 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-138-57.ec2.internal\" not found" Apr 22 17:34:53.412642 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:53.412602 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-138-57.ec2.internal\" not found" Apr 22 17:34:53.427970 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:53.427946 2581 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Apr 22 17:34:53.428138 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:53.428109 2581 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" err="very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received" Apr 22 17:34:53.428200 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:53.428170 2581 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" err="very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received" Apr 22 17:34:53.513607 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:53.513569 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-138-57.ec2.internal\" not found" Apr 22 17:34:53.521959 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:53.521929 2581 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kubelet-serving" Apr 22 17:34:53.533786 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:53.533759 2581 reflector.go:430] "Caches populated" logger="kubernetes.io/kubelet-serving" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Apr 22 17:34:53.543915 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:53.543868 2581 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kube-apiserver-client-kubelet" expiration="2028-04-21 17:29:52 +0000 UTC" deadline="2028-02-06 05:52:41.278976916 +0000 UTC" Apr 22 17:34:53.543915 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:53.543915 2581 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kube-apiserver-client-kubelet" sleep="15708h17m47.735065089s" Apr 22 17:34:53.567897 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:53.567811 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb835e279c462c30381fc4ead00ea7937.slice/crio-223dc670a86a9e1a5e836996ac68c7830ee280851d05812c7c400fb99cb79ca3 WatchSource:0}: Error finding container 223dc670a86a9e1a5e836996ac68c7830ee280851d05812c7c400fb99cb79ca3: Status 404 returned error can't find the container with id 223dc670a86a9e1a5e836996ac68c7830ee280851d05812c7c400fb99cb79ca3 Apr 22 17:34:53.568130 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:53.568102 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8df1f5cf5626886b4ad3778d0dba9ac9.slice/crio-d898bcfd053ca9648e57f6297c4fac11bc3fca47dc8547583990d0c3535bab6c WatchSource:0}: Error finding container d898bcfd053ca9648e57f6297c4fac11bc3fca47dc8547583990d0c3535bab6c: Status 404 returned error can't find the container with id d898bcfd053ca9648e57f6297c4fac11bc3fca47dc8547583990d0c3535bab6c Apr 22 17:34:53.572230 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:53.572213 2581 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 22 17:34:53.573673 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:53.573652 2581 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kubelet-serving" csr="csr-xjrbq" Apr 22 17:34:53.584748 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:53.584726 2581 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kubelet-serving" csr="csr-xjrbq" Apr 22 17:34:53.613758 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:53.613726 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-138-57.ec2.internal\" not found" Apr 22 17:34:53.638397 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:53.638348 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-138-57.ec2.internal" event={"ID":"b835e279c462c30381fc4ead00ea7937","Type":"ContainerStarted","Data":"223dc670a86a9e1a5e836996ac68c7830ee280851d05812c7c400fb99cb79ca3"} Apr 22 17:34:53.639292 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:53.639262 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-138-57.ec2.internal" event={"ID":"8df1f5cf5626886b4ad3778d0dba9ac9","Type":"ContainerStarted","Data":"d898bcfd053ca9648e57f6297c4fac11bc3fca47dc8547583990d0c3535bab6c"} Apr 22 17:34:53.707690 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:53.707661 2581 reflector.go:430] "Caches populated" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" Apr 22 17:34:53.713844 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:53.713822 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-138-57.ec2.internal\" not found" Apr 22 17:34:53.814387 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:53.814360 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-138-57.ec2.internal\" not found" Apr 22 17:34:53.914972 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:53.914856 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-138-57.ec2.internal\" not found" Apr 22 17:34:54.015783 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:54.015743 2581 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-138-57.ec2.internal\" not found" Apr 22 17:34:54.098962 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.098929 2581 reflector.go:430] "Caches populated" type="*v1.Node" reflector="k8s.io/client-go/informers/factory.go:160" Apr 22 17:34:54.121980 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.121930 2581 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-138-57.ec2.internal" Apr 22 17:34:54.131102 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.130980 2581 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 22 17:34:54.132243 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.132006 2581 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-138-57.ec2.internal" Apr 22 17:34:54.142150 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.142121 2581 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 22 17:34:54.432969 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.432931 2581 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Apr 22 17:34:54.503130 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.503091 2581 apiserver.go:52] "Watching apiserver" Apr 22 17:34:54.508481 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.508452 2581 reflector.go:430] "Caches populated" type="*v1.Pod" reflector="pkg/kubelet/config/apiserver.go:66" Apr 22 17:34:54.511332 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.511298 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-p72t5","openshift-network-operator/iptables-alerter-kctrb","openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79","openshift-cluster-node-tuning-operator/tuned-4tdc7","openshift-dns/node-resolver-4m4hg","openshift-multus/network-metrics-daemon-zhl6m","openshift-ovn-kubernetes/ovnkube-node-tn68c","kube-system/konnectivity-agent-vtsck","kube-system/kube-apiserver-proxy-ip-10-0-138-57.ec2.internal","openshift-image-registry/node-ca-6zmdm","openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-138-57.ec2.internal","openshift-multus/multus-additional-cni-plugins-pnh8b","openshift-multus/multus-nwhzz"] Apr 22 17:34:54.513799 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.513773 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.515023 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.515000 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-kctrb" Apr 22 17:34:54.516064 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.516043 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" Apr 22 17:34:54.516786 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.516757 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-node-metrics-cert\"" Apr 22 17:34:54.516957 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.516818 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-config\"" Apr 22 17:34:54.516957 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.516861 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-node-dockercfg-8n2vv\"" Apr 22 17:34:54.517076 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.516975 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"openshift-service-ca.crt\"" Apr 22 17:34:54.517230 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.517212 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"kube-root-ca.crt\"" Apr 22 17:34:54.517304 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.517234 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-script-lib\"" Apr 22 17:34:54.517304 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.517212 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"openshift-service-ca.crt\"" Apr 22 17:34:54.517304 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.517235 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"env-overrides\"" Apr 22 17:34:54.517431 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.517339 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-script\"" Apr 22 17:34:54.517463 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.517430 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-dockercfg-ntf6c\"" Apr 22 17:34:54.517525 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.517510 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"kube-root-ca.crt\"" Apr 22 17:34:54.517650 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.517631 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.518055 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.518036 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-csi-drivers\"/\"aws-ebs-csi-driver-node-metrics-serving-cert\"" Apr 22 17:34:54.518171 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.518135 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-csi-drivers\"/\"kube-root-ca.crt\"" Apr 22 17:34:54.518354 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.518337 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-csi-drivers\"/\"aws-ebs-csi-driver-node-sa-dockercfg-9vmk9\"" Apr 22 17:34:54.518830 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.518812 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-4m4hg" Apr 22 17:34:54.519194 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.519178 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-csi-drivers\"/\"openshift-service-ca.crt\"" Apr 22 17:34:54.520562 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.520546 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:34:54.520651 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:54.520628 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zhl6m" podUID="cff34b01-53b3-4374-b73d-d3146e3c558d" Apr 22 17:34:54.520856 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.520838 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"kube-root-ca.crt\"" Apr 22 17:34:54.520969 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.520935 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"tuned-dockercfg-fnvgp\"" Apr 22 17:34:54.520969 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.520962 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"openshift-service-ca.crt\"" Apr 22 17:34:54.521268 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.521251 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"openshift-service-ca.crt\"" Apr 22 17:34:54.521413 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.521396 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"node-resolver-dockercfg-ptgzc\"" Apr 22 17:34:54.521542 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.521525 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"kube-root-ca.crt\"" Apr 22 17:34:54.521821 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.521803 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:34:54.521904 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:54.521871 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-p72t5" podUID="c858366e-d830-4c42-b18e-1eab94ea3113" Apr 22 17:34:54.524347 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.523963 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/konnectivity-agent-vtsck" Apr 22 17:34:54.524347 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.524094 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-6zmdm" Apr 22 17:34:54.525799 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.525740 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kube-system\"/\"konnectivity-ca-bundle\"" Apr 22 17:34:54.525960 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.525936 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"default-dockercfg-t4d8l\"" Apr 22 17:34:54.526158 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.526138 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-pnh8b" Apr 22 17:34:54.526460 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.526441 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"kube-root-ca.crt\"" Apr 22 17:34:54.526697 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.526679 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"openshift-service-ca.crt\"" Apr 22 17:34:54.526780 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.526679 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"konnectivity-agent\"" Apr 22 17:34:54.526780 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.526751 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"node-ca-dockercfg-xphvg\"" Apr 22 17:34:54.526915 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.526683 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"image-registry-certificates\"" Apr 22 17:34:54.527470 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.527451 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.528794 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.528776 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"kube-root-ca.crt\"" Apr 22 17:34:54.529100 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.529082 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"default-cni-sysctl-allowlist\"" Apr 22 17:34:54.529439 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.529410 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"openshift-service-ca.crt\"" Apr 22 17:34:54.529517 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.529442 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"multus-daemon-config\"" Apr 22 17:34:54.529705 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.529687 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"default-dockercfg-58pbr\"" Apr 22 17:34:54.529776 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.529741 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-9pgcf\"" Apr 22 17:34:54.529833 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.529822 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"whereabouts-flatfile-config\"" Apr 22 17:34:54.529899 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.529833 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"cni-copy-resources\"" Apr 22 17:34:54.534428 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.534407 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9h6z\" (UniqueName: \"kubernetes.io/projected/cff34b01-53b3-4374-b73d-d3146e3c558d-kube-api-access-t9h6z\") pod \"network-metrics-daemon-zhl6m\" (UID: \"cff34b01-53b3-4374-b73d-d3146e3c558d\") " pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:34:54.534529 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.534441 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f57a62ce-7e60-4525-ac30-5348379d96f7-registration-dir\") pod \"aws-ebs-csi-driver-node-cvp79\" (UID: \"f57a62ce-7e60-4525-ac30-5348379d96f7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" Apr 22 17:34:54.534529 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.534512 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-run-ovn\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.534633 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.534533 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-host-cni-bin\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.534633 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.534548 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/24c19adf-c002-4c22-9098-f1f3be79bb45-ovnkube-config\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.534633 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.534607 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/24c19adf-c002-4c22-9098-f1f3be79bb45-ovn-node-metrics-cert\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.534781 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.534652 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-var-lib-kubelet\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.534781 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.534688 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-host-kubelet\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.534781 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.534724 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-etc-openvswitch\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.534781 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.534744 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-host-run-ovn-kubernetes\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.534781 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.534759 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-host-cni-netd\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.535004 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.534780 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.535004 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.534803 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-576vc\" (UniqueName: \"kubernetes.io/projected/24c19adf-c002-4c22-9098-f1f3be79bb45-kube-api-access-576vc\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.535004 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.534820 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fde77089-965b-4384-b901-80ddae4dbc8b-host-slash\") pod \"iptables-alerter-kctrb\" (UID: \"fde77089-965b-4384-b901-80ddae4dbc8b\") " pod="openshift-network-operator/iptables-alerter-kctrb" Apr 22 17:34:54.535004 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.534834 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nm26t\" (UniqueName: \"kubernetes.io/projected/fde77089-965b-4384-b901-80ddae4dbc8b-kube-api-access-nm26t\") pod \"iptables-alerter-kctrb\" (UID: \"fde77089-965b-4384-b901-80ddae4dbc8b\") " pod="openshift-network-operator/iptables-alerter-kctrb" Apr 22 17:34:54.535004 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.534854 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/f57a62ce-7e60-4525-ac30-5348379d96f7-etc-selinux\") pod \"aws-ebs-csi-driver-node-cvp79\" (UID: \"f57a62ce-7e60-4525-ac30-5348379d96f7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" Apr 22 17:34:54.535004 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.534929 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbd4h\" (UniqueName: \"kubernetes.io/projected/f57a62ce-7e60-4525-ac30-5348379d96f7-kube-api-access-mbd4h\") pod \"aws-ebs-csi-driver-node-cvp79\" (UID: \"f57a62ce-7e60-4525-ac30-5348379d96f7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" Apr 22 17:34:54.535218 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.534964 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-sys\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.535272 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.535234 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/5db20131-70d2-456b-a852-663399260efa-tmp-dir\") pod \"node-resolver-4m4hg\" (UID: \"5db20131-70d2-456b-a852-663399260efa\") " pod="openshift-dns/node-resolver-4m4hg" Apr 22 17:34:54.535306 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.535274 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/f57a62ce-7e60-4525-ac30-5348379d96f7-device-dir\") pod \"aws-ebs-csi-driver-node-cvp79\" (UID: \"f57a62ce-7e60-4525-ac30-5348379d96f7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" Apr 22 17:34:54.535306 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.535295 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-host-run-netns\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.535507 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.535314 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-run-systemd\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.535507 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.535474 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-etc-kubernetes\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.535637 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.535519 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-etc-systemd\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.535637 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.535550 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-lib-modules\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.535637 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.535580 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-host\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.535747 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.535631 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/bd9eb7e6-a061-4131-af0b-bfe904a22af3-etc-tuned\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.535747 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.535694 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f57a62ce-7e60-4525-ac30-5348379d96f7-kubelet-dir\") pod \"aws-ebs-csi-driver-node-cvp79\" (UID: \"f57a62ce-7e60-4525-ac30-5348379d96f7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" Apr 22 17:34:54.535747 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.535731 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-systemd-units\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.535927 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.535896 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-node-log\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.536051 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.535934 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-log-socket\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.536051 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.535966 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/fde77089-965b-4384-b901-80ddae4dbc8b-iptables-alerter-script\") pod \"iptables-alerter-kctrb\" (UID: \"fde77089-965b-4384-b901-80ddae4dbc8b\") " pod="openshift-network-operator/iptables-alerter-kctrb" Apr 22 17:34:54.536051 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.536021 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-etc-modprobe-d\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.536222 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.536064 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-etc-sysctl-conf\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.536222 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.536112 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/0288dec2-ebf5-4de0-82a8-4785af694080-agent-certs\") pod \"konnectivity-agent-vtsck\" (UID: \"0288dec2-ebf5-4de0-82a8-4785af694080\") " pod="kube-system/konnectivity-agent-vtsck" Apr 22 17:34:54.536222 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.536151 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/f57a62ce-7e60-4525-ac30-5348379d96f7-sys-fs\") pod \"aws-ebs-csi-driver-node-cvp79\" (UID: \"f57a62ce-7e60-4525-ac30-5348379d96f7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" Apr 22 17:34:54.536222 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.536182 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-host-slash\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.536222 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.536212 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-run-openvswitch\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.536441 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.536265 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/24c19adf-c002-4c22-9098-f1f3be79bb45-ovnkube-script-lib\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.536441 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.536297 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-etc-sysctl-d\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.536441 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.536325 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-run\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.536441 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.536342 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/5db20131-70d2-456b-a852-663399260efa-hosts-file\") pod \"node-resolver-4m4hg\" (UID: \"5db20131-70d2-456b-a852-663399260efa\") " pod="openshift-dns/node-resolver-4m4hg" Apr 22 17:34:54.536441 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.536361 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs\") pod \"network-metrics-daemon-zhl6m\" (UID: \"cff34b01-53b3-4374-b73d-d3146e3c558d\") " pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:34:54.536441 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.536380 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f57a62ce-7e60-4525-ac30-5348379d96f7-socket-dir\") pod \"aws-ebs-csi-driver-node-cvp79\" (UID: \"f57a62ce-7e60-4525-ac30-5348379d96f7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" Apr 22 17:34:54.536441 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.536411 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gtkc\" (UniqueName: \"kubernetes.io/projected/c858366e-d830-4c42-b18e-1eab94ea3113-kube-api-access-4gtkc\") pod \"network-check-target-p72t5\" (UID: \"c858366e-d830-4c42-b18e-1eab94ea3113\") " pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:34:54.536733 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.536601 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-var-lib-openvswitch\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.536733 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.536633 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/24c19adf-c002-4c22-9098-f1f3be79bb45-env-overrides\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.536733 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.536665 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-etc-sysconfig\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.536733 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.536700 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/bd9eb7e6-a061-4131-af0b-bfe904a22af3-tmp\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.536908 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.536750 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t772\" (UniqueName: \"kubernetes.io/projected/bd9eb7e6-a061-4131-af0b-bfe904a22af3-kube-api-access-7t772\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.536908 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.536782 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/0288dec2-ebf5-4de0-82a8-4785af694080-konnectivity-ca\") pod \"konnectivity-agent-vtsck\" (UID: \"0288dec2-ebf5-4de0-82a8-4785af694080\") " pod="kube-system/konnectivity-agent-vtsck" Apr 22 17:34:54.536908 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.536808 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kr54l\" (UniqueName: \"kubernetes.io/projected/5db20131-70d2-456b-a852-663399260efa-kube-api-access-kr54l\") pod \"node-resolver-4m4hg\" (UID: \"5db20131-70d2-456b-a852-663399260efa\") " pod="openshift-dns/node-resolver-4m4hg" Apr 22 17:34:54.566133 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.566096 2581 reflector.go:430] "Caches populated" type="*v1.CSIDriver" reflector="k8s.io/client-go/informers/factory.go:160" Apr 22 17:34:54.585466 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.585419 2581 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kubelet-serving" expiration="2028-04-21 17:29:53 +0000 UTC" deadline="2027-09-29 18:59:31.494693959 +0000 UTC" Apr 22 17:34:54.585466 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.585454 2581 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kubelet-serving" sleep="12601h24m36.909243843s" Apr 22 17:34:54.623521 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.623488 2581 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Apr 22 17:34:54.637089 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637063 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/f57a62ce-7e60-4525-ac30-5348379d96f7-sys-fs\") pod \"aws-ebs-csi-driver-node-cvp79\" (UID: \"f57a62ce-7e60-4525-ac30-5348379d96f7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" Apr 22 17:34:54.637089 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637098 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-run-openvswitch\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.637311 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637116 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-etc-sysctl-d\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.637311 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637143 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqxqb\" (UniqueName: \"kubernetes.io/projected/78313ad4-d7f9-4788-9293-7cc067465f5a-kube-api-access-nqxqb\") pod \"multus-additional-cni-plugins-pnh8b\" (UID: \"78313ad4-d7f9-4788-9293-7cc067465f5a\") " pod="openshift-multus/multus-additional-cni-plugins-pnh8b" Apr 22 17:34:54.637311 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637167 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-system-cni-dir\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.637311 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637200 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-cnibin\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.637311 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637208 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/f57a62ce-7e60-4525-ac30-5348379d96f7-sys-fs\") pod \"aws-ebs-csi-driver-node-cvp79\" (UID: \"f57a62ce-7e60-4525-ac30-5348379d96f7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" Apr 22 17:34:54.637311 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637196 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-run-openvswitch\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.637311 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637220 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-os-release\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.637311 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637279 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-host-run-k8s-cni-cncf-io\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.637311 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637290 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-etc-sysctl-d\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.637311 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637315 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs\") pod \"network-metrics-daemon-zhl6m\" (UID: \"cff34b01-53b3-4374-b73d-d3146e3c558d\") " pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:34:54.637734 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637343 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f57a62ce-7e60-4525-ac30-5348379d96f7-socket-dir\") pod \"aws-ebs-csi-driver-node-cvp79\" (UID: \"f57a62ce-7e60-4525-ac30-5348379d96f7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" Apr 22 17:34:54.637734 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637370 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-hostroot\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.637734 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637397 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/29e69548-2984-41bc-b4bd-8554ba75d702-multus-daemon-config\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.637734 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637425 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-kr54l\" (UniqueName: \"kubernetes.io/projected/5db20131-70d2-456b-a852-663399260efa-kube-api-access-kr54l\") pod \"node-resolver-4m4hg\" (UID: \"5db20131-70d2-456b-a852-663399260efa\") " pod="openshift-dns/node-resolver-4m4hg" Apr 22 17:34:54.637734 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637454 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/78313ad4-d7f9-4788-9293-7cc067465f5a-system-cni-dir\") pod \"multus-additional-cni-plugins-pnh8b\" (UID: \"78313ad4-d7f9-4788-9293-7cc067465f5a\") " pod="openshift-multus/multus-additional-cni-plugins-pnh8b" Apr 22 17:34:54.637734 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637480 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/78313ad4-d7f9-4788-9293-7cc067465f5a-cnibin\") pod \"multus-additional-cni-plugins-pnh8b\" (UID: \"78313ad4-d7f9-4788-9293-7cc067465f5a\") " pod="openshift-multus/multus-additional-cni-plugins-pnh8b" Apr 22 17:34:54.637734 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637508 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/bd9eb7e6-a061-4131-af0b-bfe904a22af3-tmp\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.637734 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637511 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f57a62ce-7e60-4525-ac30-5348379d96f7-socket-dir\") pod \"aws-ebs-csi-driver-node-cvp79\" (UID: \"f57a62ce-7e60-4525-ac30-5348379d96f7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" Apr 22 17:34:54.637734 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637533 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-host-cni-bin\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.637734 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637572 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/24c19adf-c002-4c22-9098-f1f3be79bb45-ovnkube-config\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.637734 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637598 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/24c19adf-c002-4c22-9098-f1f3be79bb45-ovn-node-metrics-cert\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.637734 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637626 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-var-lib-kubelet\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.637734 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637651 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-host\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.637734 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637678 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/bd9eb7e6-a061-4131-af0b-bfe904a22af3-etc-tuned\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.637734 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637707 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thc9s\" (UniqueName: \"kubernetes.io/projected/29e69548-2984-41bc-b4bd-8554ba75d702-kube-api-access-thc9s\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.637734 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637736 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-host-kubelet\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.638490 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637764 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-host-cni-netd\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.638490 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637791 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-576vc\" (UniqueName: \"kubernetes.io/projected/24c19adf-c002-4c22-9098-f1f3be79bb45-kube-api-access-576vc\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.638490 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637819 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nm26t\" (UniqueName: \"kubernetes.io/projected/fde77089-965b-4384-b901-80ddae4dbc8b-kube-api-access-nm26t\") pod \"iptables-alerter-kctrb\" (UID: \"fde77089-965b-4384-b901-80ddae4dbc8b\") " pod="openshift-network-operator/iptables-alerter-kctrb" Apr 22 17:34:54.638490 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637846 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-sys\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.638490 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637893 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/78313ad4-d7f9-4788-9293-7cc067465f5a-os-release\") pod \"multus-additional-cni-plugins-pnh8b\" (UID: \"78313ad4-d7f9-4788-9293-7cc067465f5a\") " pod="openshift-multus/multus-additional-cni-plugins-pnh8b" Apr 22 17:34:54.638490 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637927 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/f57a62ce-7e60-4525-ac30-5348379d96f7-etc-selinux\") pod \"aws-ebs-csi-driver-node-cvp79\" (UID: \"f57a62ce-7e60-4525-ac30-5348379d96f7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" Apr 22 17:34:54.638490 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637940 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-host\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.638490 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637957 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-mbd4h\" (UniqueName: \"kubernetes.io/projected/f57a62ce-7e60-4525-ac30-5348379d96f7-kube-api-access-mbd4h\") pod \"aws-ebs-csi-driver-node-cvp79\" (UID: \"f57a62ce-7e60-4525-ac30-5348379d96f7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" Apr 22 17:34:54.638490 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637987 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-multus-socket-dir-parent\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.638490 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.637990 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-host-cni-netd\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.638490 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638020 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/f57a62ce-7e60-4525-ac30-5348379d96f7-device-dir\") pod \"aws-ebs-csi-driver-node-cvp79\" (UID: \"f57a62ce-7e60-4525-ac30-5348379d96f7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" Apr 22 17:34:54.638490 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638089 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/f57a62ce-7e60-4525-ac30-5348379d96f7-device-dir\") pod \"aws-ebs-csi-driver-node-cvp79\" (UID: \"f57a62ce-7e60-4525-ac30-5348379d96f7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" Apr 22 17:34:54.638490 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638212 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-run-systemd\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.638490 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638257 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-etc-kubernetes\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.638490 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638258 2581 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Apr 22 17:34:54.638490 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638287 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-etc-systemd\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.638490 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638315 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/29e69548-2984-41bc-b4bd-8554ba75d702-cni-binary-copy\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.639410 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638344 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f57a62ce-7e60-4525-ac30-5348379d96f7-kubelet-dir\") pod \"aws-ebs-csi-driver-node-cvp79\" (UID: \"f57a62ce-7e60-4525-ac30-5348379d96f7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" Apr 22 17:34:54.639410 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638370 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-log-socket\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.639410 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638395 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-etc-modprobe-d\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.639410 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638420 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/5db20131-70d2-456b-a852-663399260efa-hosts-file\") pod \"node-resolver-4m4hg\" (UID: \"5db20131-70d2-456b-a852-663399260efa\") " pod="openshift-dns/node-resolver-4m4hg" Apr 22 17:34:54.639410 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638406 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-sys\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.639410 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638450 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8dedc3a1-a7e9-4684-b53d-a1b087d23cdb-host\") pod \"node-ca-6zmdm\" (UID: \"8dedc3a1-a7e9-4684-b53d-a1b087d23cdb\") " pod="openshift-image-registry/node-ca-6zmdm" Apr 22 17:34:54.639410 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638482 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/0288dec2-ebf5-4de0-82a8-4785af694080-agent-certs\") pod \"konnectivity-agent-vtsck\" (UID: \"0288dec2-ebf5-4de0-82a8-4785af694080\") " pod="kube-system/konnectivity-agent-vtsck" Apr 22 17:34:54.639410 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638508 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-host-slash\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.639410 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638536 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/24c19adf-c002-4c22-9098-f1f3be79bb45-ovnkube-script-lib\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.639410 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638553 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/f57a62ce-7e60-4525-ac30-5348379d96f7-etc-selinux\") pod \"aws-ebs-csi-driver-node-cvp79\" (UID: \"f57a62ce-7e60-4525-ac30-5348379d96f7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" Apr 22 17:34:54.639410 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638567 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/78313ad4-d7f9-4788-9293-7cc067465f5a-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-pnh8b\" (UID: \"78313ad4-d7f9-4788-9293-7cc067465f5a\") " pod="openshift-multus/multus-additional-cni-plugins-pnh8b" Apr 22 17:34:54.639410 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638595 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-4gtkc\" (UniqueName: \"kubernetes.io/projected/c858366e-d830-4c42-b18e-1eab94ea3113-kube-api-access-4gtkc\") pod \"network-check-target-p72t5\" (UID: \"c858366e-d830-4c42-b18e-1eab94ea3113\") " pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:34:54.639410 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638612 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-var-lib-openvswitch\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.639410 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638627 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/24c19adf-c002-4c22-9098-f1f3be79bb45-env-overrides\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.639410 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638643 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-etc-sysconfig\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.639410 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638690 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/78313ad4-d7f9-4788-9293-7cc067465f5a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-pnh8b\" (UID: \"78313ad4-d7f9-4788-9293-7cc067465f5a\") " pod="openshift-multus/multus-additional-cni-plugins-pnh8b" Apr 22 17:34:54.639410 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638707 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-host-var-lib-kubelet\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.640238 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638722 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-etc-kubernetes\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.640238 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638746 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/0288dec2-ebf5-4de0-82a8-4785af694080-konnectivity-ca\") pod \"konnectivity-agent-vtsck\" (UID: \"0288dec2-ebf5-4de0-82a8-4785af694080\") " pod="kube-system/konnectivity-agent-vtsck" Apr 22 17:34:54.640238 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638792 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8dedc3a1-a7e9-4684-b53d-a1b087d23cdb-serviceca\") pod \"node-ca-6zmdm\" (UID: \"8dedc3a1-a7e9-4684-b53d-a1b087d23cdb\") " pod="openshift-image-registry/node-ca-6zmdm" Apr 22 17:34:54.640238 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638852 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-host-kubelet\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.640238 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638872 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-var-lib-kubelet\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.640238 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:54.637628 2581 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 17:34:54.640238 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638908 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-run-systemd\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.640238 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638950 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-etc-kubernetes\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.640238 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:54.638979 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs podName:cff34b01-53b3-4374-b73d-d3146e3c558d nodeName:}" failed. No retries permitted until 2026-04-22 17:34:55.138945077 +0000 UTC m=+3.031640782 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs") pod "network-metrics-daemon-zhl6m" (UID: "cff34b01-53b3-4374-b73d-d3146e3c558d") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 17:34:54.640238 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.638990 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-etc-systemd\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.640238 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.639035 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f57a62ce-7e60-4525-ac30-5348379d96f7-kubelet-dir\") pod \"aws-ebs-csi-driver-node-cvp79\" (UID: \"f57a62ce-7e60-4525-ac30-5348379d96f7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" Apr 22 17:34:54.640238 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.639185 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-log-socket\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.640238 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.639274 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-etc-modprobe-d\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.640238 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.639320 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/5db20131-70d2-456b-a852-663399260efa-hosts-file\") pod \"node-resolver-4m4hg\" (UID: \"5db20131-70d2-456b-a852-663399260efa\") " pod="openshift-dns/node-resolver-4m4hg" Apr 22 17:34:54.640238 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.639400 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/78313ad4-d7f9-4788-9293-7cc067465f5a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-pnh8b\" (UID: \"78313ad4-d7f9-4788-9293-7cc067465f5a\") " pod="openshift-multus/multus-additional-cni-plugins-pnh8b" Apr 22 17:34:54.640238 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.639530 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-host-var-lib-cni-multus\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.640238 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.639648 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-7t772\" (UniqueName: \"kubernetes.io/projected/bd9eb7e6-a061-4131-af0b-bfe904a22af3-kube-api-access-7t772\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.641048 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.639681 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-t9h6z\" (UniqueName: \"kubernetes.io/projected/cff34b01-53b3-4374-b73d-d3146e3c558d-kube-api-access-t9h6z\") pod \"network-metrics-daemon-zhl6m\" (UID: \"cff34b01-53b3-4374-b73d-d3146e3c558d\") " pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:34:54.641048 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.639697 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-host-slash\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.641048 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.639708 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f57a62ce-7e60-4525-ac30-5348379d96f7-registration-dir\") pod \"aws-ebs-csi-driver-node-cvp79\" (UID: \"f57a62ce-7e60-4525-ac30-5348379d96f7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" Apr 22 17:34:54.641048 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.639735 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-run-ovn\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.641048 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.639785 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-host-run-netns\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.641048 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.639814 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-etc-openvswitch\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.641048 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.639842 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-host-run-ovn-kubernetes\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.641048 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.639871 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.641048 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.639960 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fde77089-965b-4384-b901-80ddae4dbc8b-host-slash\") pod \"iptables-alerter-kctrb\" (UID: \"fde77089-965b-4384-b901-80ddae4dbc8b\") " pod="openshift-network-operator/iptables-alerter-kctrb" Apr 22 17:34:54.641048 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.639987 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/5db20131-70d2-456b-a852-663399260efa-tmp-dir\") pod \"node-resolver-4m4hg\" (UID: \"5db20131-70d2-456b-a852-663399260efa\") " pod="openshift-dns/node-resolver-4m4hg" Apr 22 17:34:54.641048 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.640018 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-multus-conf-dir\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.641048 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.640045 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-host-run-multus-certs\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.641048 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.640072 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/78313ad4-d7f9-4788-9293-7cc067465f5a-cni-binary-copy\") pod \"multus-additional-cni-plugins-pnh8b\" (UID: \"78313ad4-d7f9-4788-9293-7cc067465f5a\") " pod="openshift-multus/multus-additional-cni-plugins-pnh8b" Apr 22 17:34:54.641048 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.640101 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-host-run-netns\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.641048 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.640139 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-lib-modules\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.641048 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.640168 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/24c19adf-c002-4c22-9098-f1f3be79bb45-env-overrides\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.641048 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.640238 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-host-cni-bin\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.641810 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.640165 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zv56\" (UniqueName: \"kubernetes.io/projected/8dedc3a1-a7e9-4684-b53d-a1b087d23cdb-kube-api-access-6zv56\") pod \"node-ca-6zmdm\" (UID: \"8dedc3a1-a7e9-4684-b53d-a1b087d23cdb\") " pod="openshift-image-registry/node-ca-6zmdm" Apr 22 17:34:54.641810 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.640269 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/0288dec2-ebf5-4de0-82a8-4785af694080-konnectivity-ca\") pod \"konnectivity-agent-vtsck\" (UID: \"0288dec2-ebf5-4de0-82a8-4785af694080\") " pod="kube-system/konnectivity-agent-vtsck" Apr 22 17:34:54.641810 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.640287 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-host-var-lib-cni-bin\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.641810 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.640323 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-systemd-units\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.641810 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.640355 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-node-log\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.641810 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.640399 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/fde77089-965b-4384-b901-80ddae4dbc8b-iptables-alerter-script\") pod \"iptables-alerter-kctrb\" (UID: \"fde77089-965b-4384-b901-80ddae4dbc8b\") " pod="openshift-network-operator/iptables-alerter-kctrb" Apr 22 17:34:54.641810 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.640437 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-etc-sysctl-conf\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.641810 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.640464 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-run\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.641810 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.640509 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-multus-cni-dir\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.641810 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.640620 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-etc-sysconfig\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.641810 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.640670 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-systemd-units\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.641810 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.640711 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-node-log\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.641810 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.640724 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/24c19adf-c002-4c22-9098-f1f3be79bb45-ovnkube-config\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.641810 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.639409 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-var-lib-openvswitch\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.641810 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.641182 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f57a62ce-7e60-4525-ac30-5348379d96f7-registration-dir\") pod \"aws-ebs-csi-driver-node-cvp79\" (UID: \"f57a62ce-7e60-4525-ac30-5348379d96f7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" Apr 22 17:34:54.641810 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.641229 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-run-ovn\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.641810 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.641288 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-etc-openvswitch\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.642404 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.641321 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-host-run-ovn-kubernetes\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.642404 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.641359 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.642404 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.641400 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fde77089-965b-4384-b901-80ddae4dbc8b-host-slash\") pod \"iptables-alerter-kctrb\" (UID: \"fde77089-965b-4384-b901-80ddae4dbc8b\") " pod="openshift-network-operator/iptables-alerter-kctrb" Apr 22 17:34:54.642404 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.641957 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/0288dec2-ebf5-4de0-82a8-4785af694080-agent-certs\") pod \"konnectivity-agent-vtsck\" (UID: \"0288dec2-ebf5-4de0-82a8-4785af694080\") " pod="kube-system/konnectivity-agent-vtsck" Apr 22 17:34:54.642404 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.641963 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/bd9eb7e6-a061-4131-af0b-bfe904a22af3-tmp\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.642404 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.642078 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-run\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.642404 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.642080 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/24c19adf-c002-4c22-9098-f1f3be79bb45-ovnkube-script-lib\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.642404 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.642092 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/5db20131-70d2-456b-a852-663399260efa-tmp-dir\") pod \"node-resolver-4m4hg\" (UID: \"5db20131-70d2-456b-a852-663399260efa\") " pod="openshift-dns/node-resolver-4m4hg" Apr 22 17:34:54.642404 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.642117 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-etc-sysctl-conf\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.642404 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.642171 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/24c19adf-c002-4c22-9098-f1f3be79bb45-host-run-netns\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.642404 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.642197 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bd9eb7e6-a061-4131-af0b-bfe904a22af3-lib-modules\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.642404 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.642302 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/bd9eb7e6-a061-4131-af0b-bfe904a22af3-etc-tuned\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.642404 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.642361 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/24c19adf-c002-4c22-9098-f1f3be79bb45-ovn-node-metrics-cert\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.643254 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.643229 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/fde77089-965b-4384-b901-80ddae4dbc8b-iptables-alerter-script\") pod \"iptables-alerter-kctrb\" (UID: \"fde77089-965b-4384-b901-80ddae4dbc8b\") " pod="openshift-network-operator/iptables-alerter-kctrb" Apr 22 17:34:54.647738 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:54.645030 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 22 17:34:54.647738 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:54.645058 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 22 17:34:54.647738 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:54.645075 2581 projected.go:194] Error preparing data for projected volume kube-api-access-4gtkc for pod openshift-network-diagnostics/network-check-target-p72t5: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 17:34:54.647738 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:54.645157 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c858366e-d830-4c42-b18e-1eab94ea3113-kube-api-access-4gtkc podName:c858366e-d830-4c42-b18e-1eab94ea3113 nodeName:}" failed. No retries permitted until 2026-04-22 17:34:55.145136302 +0000 UTC m=+3.037832021 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-4gtkc" (UniqueName: "kubernetes.io/projected/c858366e-d830-4c42-b18e-1eab94ea3113-kube-api-access-4gtkc") pod "network-check-target-p72t5" (UID: "c858366e-d830-4c42-b18e-1eab94ea3113") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 17:34:54.649512 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.649484 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-kr54l\" (UniqueName: \"kubernetes.io/projected/5db20131-70d2-456b-a852-663399260efa-kube-api-access-kr54l\") pod \"node-resolver-4m4hg\" (UID: \"5db20131-70d2-456b-a852-663399260efa\") " pod="openshift-dns/node-resolver-4m4hg" Apr 22 17:34:54.651301 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.651249 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-nm26t\" (UniqueName: \"kubernetes.io/projected/fde77089-965b-4384-b901-80ddae4dbc8b-kube-api-access-nm26t\") pod \"iptables-alerter-kctrb\" (UID: \"fde77089-965b-4384-b901-80ddae4dbc8b\") " pod="openshift-network-operator/iptables-alerter-kctrb" Apr 22 17:34:54.651497 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.651473 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t772\" (UniqueName: \"kubernetes.io/projected/bd9eb7e6-a061-4131-af0b-bfe904a22af3-kube-api-access-7t772\") pod \"tuned-4tdc7\" (UID: \"bd9eb7e6-a061-4131-af0b-bfe904a22af3\") " pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.651807 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.651788 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-576vc\" (UniqueName: \"kubernetes.io/projected/24c19adf-c002-4c22-9098-f1f3be79bb45-kube-api-access-576vc\") pod \"ovnkube-node-tn68c\" (UID: \"24c19adf-c002-4c22-9098-f1f3be79bb45\") " pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.652024 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.652004 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbd4h\" (UniqueName: \"kubernetes.io/projected/f57a62ce-7e60-4525-ac30-5348379d96f7-kube-api-access-mbd4h\") pod \"aws-ebs-csi-driver-node-cvp79\" (UID: \"f57a62ce-7e60-4525-ac30-5348379d96f7\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" Apr 22 17:34:54.653035 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.653013 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9h6z\" (UniqueName: \"kubernetes.io/projected/cff34b01-53b3-4374-b73d-d3146e3c558d-kube-api-access-t9h6z\") pod \"network-metrics-daemon-zhl6m\" (UID: \"cff34b01-53b3-4374-b73d-d3146e3c558d\") " pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:34:54.741278 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741185 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/78313ad4-d7f9-4788-9293-7cc067465f5a-system-cni-dir\") pod \"multus-additional-cni-plugins-pnh8b\" (UID: \"78313ad4-d7f9-4788-9293-7cc067465f5a\") " pod="openshift-multus/multus-additional-cni-plugins-pnh8b" Apr 22 17:34:54.741278 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741238 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/78313ad4-d7f9-4788-9293-7cc067465f5a-cnibin\") pod \"multus-additional-cni-plugins-pnh8b\" (UID: \"78313ad4-d7f9-4788-9293-7cc067465f5a\") " pod="openshift-multus/multus-additional-cni-plugins-pnh8b" Apr 22 17:34:54.741278 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741272 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-thc9s\" (UniqueName: \"kubernetes.io/projected/29e69548-2984-41bc-b4bd-8554ba75d702-kube-api-access-thc9s\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.741547 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741302 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/78313ad4-d7f9-4788-9293-7cc067465f5a-os-release\") pod \"multus-additional-cni-plugins-pnh8b\" (UID: \"78313ad4-d7f9-4788-9293-7cc067465f5a\") " pod="openshift-multus/multus-additional-cni-plugins-pnh8b" Apr 22 17:34:54.741547 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741305 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/78313ad4-d7f9-4788-9293-7cc067465f5a-system-cni-dir\") pod \"multus-additional-cni-plugins-pnh8b\" (UID: \"78313ad4-d7f9-4788-9293-7cc067465f5a\") " pod="openshift-multus/multus-additional-cni-plugins-pnh8b" Apr 22 17:34:54.741547 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741330 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-multus-socket-dir-parent\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.741547 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741371 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/78313ad4-d7f9-4788-9293-7cc067465f5a-cnibin\") pod \"multus-additional-cni-plugins-pnh8b\" (UID: \"78313ad4-d7f9-4788-9293-7cc067465f5a\") " pod="openshift-multus/multus-additional-cni-plugins-pnh8b" Apr 22 17:34:54.741547 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741388 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/29e69548-2984-41bc-b4bd-8554ba75d702-cni-binary-copy\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.741547 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741392 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-multus-socket-dir-parent\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.741547 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741430 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/78313ad4-d7f9-4788-9293-7cc067465f5a-os-release\") pod \"multus-additional-cni-plugins-pnh8b\" (UID: \"78313ad4-d7f9-4788-9293-7cc067465f5a\") " pod="openshift-multus/multus-additional-cni-plugins-pnh8b" Apr 22 17:34:54.741547 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741436 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8dedc3a1-a7e9-4684-b53d-a1b087d23cdb-host\") pod \"node-ca-6zmdm\" (UID: \"8dedc3a1-a7e9-4684-b53d-a1b087d23cdb\") " pod="openshift-image-registry/node-ca-6zmdm" Apr 22 17:34:54.741547 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741462 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8dedc3a1-a7e9-4684-b53d-a1b087d23cdb-host\") pod \"node-ca-6zmdm\" (UID: \"8dedc3a1-a7e9-4684-b53d-a1b087d23cdb\") " pod="openshift-image-registry/node-ca-6zmdm" Apr 22 17:34:54.741547 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741495 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/78313ad4-d7f9-4788-9293-7cc067465f5a-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-pnh8b\" (UID: \"78313ad4-d7f9-4788-9293-7cc067465f5a\") " pod="openshift-multus/multus-additional-cni-plugins-pnh8b" Apr 22 17:34:54.741547 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741544 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/78313ad4-d7f9-4788-9293-7cc067465f5a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-pnh8b\" (UID: \"78313ad4-d7f9-4788-9293-7cc067465f5a\") " pod="openshift-multus/multus-additional-cni-plugins-pnh8b" Apr 22 17:34:54.742066 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741569 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-host-var-lib-kubelet\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.742066 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741593 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-etc-kubernetes\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.742066 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741619 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8dedc3a1-a7e9-4684-b53d-a1b087d23cdb-serviceca\") pod \"node-ca-6zmdm\" (UID: \"8dedc3a1-a7e9-4684-b53d-a1b087d23cdb\") " pod="openshift-image-registry/node-ca-6zmdm" Apr 22 17:34:54.742066 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741647 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/78313ad4-d7f9-4788-9293-7cc067465f5a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-pnh8b\" (UID: \"78313ad4-d7f9-4788-9293-7cc067465f5a\") " pod="openshift-multus/multus-additional-cni-plugins-pnh8b" Apr 22 17:34:54.742066 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741673 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-host-var-lib-cni-multus\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.742066 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741702 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-host-run-netns\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.742066 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741713 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-etc-kubernetes\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.742066 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741732 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-multus-conf-dir\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.742066 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741718 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/78313ad4-d7f9-4788-9293-7cc067465f5a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-pnh8b\" (UID: \"78313ad4-d7f9-4788-9293-7cc067465f5a\") " pod="openshift-multus/multus-additional-cni-plugins-pnh8b" Apr 22 17:34:54.742066 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741773 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-host-var-lib-kubelet\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.742066 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741786 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-host-var-lib-cni-multus\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.742066 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741815 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-host-run-netns\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.742066 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741825 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-multus-conf-dir\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.742066 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741854 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-host-run-multus-certs\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.742066 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741903 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/78313ad4-d7f9-4788-9293-7cc067465f5a-cni-binary-copy\") pod \"multus-additional-cni-plugins-pnh8b\" (UID: \"78313ad4-d7f9-4788-9293-7cc067465f5a\") " pod="openshift-multus/multus-additional-cni-plugins-pnh8b" Apr 22 17:34:54.742066 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741920 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/29e69548-2984-41bc-b4bd-8554ba75d702-cni-binary-copy\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.742066 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741932 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6zv56\" (UniqueName: \"kubernetes.io/projected/8dedc3a1-a7e9-4684-b53d-a1b087d23cdb-kube-api-access-6zv56\") pod \"node-ca-6zmdm\" (UID: \"8dedc3a1-a7e9-4684-b53d-a1b087d23cdb\") " pod="openshift-image-registry/node-ca-6zmdm" Apr 22 17:34:54.742066 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741969 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-host-var-lib-cni-bin\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.742826 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.741991 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-multus-cni-dir\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.742826 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.742010 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nqxqb\" (UniqueName: \"kubernetes.io/projected/78313ad4-d7f9-4788-9293-7cc067465f5a-kube-api-access-nqxqb\") pod \"multus-additional-cni-plugins-pnh8b\" (UID: \"78313ad4-d7f9-4788-9293-7cc067465f5a\") " pod="openshift-multus/multus-additional-cni-plugins-pnh8b" Apr 22 17:34:54.742826 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.742030 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-system-cni-dir\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.742826 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.742044 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-cnibin\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.742826 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.742058 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-os-release\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.742826 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.742072 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-host-run-k8s-cni-cncf-io\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.742826 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.742105 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-hostroot\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.742826 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.742110 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/78313ad4-d7f9-4788-9293-7cc067465f5a-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-pnh8b\" (UID: \"78313ad4-d7f9-4788-9293-7cc067465f5a\") " pod="openshift-multus/multus-additional-cni-plugins-pnh8b" Apr 22 17:34:54.742826 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.742120 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/29e69548-2984-41bc-b4bd-8554ba75d702-multus-daemon-config\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.742826 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.742168 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-host-run-k8s-cni-cncf-io\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.742826 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.742202 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-cnibin\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.742826 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.742222 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/78313ad4-d7f9-4788-9293-7cc067465f5a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-pnh8b\" (UID: \"78313ad4-d7f9-4788-9293-7cc067465f5a\") " pod="openshift-multus/multus-additional-cni-plugins-pnh8b" Apr 22 17:34:54.742826 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.742225 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8dedc3a1-a7e9-4684-b53d-a1b087d23cdb-serviceca\") pod \"node-ca-6zmdm\" (UID: \"8dedc3a1-a7e9-4684-b53d-a1b087d23cdb\") " pod="openshift-image-registry/node-ca-6zmdm" Apr 22 17:34:54.742826 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.742247 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-os-release\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.742826 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.742258 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-host-run-multus-certs\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.742826 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.742275 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-host-var-lib-cni-bin\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.742826 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.742319 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-multus-cni-dir\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.742826 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.742332 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-system-cni-dir\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.743568 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.742347 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/29e69548-2984-41bc-b4bd-8554ba75d702-hostroot\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.743568 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.742703 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/29e69548-2984-41bc-b4bd-8554ba75d702-multus-daemon-config\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.743568 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.743058 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/78313ad4-d7f9-4788-9293-7cc067465f5a-cni-binary-copy\") pod \"multus-additional-cni-plugins-pnh8b\" (UID: \"78313ad4-d7f9-4788-9293-7cc067465f5a\") " pod="openshift-multus/multus-additional-cni-plugins-pnh8b" Apr 22 17:34:54.750598 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.750569 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqxqb\" (UniqueName: \"kubernetes.io/projected/78313ad4-d7f9-4788-9293-7cc067465f5a-kube-api-access-nqxqb\") pod \"multus-additional-cni-plugins-pnh8b\" (UID: \"78313ad4-d7f9-4788-9293-7cc067465f5a\") " pod="openshift-multus/multus-additional-cni-plugins-pnh8b" Apr 22 17:34:54.750743 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.750668 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zv56\" (UniqueName: \"kubernetes.io/projected/8dedc3a1-a7e9-4684-b53d-a1b087d23cdb-kube-api-access-6zv56\") pod \"node-ca-6zmdm\" (UID: \"8dedc3a1-a7e9-4684-b53d-a1b087d23cdb\") " pod="openshift-image-registry/node-ca-6zmdm" Apr 22 17:34:54.750803 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.750776 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-thc9s\" (UniqueName: \"kubernetes.io/projected/29e69548-2984-41bc-b4bd-8554ba75d702-kube-api-access-thc9s\") pod \"multus-nwhzz\" (UID: \"29e69548-2984-41bc-b4bd-8554ba75d702\") " pod="openshift-multus/multus-nwhzz" Apr 22 17:34:54.828024 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.827982 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:34:54.835527 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.834973 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-kctrb" Apr 22 17:34:54.843994 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.843961 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" Apr 22 17:34:54.849637 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.849612 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" Apr 22 17:34:54.857240 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.857216 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-4m4hg" Apr 22 17:34:54.863852 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.863827 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/konnectivity-agent-vtsck" Apr 22 17:34:54.869525 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.869501 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-6zmdm" Apr 22 17:34:54.876096 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.876072 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-pnh8b" Apr 22 17:34:54.882764 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:54.882742 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-nwhzz" Apr 22 17:34:55.145271 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:55.145184 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-4gtkc\" (UniqueName: \"kubernetes.io/projected/c858366e-d830-4c42-b18e-1eab94ea3113-kube-api-access-4gtkc\") pod \"network-check-target-p72t5\" (UID: \"c858366e-d830-4c42-b18e-1eab94ea3113\") " pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:34:55.145271 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:55.145260 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs\") pod \"network-metrics-daemon-zhl6m\" (UID: \"cff34b01-53b3-4374-b73d-d3146e3c558d\") " pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:34:55.145466 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:55.145360 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 22 17:34:55.145466 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:55.145380 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 22 17:34:55.145466 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:55.145382 2581 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 17:34:55.145466 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:55.145389 2581 projected.go:194] Error preparing data for projected volume kube-api-access-4gtkc for pod openshift-network-diagnostics/network-check-target-p72t5: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 17:34:55.145466 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:55.145445 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c858366e-d830-4c42-b18e-1eab94ea3113-kube-api-access-4gtkc podName:c858366e-d830-4c42-b18e-1eab94ea3113 nodeName:}" failed. No retries permitted until 2026-04-22 17:34:56.145430331 +0000 UTC m=+4.038126033 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-4gtkc" (UniqueName: "kubernetes.io/projected/c858366e-d830-4c42-b18e-1eab94ea3113-kube-api-access-4gtkc") pod "network-check-target-p72t5" (UID: "c858366e-d830-4c42-b18e-1eab94ea3113") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 17:34:55.145466 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:55.145459 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs podName:cff34b01-53b3-4374-b73d-d3146e3c558d nodeName:}" failed. No retries permitted until 2026-04-22 17:34:56.145452906 +0000 UTC m=+4.038148608 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs") pod "network-metrics-daemon-zhl6m" (UID: "cff34b01-53b3-4374-b73d-d3146e3c558d") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 17:34:55.188612 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:55.188462 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24c19adf_c002_4c22_9098_f1f3be79bb45.slice/crio-187759799d81d97f0dd0ef17b0627926879f9ac15bd4898d71e00e9d1b17fc9b WatchSource:0}: Error finding container 187759799d81d97f0dd0ef17b0627926879f9ac15bd4898d71e00e9d1b17fc9b: Status 404 returned error can't find the container with id 187759799d81d97f0dd0ef17b0627926879f9ac15bd4898d71e00e9d1b17fc9b Apr 22 17:34:55.191504 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:55.191469 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf57a62ce_7e60_4525_ac30_5348379d96f7.slice/crio-dada65abe29155d0afe0ed74d33bfa339b62adc9c4ab63737eabba3402bc0f5a WatchSource:0}: Error finding container dada65abe29155d0afe0ed74d33bfa339b62adc9c4ab63737eabba3402bc0f5a: Status 404 returned error can't find the container with id dada65abe29155d0afe0ed74d33bfa339b62adc9c4ab63737eabba3402bc0f5a Apr 22 17:34:55.193470 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:55.193376 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod78313ad4_d7f9_4788_9293_7cc067465f5a.slice/crio-6af021a9f7f25362c86be068c9a44a7c224ac37596861efbd980b82c0eb2a1fa WatchSource:0}: Error finding container 6af021a9f7f25362c86be068c9a44a7c224ac37596861efbd980b82c0eb2a1fa: Status 404 returned error can't find the container with id 6af021a9f7f25362c86be068c9a44a7c224ac37596861efbd980b82c0eb2a1fa Apr 22 17:34:55.195354 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:55.195334 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5db20131_70d2_456b_a852_663399260efa.slice/crio-41d1e8bb020c55e243836409831f6d010f1b76afdc139905f92e48118dc4684e WatchSource:0}: Error finding container 41d1e8bb020c55e243836409831f6d010f1b76afdc139905f92e48118dc4684e: Status 404 returned error can't find the container with id 41d1e8bb020c55e243836409831f6d010f1b76afdc139905f92e48118dc4684e Apr 22 17:34:55.196152 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:55.196128 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0288dec2_ebf5_4de0_82a8_4785af694080.slice/crio-c02a63a77be3a2a33ae280b5e23bdeb01892a06da0405e6220350df3a1525197 WatchSource:0}: Error finding container c02a63a77be3a2a33ae280b5e23bdeb01892a06da0405e6220350df3a1525197: Status 404 returned error can't find the container with id c02a63a77be3a2a33ae280b5e23bdeb01892a06da0405e6220350df3a1525197 Apr 22 17:34:55.197599 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:55.197580 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfde77089_965b_4384_b901_80ddae4dbc8b.slice/crio-d77ec947e882e6d86f1a6924408ba80ff1f35e07dcabd299253bb8c427558e16 WatchSource:0}: Error finding container d77ec947e882e6d86f1a6924408ba80ff1f35e07dcabd299253bb8c427558e16: Status 404 returned error can't find the container with id d77ec947e882e6d86f1a6924408ba80ff1f35e07dcabd299253bb8c427558e16 Apr 22 17:34:55.197860 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:55.197828 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29e69548_2984_41bc_b4bd_8554ba75d702.slice/crio-5ff0dab4b68d08aa67ffd0f93a2e5366c5259f6774930a05d962b48a7a8d2772 WatchSource:0}: Error finding container 5ff0dab4b68d08aa67ffd0f93a2e5366c5259f6774930a05d962b48a7a8d2772: Status 404 returned error can't find the container with id 5ff0dab4b68d08aa67ffd0f93a2e5366c5259f6774930a05d962b48a7a8d2772 Apr 22 17:34:55.199250 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:55.199135 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8dedc3a1_a7e9_4684_b53d_a1b087d23cdb.slice/crio-5ca22cf350ba853315a0b0e5b53857b49e36e0fc45433fe907fe63e2ab620fb7 WatchSource:0}: Error finding container 5ca22cf350ba853315a0b0e5b53857b49e36e0fc45433fe907fe63e2ab620fb7: Status 404 returned error can't find the container with id 5ca22cf350ba853315a0b0e5b53857b49e36e0fc45433fe907fe63e2ab620fb7 Apr 22 17:34:55.199825 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:34:55.199808 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd9eb7e6_a061_4131_af0b_bfe904a22af3.slice/crio-33f4c493878a30c7504d75ae03af66591e98623e6c9c4be6503c6ee556bb6c4b WatchSource:0}: Error finding container 33f4c493878a30c7504d75ae03af66591e98623e6c9c4be6503c6ee556bb6c4b: Status 404 returned error can't find the container with id 33f4c493878a30c7504d75ae03af66591e98623e6c9c4be6503c6ee556bb6c4b Apr 22 17:34:55.586710 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:55.586424 2581 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kubelet-serving" expiration="2028-04-21 17:29:53 +0000 UTC" deadline="2028-01-04 11:15:45.879745798 +0000 UTC" Apr 22 17:34:55.586710 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:55.586702 2581 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kubelet-serving" sleep="14921h40m50.293049539s" Apr 22 17:34:55.636012 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:55.635401 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:34:55.636012 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:55.635556 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zhl6m" podUID="cff34b01-53b3-4374-b73d-d3146e3c558d" Apr 22 17:34:55.655350 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:55.655309 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-6zmdm" event={"ID":"8dedc3a1-a7e9-4684-b53d-a1b087d23cdb","Type":"ContainerStarted","Data":"5ca22cf350ba853315a0b0e5b53857b49e36e0fc45433fe907fe63e2ab620fb7"} Apr 22 17:34:55.658970 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:55.658903 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nwhzz" event={"ID":"29e69548-2984-41bc-b4bd-8554ba75d702","Type":"ContainerStarted","Data":"5ff0dab4b68d08aa67ffd0f93a2e5366c5259f6774930a05d962b48a7a8d2772"} Apr 22 17:34:55.666171 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:55.666097 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-kctrb" event={"ID":"fde77089-965b-4384-b901-80ddae4dbc8b","Type":"ContainerStarted","Data":"d77ec947e882e6d86f1a6924408ba80ff1f35e07dcabd299253bb8c427558e16"} Apr 22 17:34:55.671853 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:55.671805 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" event={"ID":"24c19adf-c002-4c22-9098-f1f3be79bb45","Type":"ContainerStarted","Data":"187759799d81d97f0dd0ef17b0627926879f9ac15bd4898d71e00e9d1b17fc9b"} Apr 22 17:34:55.680510 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:55.680473 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" event={"ID":"bd9eb7e6-a061-4131-af0b-bfe904a22af3","Type":"ContainerStarted","Data":"33f4c493878a30c7504d75ae03af66591e98623e6c9c4be6503c6ee556bb6c4b"} Apr 22 17:34:55.683204 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:55.683151 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/konnectivity-agent-vtsck" event={"ID":"0288dec2-ebf5-4de0-82a8-4785af694080","Type":"ContainerStarted","Data":"c02a63a77be3a2a33ae280b5e23bdeb01892a06da0405e6220350df3a1525197"} Apr 22 17:34:55.685290 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:55.685225 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-4m4hg" event={"ID":"5db20131-70d2-456b-a852-663399260efa","Type":"ContainerStarted","Data":"41d1e8bb020c55e243836409831f6d010f1b76afdc139905f92e48118dc4684e"} Apr 22 17:34:55.686941 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:55.686870 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pnh8b" event={"ID":"78313ad4-d7f9-4788-9293-7cc067465f5a","Type":"ContainerStarted","Data":"6af021a9f7f25362c86be068c9a44a7c224ac37596861efbd980b82c0eb2a1fa"} Apr 22 17:34:55.692571 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:55.692513 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" event={"ID":"f57a62ce-7e60-4525-ac30-5348379d96f7","Type":"ContainerStarted","Data":"dada65abe29155d0afe0ed74d33bfa339b62adc9c4ab63737eabba3402bc0f5a"} Apr 22 17:34:55.707588 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:55.704732 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-138-57.ec2.internal" event={"ID":"8df1f5cf5626886b4ad3778d0dba9ac9","Type":"ContainerStarted","Data":"3384f6c98f8ed7cc4024d74ed2694a0e2cd5653bc3f061cf361c496846e87bfe"} Apr 22 17:34:55.718897 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:55.718817 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-proxy-ip-10-0-138-57.ec2.internal" podStartSLOduration=1.718797669 podStartE2EDuration="1.718797669s" podCreationTimestamp="2026-04-22 17:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-22 17:34:55.718548967 +0000 UTC m=+3.611244693" watchObservedRunningTime="2026-04-22 17:34:55.718797669 +0000 UTC m=+3.611493394" Apr 22 17:34:56.154356 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:56.153538 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs\") pod \"network-metrics-daemon-zhl6m\" (UID: \"cff34b01-53b3-4374-b73d-d3146e3c558d\") " pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:34:56.154356 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:56.153604 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-4gtkc\" (UniqueName: \"kubernetes.io/projected/c858366e-d830-4c42-b18e-1eab94ea3113-kube-api-access-4gtkc\") pod \"network-check-target-p72t5\" (UID: \"c858366e-d830-4c42-b18e-1eab94ea3113\") " pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:34:56.154356 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:56.153737 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 22 17:34:56.154356 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:56.153757 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 22 17:34:56.154356 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:56.153771 2581 projected.go:194] Error preparing data for projected volume kube-api-access-4gtkc for pod openshift-network-diagnostics/network-check-target-p72t5: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 17:34:56.154356 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:56.153829 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c858366e-d830-4c42-b18e-1eab94ea3113-kube-api-access-4gtkc podName:c858366e-d830-4c42-b18e-1eab94ea3113 nodeName:}" failed. No retries permitted until 2026-04-22 17:34:58.153810273 +0000 UTC m=+6.046505979 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-4gtkc" (UniqueName: "kubernetes.io/projected/c858366e-d830-4c42-b18e-1eab94ea3113-kube-api-access-4gtkc") pod "network-check-target-p72t5" (UID: "c858366e-d830-4c42-b18e-1eab94ea3113") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 17:34:56.154356 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:56.154263 2581 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 17:34:56.154356 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:56.154315 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs podName:cff34b01-53b3-4374-b73d-d3146e3c558d nodeName:}" failed. No retries permitted until 2026-04-22 17:34:58.154300089 +0000 UTC m=+6.046995797 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs") pod "network-metrics-daemon-zhl6m" (UID: "cff34b01-53b3-4374-b73d-d3146e3c558d") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 17:34:56.635163 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:56.635125 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:34:56.635615 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:56.635261 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-p72t5" podUID="c858366e-d830-4c42-b18e-1eab94ea3113" Apr 22 17:34:56.718233 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:56.718194 2581 generic.go:358] "Generic (PLEG): container finished" podID="b835e279c462c30381fc4ead00ea7937" containerID="d0232701c82fd33672fdc8803dd9b637f3a745a71edd30899014e17516430cab" exitCode=0 Apr 22 17:34:56.719007 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:56.718731 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-138-57.ec2.internal" event={"ID":"b835e279c462c30381fc4ead00ea7937","Type":"ContainerDied","Data":"d0232701c82fd33672fdc8803dd9b637f3a745a71edd30899014e17516430cab"} Apr 22 17:34:57.635111 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:57.635075 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:34:57.635300 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:57.635226 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zhl6m" podUID="cff34b01-53b3-4374-b73d-d3146e3c558d" Apr 22 17:34:57.740452 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:57.738050 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-138-57.ec2.internal" event={"ID":"b835e279c462c30381fc4ead00ea7937","Type":"ContainerStarted","Data":"319d509c62818fc399e47769f2f9bd6330910f39e3048576623d7228568c519d"} Apr 22 17:34:57.751706 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:57.751651 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-138-57.ec2.internal" podStartSLOduration=3.751633424 podStartE2EDuration="3.751633424s" podCreationTimestamp="2026-04-22 17:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-22 17:34:57.751327852 +0000 UTC m=+5.644023577" watchObservedRunningTime="2026-04-22 17:34:57.751633424 +0000 UTC m=+5.644329149" Apr 22 17:34:58.175381 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:58.174585 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-4gtkc\" (UniqueName: \"kubernetes.io/projected/c858366e-d830-4c42-b18e-1eab94ea3113-kube-api-access-4gtkc\") pod \"network-check-target-p72t5\" (UID: \"c858366e-d830-4c42-b18e-1eab94ea3113\") " pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:34:58.175381 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:58.174654 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs\") pod \"network-metrics-daemon-zhl6m\" (UID: \"cff34b01-53b3-4374-b73d-d3146e3c558d\") " pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:34:58.175381 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:58.174781 2581 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 17:34:58.175381 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:58.174837 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs podName:cff34b01-53b3-4374-b73d-d3146e3c558d nodeName:}" failed. No retries permitted until 2026-04-22 17:35:02.174818413 +0000 UTC m=+10.067514138 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs") pod "network-metrics-daemon-zhl6m" (UID: "cff34b01-53b3-4374-b73d-d3146e3c558d") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 17:34:58.175381 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:58.175249 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 22 17:34:58.175381 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:58.175267 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 22 17:34:58.175381 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:58.175278 2581 projected.go:194] Error preparing data for projected volume kube-api-access-4gtkc for pod openshift-network-diagnostics/network-check-target-p72t5: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 17:34:58.175381 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:58.175321 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c858366e-d830-4c42-b18e-1eab94ea3113-kube-api-access-4gtkc podName:c858366e-d830-4c42-b18e-1eab94ea3113 nodeName:}" failed. No retries permitted until 2026-04-22 17:35:02.175307165 +0000 UTC m=+10.068002871 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-4gtkc" (UniqueName: "kubernetes.io/projected/c858366e-d830-4c42-b18e-1eab94ea3113-kube-api-access-4gtkc") pod "network-check-target-p72t5" (UID: "c858366e-d830-4c42-b18e-1eab94ea3113") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 17:34:58.635046 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:58.635011 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:34:58.635233 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:58.635145 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-p72t5" podUID="c858366e-d830-4c42-b18e-1eab94ea3113" Apr 22 17:34:59.634895 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:34:59.634853 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:34:59.635349 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:34:59.635021 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zhl6m" podUID="cff34b01-53b3-4374-b73d-d3146e3c558d" Apr 22 17:35:00.637210 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:00.637171 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:35:00.637675 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:00.637343 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-p72t5" podUID="c858366e-d830-4c42-b18e-1eab94ea3113" Apr 22 17:35:01.635485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:01.634934 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:35:01.635485 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:01.635088 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zhl6m" podUID="cff34b01-53b3-4374-b73d-d3146e3c558d" Apr 22 17:35:02.208451 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:02.207918 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs\") pod \"network-metrics-daemon-zhl6m\" (UID: \"cff34b01-53b3-4374-b73d-d3146e3c558d\") " pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:35:02.208451 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:02.207974 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-4gtkc\" (UniqueName: \"kubernetes.io/projected/c858366e-d830-4c42-b18e-1eab94ea3113-kube-api-access-4gtkc\") pod \"network-check-target-p72t5\" (UID: \"c858366e-d830-4c42-b18e-1eab94ea3113\") " pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:35:02.208451 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:02.208075 2581 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 17:35:02.208451 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:02.208090 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 22 17:35:02.208451 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:02.208103 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 22 17:35:02.208451 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:02.208115 2581 projected.go:194] Error preparing data for projected volume kube-api-access-4gtkc for pod openshift-network-diagnostics/network-check-target-p72t5: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 17:35:02.208451 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:02.208150 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs podName:cff34b01-53b3-4374-b73d-d3146e3c558d nodeName:}" failed. No retries permitted until 2026-04-22 17:35:10.208129288 +0000 UTC m=+18.100824990 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs") pod "network-metrics-daemon-zhl6m" (UID: "cff34b01-53b3-4374-b73d-d3146e3c558d") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 17:35:02.208451 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:02.208167 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c858366e-d830-4c42-b18e-1eab94ea3113-kube-api-access-4gtkc podName:c858366e-d830-4c42-b18e-1eab94ea3113 nodeName:}" failed. No retries permitted until 2026-04-22 17:35:10.208158204 +0000 UTC m=+18.100853906 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-4gtkc" (UniqueName: "kubernetes.io/projected/c858366e-d830-4c42-b18e-1eab94ea3113-kube-api-access-4gtkc") pod "network-check-target-p72t5" (UID: "c858366e-d830-4c42-b18e-1eab94ea3113") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 17:35:02.636003 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:02.635968 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:35:02.636206 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:02.636086 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-p72t5" podUID="c858366e-d830-4c42-b18e-1eab94ea3113" Apr 22 17:35:03.634684 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:03.634651 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:35:03.635087 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:03.634774 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zhl6m" podUID="cff34b01-53b3-4374-b73d-d3146e3c558d" Apr 22 17:35:04.635042 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:04.634963 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:35:04.635441 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:04.635094 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-p72t5" podUID="c858366e-d830-4c42-b18e-1eab94ea3113" Apr 22 17:35:05.635301 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:05.635254 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:35:05.635702 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:05.635388 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zhl6m" podUID="cff34b01-53b3-4374-b73d-d3146e3c558d" Apr 22 17:35:06.635532 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:06.635494 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:35:06.636010 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:06.635621 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-p72t5" podUID="c858366e-d830-4c42-b18e-1eab94ea3113" Apr 22 17:35:07.634788 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:07.634747 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:35:07.635023 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:07.634919 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zhl6m" podUID="cff34b01-53b3-4374-b73d-d3146e3c558d" Apr 22 17:35:08.634979 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:08.634944 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:35:08.635443 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:08.635083 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-p72t5" podUID="c858366e-d830-4c42-b18e-1eab94ea3113" Apr 22 17:35:09.635247 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:09.635212 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:35:09.635690 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:09.635351 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zhl6m" podUID="cff34b01-53b3-4374-b73d-d3146e3c558d" Apr 22 17:35:10.268130 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:10.268094 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-4gtkc\" (UniqueName: \"kubernetes.io/projected/c858366e-d830-4c42-b18e-1eab94ea3113-kube-api-access-4gtkc\") pod \"network-check-target-p72t5\" (UID: \"c858366e-d830-4c42-b18e-1eab94ea3113\") " pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:35:10.268339 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:10.268167 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs\") pod \"network-metrics-daemon-zhl6m\" (UID: \"cff34b01-53b3-4374-b73d-d3146e3c558d\") " pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:35:10.268339 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:10.268259 2581 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 17:35:10.268339 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:10.268288 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 22 17:35:10.268339 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:10.268315 2581 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 22 17:35:10.268339 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:10.268325 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs podName:cff34b01-53b3-4374-b73d-d3146e3c558d nodeName:}" failed. No retries permitted until 2026-04-22 17:35:26.268303095 +0000 UTC m=+34.160998812 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs") pod "network-metrics-daemon-zhl6m" (UID: "cff34b01-53b3-4374-b73d-d3146e3c558d") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 22 17:35:10.268339 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:10.268327 2581 projected.go:194] Error preparing data for projected volume kube-api-access-4gtkc for pod openshift-network-diagnostics/network-check-target-p72t5: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 17:35:10.268586 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:10.268382 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c858366e-d830-4c42-b18e-1eab94ea3113-kube-api-access-4gtkc podName:c858366e-d830-4c42-b18e-1eab94ea3113 nodeName:}" failed. No retries permitted until 2026-04-22 17:35:26.268365525 +0000 UTC m=+34.161061227 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-4gtkc" (UniqueName: "kubernetes.io/projected/c858366e-d830-4c42-b18e-1eab94ea3113-kube-api-access-4gtkc") pod "network-check-target-p72t5" (UID: "c858366e-d830-4c42-b18e-1eab94ea3113") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 22 17:35:10.635396 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:10.635296 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:35:10.635861 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:10.635429 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-p72t5" podUID="c858366e-d830-4c42-b18e-1eab94ea3113" Apr 22 17:35:11.635309 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:11.635271 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:35:11.635524 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:11.635431 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zhl6m" podUID="cff34b01-53b3-4374-b73d-d3146e3c558d" Apr 22 17:35:12.635445 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:12.635212 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:35:12.635591 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:12.635561 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-p72t5" podUID="c858366e-d830-4c42-b18e-1eab94ea3113" Apr 22 17:35:12.763604 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:12.763129 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" event={"ID":"bd9eb7e6-a061-4131-af0b-bfe904a22af3","Type":"ContainerStarted","Data":"ba5d58a7d6b874dc14712b0da0605f7ea9a989634efa0dd6e6e2bededea02275"} Apr 22 17:35:12.764394 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:12.764360 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/konnectivity-agent-vtsck" event={"ID":"0288dec2-ebf5-4de0-82a8-4785af694080","Type":"ContainerStarted","Data":"9e27acfda87e602775c6b91a03dc2ae1febf64ca36b024f86dbcd84fec0bfd10"} Apr 22 17:35:12.765595 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:12.765568 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-4m4hg" event={"ID":"5db20131-70d2-456b-a852-663399260efa","Type":"ContainerStarted","Data":"bed651355d2f58cc62b0dae9119345792ec90de3316462370fed6ea6c95712b3"} Apr 22 17:35:12.766645 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:12.766617 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pnh8b" event={"ID":"78313ad4-d7f9-4788-9293-7cc067465f5a","Type":"ContainerStarted","Data":"07c4564d5a39da04ff541dd9119a0f59280e888b148e58f8eb9bef49739c9378"} Apr 22 17:35:12.767745 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:12.767725 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" event={"ID":"f57a62ce-7e60-4525-ac30-5348379d96f7","Type":"ContainerStarted","Data":"fd1820917fdf90b4b1e65d418e55a62921ec320617d83c54eaf7849b5e2aabac"} Apr 22 17:35:12.768718 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:12.768700 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-6zmdm" event={"ID":"8dedc3a1-a7e9-4684-b53d-a1b087d23cdb","Type":"ContainerStarted","Data":"35e627c126f938daba0cc6cf7571d4d18dfae5a5733bdb64062617d4572b52fc"} Apr 22 17:35:12.769741 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:12.769720 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-nwhzz" event={"ID":"29e69548-2984-41bc-b4bd-8554ba75d702","Type":"ContainerStarted","Data":"53ac1e1fca1fbd8596c1dbd9d06a7ddf4831951d35d5aea676b2865422bf45c6"} Apr 22 17:35:12.771182 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:12.771138 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 17:35:12.771424 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:12.771407 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" event={"ID":"24c19adf-c002-4c22-9098-f1f3be79bb45","Type":"ContainerStarted","Data":"96d90d56fea00b8d2e3504befd041470474bfe79206128a339852648759178c8"} Apr 22 17:35:12.771481 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:12.771429 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" event={"ID":"24c19adf-c002-4c22-9098-f1f3be79bb45","Type":"ContainerStarted","Data":"c2ef05cf79a2b59a4df20ae88b7e5d799a6059890b9c399be5b875b5ef303114"} Apr 22 17:35:12.787856 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:12.787812 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-node-tuning-operator/tuned-4tdc7" podStartSLOduration=3.715209003 podStartE2EDuration="20.787798386s" podCreationTimestamp="2026-04-22 17:34:52 +0000 UTC" firstStartedPulling="2026-04-22 17:34:55.201807409 +0000 UTC m=+3.094503112" lastFinishedPulling="2026-04-22 17:35:12.274396777 +0000 UTC m=+20.167092495" observedRunningTime="2026-04-22 17:35:12.78744428 +0000 UTC m=+20.680140004" watchObservedRunningTime="2026-04-22 17:35:12.787798386 +0000 UTC m=+20.680494158" Apr 22 17:35:12.842562 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:12.842510 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-nwhzz" podStartSLOduration=3.646927265 podStartE2EDuration="20.842490851s" podCreationTimestamp="2026-04-22 17:34:52 +0000 UTC" firstStartedPulling="2026-04-22 17:34:55.19980472 +0000 UTC m=+3.092500423" lastFinishedPulling="2026-04-22 17:35:12.395368298 +0000 UTC m=+20.288064009" observedRunningTime="2026-04-22 17:35:12.842113939 +0000 UTC m=+20.734809666" watchObservedRunningTime="2026-04-22 17:35:12.842490851 +0000 UTC m=+20.735186575" Apr 22 17:35:12.861050 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:12.861003 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-4m4hg" podStartSLOduration=3.793200118 podStartE2EDuration="20.860988537s" podCreationTimestamp="2026-04-22 17:34:52 +0000 UTC" firstStartedPulling="2026-04-22 17:34:55.198529556 +0000 UTC m=+3.091225257" lastFinishedPulling="2026-04-22 17:35:12.266317975 +0000 UTC m=+20.159013676" observedRunningTime="2026-04-22 17:35:12.860842368 +0000 UTC m=+20.753538091" watchObservedRunningTime="2026-04-22 17:35:12.860988537 +0000 UTC m=+20.753684261" Apr 22 17:35:12.883108 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:12.883067 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/konnectivity-agent-vtsck" podStartSLOduration=8.278730645 podStartE2EDuration="20.883053235s" podCreationTimestamp="2026-04-22 17:34:52 +0000 UTC" firstStartedPulling="2026-04-22 17:34:55.198849725 +0000 UTC m=+3.091545427" lastFinishedPulling="2026-04-22 17:35:07.803172313 +0000 UTC m=+15.695868017" observedRunningTime="2026-04-22 17:35:12.882168795 +0000 UTC m=+20.774864518" watchObservedRunningTime="2026-04-22 17:35:12.883053235 +0000 UTC m=+20.775748958" Apr 22 17:35:13.100956 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:13.100725 2581 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="kube-system/konnectivity-agent-vtsck" Apr 22 17:35:13.101673 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:13.101646 2581 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="kube-system/konnectivity-agent-vtsck" Apr 22 17:35:13.119409 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:13.119349 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-6zmdm" podStartSLOduration=4.054195757 podStartE2EDuration="21.119329832s" podCreationTimestamp="2026-04-22 17:34:52 +0000 UTC" firstStartedPulling="2026-04-22 17:34:55.201090599 +0000 UTC m=+3.093786303" lastFinishedPulling="2026-04-22 17:35:12.266224675 +0000 UTC m=+20.158920378" observedRunningTime="2026-04-22 17:35:12.901490871 +0000 UTC m=+20.794186596" watchObservedRunningTime="2026-04-22 17:35:13.119329832 +0000 UTC m=+21.012025557" Apr 22 17:35:13.634692 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:13.634660 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:35:13.634864 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:13.634780 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zhl6m" podUID="cff34b01-53b3-4374-b73d-d3146e3c558d" Apr 22 17:35:13.774101 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:13.774065 2581 generic.go:358] "Generic (PLEG): container finished" podID="78313ad4-d7f9-4788-9293-7cc067465f5a" containerID="07c4564d5a39da04ff541dd9119a0f59280e888b148e58f8eb9bef49739c9378" exitCode=0 Apr 22 17:35:13.774799 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:13.774136 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pnh8b" event={"ID":"78313ad4-d7f9-4788-9293-7cc067465f5a","Type":"ContainerDied","Data":"07c4564d5a39da04ff541dd9119a0f59280e888b148e58f8eb9bef49739c9378"} Apr 22 17:35:13.775331 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:13.775309 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-kctrb" event={"ID":"fde77089-965b-4384-b901-80ddae4dbc8b","Type":"ContainerStarted","Data":"71868c71e352e038988fbf162f0055a6f903ad15bc2c803b3fed940a0af1ef3c"} Apr 22 17:35:13.777679 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:13.777662 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 17:35:13.777996 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:13.777974 2581 generic.go:358] "Generic (PLEG): container finished" podID="24c19adf-c002-4c22-9098-f1f3be79bb45" containerID="96d90d56fea00b8d2e3504befd041470474bfe79206128a339852648759178c8" exitCode=1 Apr 22 17:35:13.778092 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:13.778075 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" event={"ID":"24c19adf-c002-4c22-9098-f1f3be79bb45","Type":"ContainerDied","Data":"96d90d56fea00b8d2e3504befd041470474bfe79206128a339852648759178c8"} Apr 22 17:35:13.778146 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:13.778104 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" event={"ID":"24c19adf-c002-4c22-9098-f1f3be79bb45","Type":"ContainerStarted","Data":"2550d6b97adb2eadc80d2e5658a4518b5f995fb6320720b11ee4de706cd8dc05"} Apr 22 17:35:13.778146 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:13.778118 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" event={"ID":"24c19adf-c002-4c22-9098-f1f3be79bb45","Type":"ContainerStarted","Data":"3871f576456476e3eb1003b6e838a24bf2b742354e4bfd41dab3a3dabfe0ee8a"} Apr 22 17:35:13.778146 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:13.778129 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" event={"ID":"24c19adf-c002-4c22-9098-f1f3be79bb45","Type":"ContainerStarted","Data":"b1e72a99d60c4267b7e76e9d1ab3f24bc422f94f9e9a86bf73841631dce7dfc8"} Apr 22 17:35:13.778146 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:13.778142 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" event={"ID":"24c19adf-c002-4c22-9098-f1f3be79bb45","Type":"ContainerStarted","Data":"1c69bce68ce849f9a3986d4623a47c4887bfcb6606a2ac1cfc5993d9bd5b478b"} Apr 22 17:35:13.778694 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:13.778676 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kube-system/konnectivity-agent-vtsck" Apr 22 17:35:13.779062 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:13.779047 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kube-system/konnectivity-agent-vtsck" Apr 22 17:35:13.842723 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:13.842673 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-operator/iptables-alerter-kctrb" podStartSLOduration=4.775552023 podStartE2EDuration="21.842659061s" podCreationTimestamp="2026-04-22 17:34:52 +0000 UTC" firstStartedPulling="2026-04-22 17:34:55.199114817 +0000 UTC m=+3.091810519" lastFinishedPulling="2026-04-22 17:35:12.266221849 +0000 UTC m=+20.158917557" observedRunningTime="2026-04-22 17:35:13.842518805 +0000 UTC m=+21.735214526" watchObservedRunningTime="2026-04-22 17:35:13.842659061 +0000 UTC m=+21.735354767" Apr 22 17:35:14.025150 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:14.025115 2581 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/ebs.csi.aws.com-reg.sock" Apr 22 17:35:14.593986 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:14.593843 2581 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/ebs.csi.aws.com-reg.sock","Timestamp":"2026-04-22T17:35:14.025138209Z","UUID":"5ebb481a-3534-4288-985a-22215e0ea43c","Handler":null,"Name":"","Endpoint":""} Apr 22 17:35:14.595775 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:14.595751 2581 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: ebs.csi.aws.com endpoint: /var/lib/kubelet/plugins/ebs.csi.aws.com/csi.sock versions: 1.0.0 Apr 22 17:35:14.595909 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:14.595781 2581 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: ebs.csi.aws.com at endpoint: /var/lib/kubelet/plugins/ebs.csi.aws.com/csi.sock Apr 22 17:35:14.634515 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:14.634492 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:35:14.634694 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:14.634595 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-p72t5" podUID="c858366e-d830-4c42-b18e-1eab94ea3113" Apr 22 17:35:14.781698 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:14.781662 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" event={"ID":"f57a62ce-7e60-4525-ac30-5348379d96f7","Type":"ContainerStarted","Data":"c9fd07d81f3d6a944b0684d986fd659c54246a0ef510f57df1b9d43cda81687a"} Apr 22 17:35:15.634720 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:15.634536 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:35:15.634907 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:15.634799 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zhl6m" podUID="cff34b01-53b3-4374-b73d-d3146e3c558d" Apr 22 17:35:15.786654 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:15.786629 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 17:35:15.787151 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:15.787016 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" event={"ID":"24c19adf-c002-4c22-9098-f1f3be79bb45","Type":"ContainerStarted","Data":"0493cc53a0fdfe13516aa850bc78154a0b8220c2066e2601151d76399abe4bb6"} Apr 22 17:35:15.788975 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:15.788947 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" event={"ID":"f57a62ce-7e60-4525-ac30-5348379d96f7","Type":"ContainerStarted","Data":"59b03bad5eca1412d391fe349179e6a206cca0d0edd47283fa310862ceb02f8d"} Apr 22 17:35:15.811378 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:15.811328 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-cvp79" podStartSLOduration=3.854397252 podStartE2EDuration="23.811315565s" podCreationTimestamp="2026-04-22 17:34:52 +0000 UTC" firstStartedPulling="2026-04-22 17:34:55.193178642 +0000 UTC m=+3.085874345" lastFinishedPulling="2026-04-22 17:35:15.150096833 +0000 UTC m=+23.042792658" observedRunningTime="2026-04-22 17:35:15.811063657 +0000 UTC m=+23.703759381" watchObservedRunningTime="2026-04-22 17:35:15.811315565 +0000 UTC m=+23.704011318" Apr 22 17:35:16.634713 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:16.634677 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:35:16.634898 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:16.634814 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-p72t5" podUID="c858366e-d830-4c42-b18e-1eab94ea3113" Apr 22 17:35:17.634990 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:17.634962 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:35:17.635629 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:17.635105 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zhl6m" podUID="cff34b01-53b3-4374-b73d-d3146e3c558d" Apr 22 17:35:18.635697 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:18.635510 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:35:18.636292 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:18.635773 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-p72t5" podUID="c858366e-d830-4c42-b18e-1eab94ea3113" Apr 22 17:35:18.796220 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:18.796187 2581 generic.go:358] "Generic (PLEG): container finished" podID="78313ad4-d7f9-4788-9293-7cc067465f5a" containerID="ac12addbffe2f676344d13d0d82b1871bd4ed8f5f4ca2abfa56ffde5828046c7" exitCode=0 Apr 22 17:35:18.796361 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:18.796267 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pnh8b" event={"ID":"78313ad4-d7f9-4788-9293-7cc067465f5a","Type":"ContainerDied","Data":"ac12addbffe2f676344d13d0d82b1871bd4ed8f5f4ca2abfa56ffde5828046c7"} Apr 22 17:35:18.799356 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:18.799332 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 17:35:18.799633 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:18.799613 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" event={"ID":"24c19adf-c002-4c22-9098-f1f3be79bb45","Type":"ContainerStarted","Data":"bf26a3cae8e3f2b6cc470ec907dbfe22e839ffb57897aa64dc5b0e26255b58e1"} Apr 22 17:35:18.799990 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:18.799974 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:35:18.800042 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:18.800035 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:35:18.800090 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:18.800046 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:35:18.800179 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:18.800118 2581 scope.go:117] "RemoveContainer" containerID="96d90d56fea00b8d2e3504befd041470474bfe79206128a339852648759178c8" Apr 22 17:35:18.815643 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:18.815617 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:35:18.816110 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:18.816094 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:35:19.635169 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:19.635136 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:35:19.635348 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:19.635246 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zhl6m" podUID="cff34b01-53b3-4374-b73d-d3146e3c558d" Apr 22 17:35:19.806088 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:19.806005 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 17:35:19.806495 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:19.806314 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" event={"ID":"24c19adf-c002-4c22-9098-f1f3be79bb45","Type":"ContainerStarted","Data":"946bc12f53b1c75af79c77870c87531f452ac530b92241cfd6999792abd78f9d"} Apr 22 17:35:19.838360 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:19.838310 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" podStartSLOduration=10.557622991 podStartE2EDuration="27.838297115s" podCreationTimestamp="2026-04-22 17:34:52 +0000 UTC" firstStartedPulling="2026-04-22 17:34:55.190365131 +0000 UTC m=+3.083060833" lastFinishedPulling="2026-04-22 17:35:12.471039242 +0000 UTC m=+20.363734957" observedRunningTime="2026-04-22 17:35:19.838010612 +0000 UTC m=+27.730706337" watchObservedRunningTime="2026-04-22 17:35:19.838297115 +0000 UTC m=+27.730992883" Apr 22 17:35:19.996899 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:19.996689 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-zhl6m"] Apr 22 17:35:19.997075 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:19.997016 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:35:19.997124 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:19.997109 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zhl6m" podUID="cff34b01-53b3-4374-b73d-d3146e3c558d" Apr 22 17:35:19.999800 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:19.999773 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-target-p72t5"] Apr 22 17:35:19.999934 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:19.999906 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:35:20.000016 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:19.999995 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-p72t5" podUID="c858366e-d830-4c42-b18e-1eab94ea3113" Apr 22 17:35:20.811943 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:20.811904 2581 generic.go:358] "Generic (PLEG): container finished" podID="78313ad4-d7f9-4788-9293-7cc067465f5a" containerID="e114c54f9979f3e7002d1679ae0365d996555852c7a613576936197d5bfdc74f" exitCode=0 Apr 22 17:35:20.812333 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:20.811983 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pnh8b" event={"ID":"78313ad4-d7f9-4788-9293-7cc067465f5a","Type":"ContainerDied","Data":"e114c54f9979f3e7002d1679ae0365d996555852c7a613576936197d5bfdc74f"} Apr 22 17:35:21.634754 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:21.634721 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:35:21.634905 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:21.634721 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:35:21.634905 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:21.634845 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zhl6m" podUID="cff34b01-53b3-4374-b73d-d3146e3c558d" Apr 22 17:35:21.634971 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:21.634925 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-p72t5" podUID="c858366e-d830-4c42-b18e-1eab94ea3113" Apr 22 17:35:22.817052 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:22.817016 2581 generic.go:358] "Generic (PLEG): container finished" podID="78313ad4-d7f9-4788-9293-7cc067465f5a" containerID="4b8092eb60df4fa3e163ecc736f800847b6cda463260285afdf0033dd07e5109" exitCode=0 Apr 22 17:35:22.817629 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:22.817085 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pnh8b" event={"ID":"78313ad4-d7f9-4788-9293-7cc067465f5a","Type":"ContainerDied","Data":"4b8092eb60df4fa3e163ecc736f800847b6cda463260285afdf0033dd07e5109"} Apr 22 17:35:23.635128 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:23.635101 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:35:23.635277 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:23.635103 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:35:23.635277 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:23.635213 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-p72t5" podUID="c858366e-d830-4c42-b18e-1eab94ea3113" Apr 22 17:35:23.635392 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:23.635325 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-zhl6m" podUID="cff34b01-53b3-4374-b73d-d3146e3c558d" Apr 22 17:35:25.431484 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.431409 2581 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-138-57.ec2.internal" event="NodeReady" Apr 22 17:35:25.431972 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.431580 2581 kubelet_node_status.go:550] "Fast updating node status as it just became ready" Apr 22 17:35:25.478968 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.478934 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-nqmll"] Apr 22 17:35:25.495994 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.495960 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-jxlcn"] Apr 22 17:35:25.496192 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.496144 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-nqmll" Apr 22 17:35:25.498578 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.498550 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-default-metrics-tls\"" Apr 22 17:35:25.498712 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.498585 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-dockercfg-m6rcd\"" Apr 22 17:35:25.499082 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.499064 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"dns-default\"" Apr 22 17:35:25.513671 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.513642 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-nqmll"] Apr 22 17:35:25.513671 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.513674 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-jxlcn"] Apr 22 17:35:25.513909 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.513804 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-jxlcn" Apr 22 17:35:25.516110 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.516088 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"default-dockercfg-dt854\"" Apr 22 17:35:25.516245 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.516089 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"canary-serving-cert\"" Apr 22 17:35:25.516245 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.516119 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"kube-root-ca.crt\"" Apr 22 17:35:25.516245 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.516093 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"openshift-service-ca.crt\"" Apr 22 17:35:25.577952 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.577917 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d6515c87-3bb8-42d3-be06-5bf26230f609-config-volume\") pod \"dns-default-nqmll\" (UID: \"d6515c87-3bb8-42d3-be06-5bf26230f609\") " pod="openshift-dns/dns-default-nqmll" Apr 22 17:35:25.577952 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.577964 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/d6515c87-3bb8-42d3-be06-5bf26230f609-tmp-dir\") pod \"dns-default-nqmll\" (UID: \"d6515c87-3bb8-42d3-be06-5bf26230f609\") " pod="openshift-dns/dns-default-nqmll" Apr 22 17:35:25.578196 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.578038 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d6515c87-3bb8-42d3-be06-5bf26230f609-metrics-tls\") pod \"dns-default-nqmll\" (UID: \"d6515c87-3bb8-42d3-be06-5bf26230f609\") " pod="openshift-dns/dns-default-nqmll" Apr 22 17:35:25.578196 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.578082 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7d7m\" (UniqueName: \"kubernetes.io/projected/d6515c87-3bb8-42d3-be06-5bf26230f609-kube-api-access-l7d7m\") pod \"dns-default-nqmll\" (UID: \"d6515c87-3bb8-42d3-be06-5bf26230f609\") " pod="openshift-dns/dns-default-nqmll" Apr 22 17:35:25.635282 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.635252 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:35:25.635481 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.635458 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:35:25.637959 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.637928 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Apr 22 17:35:25.638086 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.637990 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Apr 22 17:35:25.638145 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.638138 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-qs9jr\"" Apr 22 17:35:25.638200 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.638155 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Apr 22 17:35:25.638256 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.638237 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-diagnostics\"/\"default-dockercfg-z4sh7\"" Apr 22 17:35:25.678607 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.678576 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d6515c87-3bb8-42d3-be06-5bf26230f609-config-volume\") pod \"dns-default-nqmll\" (UID: \"d6515c87-3bb8-42d3-be06-5bf26230f609\") " pod="openshift-dns/dns-default-nqmll" Apr 22 17:35:25.678607 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.678615 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/d6515c87-3bb8-42d3-be06-5bf26230f609-tmp-dir\") pod \"dns-default-nqmll\" (UID: \"d6515c87-3bb8-42d3-be06-5bf26230f609\") " pod="openshift-dns/dns-default-nqmll" Apr 22 17:35:25.678864 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.678647 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkrss\" (UniqueName: \"kubernetes.io/projected/58576398-c4a1-4369-8c0c-3b936c2c0752-kube-api-access-kkrss\") pod \"ingress-canary-jxlcn\" (UID: \"58576398-c4a1-4369-8c0c-3b936c2c0752\") " pod="openshift-ingress-canary/ingress-canary-jxlcn" Apr 22 17:35:25.678864 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.678699 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/58576398-c4a1-4369-8c0c-3b936c2c0752-cert\") pod \"ingress-canary-jxlcn\" (UID: \"58576398-c4a1-4369-8c0c-3b936c2c0752\") " pod="openshift-ingress-canary/ingress-canary-jxlcn" Apr 22 17:35:25.678864 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.678725 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d6515c87-3bb8-42d3-be06-5bf26230f609-metrics-tls\") pod \"dns-default-nqmll\" (UID: \"d6515c87-3bb8-42d3-be06-5bf26230f609\") " pod="openshift-dns/dns-default-nqmll" Apr 22 17:35:25.678864 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.678764 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l7d7m\" (UniqueName: \"kubernetes.io/projected/d6515c87-3bb8-42d3-be06-5bf26230f609-kube-api-access-l7d7m\") pod \"dns-default-nqmll\" (UID: \"d6515c87-3bb8-42d3-be06-5bf26230f609\") " pod="openshift-dns/dns-default-nqmll" Apr 22 17:35:25.678864 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:25.678849 2581 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 22 17:35:25.679148 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:25.678942 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d6515c87-3bb8-42d3-be06-5bf26230f609-metrics-tls podName:d6515c87-3bb8-42d3-be06-5bf26230f609 nodeName:}" failed. No retries permitted until 2026-04-22 17:35:26.178918761 +0000 UTC m=+34.071614477 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/d6515c87-3bb8-42d3-be06-5bf26230f609-metrics-tls") pod "dns-default-nqmll" (UID: "d6515c87-3bb8-42d3-be06-5bf26230f609") : secret "dns-default-metrics-tls" not found Apr 22 17:35:25.679148 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.679021 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/d6515c87-3bb8-42d3-be06-5bf26230f609-tmp-dir\") pod \"dns-default-nqmll\" (UID: \"d6515c87-3bb8-42d3-be06-5bf26230f609\") " pod="openshift-dns/dns-default-nqmll" Apr 22 17:35:25.679284 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.679268 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d6515c87-3bb8-42d3-be06-5bf26230f609-config-volume\") pod \"dns-default-nqmll\" (UID: \"d6515c87-3bb8-42d3-be06-5bf26230f609\") " pod="openshift-dns/dns-default-nqmll" Apr 22 17:35:25.689567 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.689506 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7d7m\" (UniqueName: \"kubernetes.io/projected/d6515c87-3bb8-42d3-be06-5bf26230f609-kube-api-access-l7d7m\") pod \"dns-default-nqmll\" (UID: \"d6515c87-3bb8-42d3-be06-5bf26230f609\") " pod="openshift-dns/dns-default-nqmll" Apr 22 17:35:25.779325 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.779286 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-kkrss\" (UniqueName: \"kubernetes.io/projected/58576398-c4a1-4369-8c0c-3b936c2c0752-kube-api-access-kkrss\") pod \"ingress-canary-jxlcn\" (UID: \"58576398-c4a1-4369-8c0c-3b936c2c0752\") " pod="openshift-ingress-canary/ingress-canary-jxlcn" Apr 22 17:35:25.779551 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.779528 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/58576398-c4a1-4369-8c0c-3b936c2c0752-cert\") pod \"ingress-canary-jxlcn\" (UID: \"58576398-c4a1-4369-8c0c-3b936c2c0752\") " pod="openshift-ingress-canary/ingress-canary-jxlcn" Apr 22 17:35:25.779700 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:25.779676 2581 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 22 17:35:25.779810 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:25.779743 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58576398-c4a1-4369-8c0c-3b936c2c0752-cert podName:58576398-c4a1-4369-8c0c-3b936c2c0752 nodeName:}" failed. No retries permitted until 2026-04-22 17:35:26.279725282 +0000 UTC m=+34.172420990 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/58576398-c4a1-4369-8c0c-3b936c2c0752-cert") pod "ingress-canary-jxlcn" (UID: "58576398-c4a1-4369-8c0c-3b936c2c0752") : secret "canary-serving-cert" not found Apr 22 17:35:25.790428 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:25.790400 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkrss\" (UniqueName: \"kubernetes.io/projected/58576398-c4a1-4369-8c0c-3b936c2c0752-kube-api-access-kkrss\") pod \"ingress-canary-jxlcn\" (UID: \"58576398-c4a1-4369-8c0c-3b936c2c0752\") " pod="openshift-ingress-canary/ingress-canary-jxlcn" Apr 22 17:35:26.182782 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:26.182744 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d6515c87-3bb8-42d3-be06-5bf26230f609-metrics-tls\") pod \"dns-default-nqmll\" (UID: \"d6515c87-3bb8-42d3-be06-5bf26230f609\") " pod="openshift-dns/dns-default-nqmll" Apr 22 17:35:26.183057 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:26.182919 2581 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 22 17:35:26.183057 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:26.182983 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d6515c87-3bb8-42d3-be06-5bf26230f609-metrics-tls podName:d6515c87-3bb8-42d3-be06-5bf26230f609 nodeName:}" failed. No retries permitted until 2026-04-22 17:35:27.182964288 +0000 UTC m=+35.075659991 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/d6515c87-3bb8-42d3-be06-5bf26230f609-metrics-tls") pod "dns-default-nqmll" (UID: "d6515c87-3bb8-42d3-be06-5bf26230f609") : secret "dns-default-metrics-tls" not found Apr 22 17:35:26.283939 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:26.283905 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-4gtkc\" (UniqueName: \"kubernetes.io/projected/c858366e-d830-4c42-b18e-1eab94ea3113-kube-api-access-4gtkc\") pod \"network-check-target-p72t5\" (UID: \"c858366e-d830-4c42-b18e-1eab94ea3113\") " pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:35:26.284113 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:26.283955 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/58576398-c4a1-4369-8c0c-3b936c2c0752-cert\") pod \"ingress-canary-jxlcn\" (UID: \"58576398-c4a1-4369-8c0c-3b936c2c0752\") " pod="openshift-ingress-canary/ingress-canary-jxlcn" Apr 22 17:35:26.284113 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:26.283984 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs\") pod \"network-metrics-daemon-zhl6m\" (UID: \"cff34b01-53b3-4374-b73d-d3146e3c558d\") " pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:35:26.284113 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:26.284080 2581 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: secret "metrics-daemon-secret" not found Apr 22 17:35:26.284228 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:26.284107 2581 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 22 17:35:26.284228 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:26.284145 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs podName:cff34b01-53b3-4374-b73d-d3146e3c558d nodeName:}" failed. No retries permitted until 2026-04-22 17:35:58.284129841 +0000 UTC m=+66.176825557 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs") pod "network-metrics-daemon-zhl6m" (UID: "cff34b01-53b3-4374-b73d-d3146e3c558d") : secret "metrics-daemon-secret" not found Apr 22 17:35:26.284228 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:26.284170 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58576398-c4a1-4369-8c0c-3b936c2c0752-cert podName:58576398-c4a1-4369-8c0c-3b936c2c0752 nodeName:}" failed. No retries permitted until 2026-04-22 17:35:27.284150371 +0000 UTC m=+35.176846072 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/58576398-c4a1-4369-8c0c-3b936c2c0752-cert") pod "ingress-canary-jxlcn" (UID: "58576398-c4a1-4369-8c0c-3b936c2c0752") : secret "canary-serving-cert" not found Apr 22 17:35:26.286483 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:26.286456 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gtkc\" (UniqueName: \"kubernetes.io/projected/c858366e-d830-4c42-b18e-1eab94ea3113-kube-api-access-4gtkc\") pod \"network-check-target-p72t5\" (UID: \"c858366e-d830-4c42-b18e-1eab94ea3113\") " pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:35:26.546624 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:26.546584 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:35:27.191493 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:27.191454 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d6515c87-3bb8-42d3-be06-5bf26230f609-metrics-tls\") pod \"dns-default-nqmll\" (UID: \"d6515c87-3bb8-42d3-be06-5bf26230f609\") " pod="openshift-dns/dns-default-nqmll" Apr 22 17:35:27.191685 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:27.191605 2581 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 22 17:35:27.191754 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:27.191696 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d6515c87-3bb8-42d3-be06-5bf26230f609-metrics-tls podName:d6515c87-3bb8-42d3-be06-5bf26230f609 nodeName:}" failed. No retries permitted until 2026-04-22 17:35:29.191674602 +0000 UTC m=+37.084370309 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/d6515c87-3bb8-42d3-be06-5bf26230f609-metrics-tls") pod "dns-default-nqmll" (UID: "d6515c87-3bb8-42d3-be06-5bf26230f609") : secret "dns-default-metrics-tls" not found Apr 22 17:35:27.292714 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:27.292669 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/58576398-c4a1-4369-8c0c-3b936c2c0752-cert\") pod \"ingress-canary-jxlcn\" (UID: \"58576398-c4a1-4369-8c0c-3b936c2c0752\") " pod="openshift-ingress-canary/ingress-canary-jxlcn" Apr 22 17:35:27.292863 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:27.292826 2581 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 22 17:35:27.292939 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:27.292920 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58576398-c4a1-4369-8c0c-3b936c2c0752-cert podName:58576398-c4a1-4369-8c0c-3b936c2c0752 nodeName:}" failed. No retries permitted until 2026-04-22 17:35:29.292898609 +0000 UTC m=+37.185594315 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/58576398-c4a1-4369-8c0c-3b936c2c0752-cert") pod "ingress-canary-jxlcn" (UID: "58576398-c4a1-4369-8c0c-3b936c2c0752") : secret "canary-serving-cert" not found Apr 22 17:35:28.620579 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:28.620386 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-target-p72t5"] Apr 22 17:35:28.754821 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:35:28.754749 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc858366e_d830_4c42_b18e_1eab94ea3113.slice/crio-6772f02781a2f8859502afc4d23f392e4ded5307a8350a771b0ea2a7d4e7bc33 WatchSource:0}: Error finding container 6772f02781a2f8859502afc4d23f392e4ded5307a8350a771b0ea2a7d4e7bc33: Status 404 returned error can't find the container with id 6772f02781a2f8859502afc4d23f392e4ded5307a8350a771b0ea2a7d4e7bc33 Apr 22 17:35:28.830183 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:28.830145 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-p72t5" event={"ID":"c858366e-d830-4c42-b18e-1eab94ea3113","Type":"ContainerStarted","Data":"6772f02781a2f8859502afc4d23f392e4ded5307a8350a771b0ea2a7d4e7bc33"} Apr 22 17:35:29.206773 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:29.206732 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d6515c87-3bb8-42d3-be06-5bf26230f609-metrics-tls\") pod \"dns-default-nqmll\" (UID: \"d6515c87-3bb8-42d3-be06-5bf26230f609\") " pod="openshift-dns/dns-default-nqmll" Apr 22 17:35:29.206951 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:29.206914 2581 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 22 17:35:29.207016 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:29.206988 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d6515c87-3bb8-42d3-be06-5bf26230f609-metrics-tls podName:d6515c87-3bb8-42d3-be06-5bf26230f609 nodeName:}" failed. No retries permitted until 2026-04-22 17:35:33.20696857 +0000 UTC m=+41.099664462 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/d6515c87-3bb8-42d3-be06-5bf26230f609-metrics-tls") pod "dns-default-nqmll" (UID: "d6515c87-3bb8-42d3-be06-5bf26230f609") : secret "dns-default-metrics-tls" not found Apr 22 17:35:29.307470 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:29.307379 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/58576398-c4a1-4369-8c0c-3b936c2c0752-cert\") pod \"ingress-canary-jxlcn\" (UID: \"58576398-c4a1-4369-8c0c-3b936c2c0752\") " pod="openshift-ingress-canary/ingress-canary-jxlcn" Apr 22 17:35:29.307618 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:29.307547 2581 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 22 17:35:29.307618 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:29.307606 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58576398-c4a1-4369-8c0c-3b936c2c0752-cert podName:58576398-c4a1-4369-8c0c-3b936c2c0752 nodeName:}" failed. No retries permitted until 2026-04-22 17:35:33.307588833 +0000 UTC m=+41.200284543 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/58576398-c4a1-4369-8c0c-3b936c2c0752-cert") pod "ingress-canary-jxlcn" (UID: "58576398-c4a1-4369-8c0c-3b936c2c0752") : secret "canary-serving-cert" not found Apr 22 17:35:29.835595 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:29.835552 2581 generic.go:358] "Generic (PLEG): container finished" podID="78313ad4-d7f9-4788-9293-7cc067465f5a" containerID="0a2cfa0c385cffe687738ec547077df5c63f9028cccdedfb55fd2f6991508645" exitCode=0 Apr 22 17:35:29.836079 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:29.835619 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pnh8b" event={"ID":"78313ad4-d7f9-4788-9293-7cc067465f5a","Type":"ContainerDied","Data":"0a2cfa0c385cffe687738ec547077df5c63f9028cccdedfb55fd2f6991508645"} Apr 22 17:35:30.840689 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:30.840485 2581 generic.go:358] "Generic (PLEG): container finished" podID="78313ad4-d7f9-4788-9293-7cc067465f5a" containerID="6015c717b713be25f8b7579d394929f6e057fd85cd5e2899203a7d72d2aa61da" exitCode=0 Apr 22 17:35:30.841092 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:30.840587 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pnh8b" event={"ID":"78313ad4-d7f9-4788-9293-7cc067465f5a","Type":"ContainerDied","Data":"6015c717b713be25f8b7579d394929f6e057fd85cd5e2899203a7d72d2aa61da"} Apr 22 17:35:31.845523 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:31.845426 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-pnh8b" event={"ID":"78313ad4-d7f9-4788-9293-7cc067465f5a","Type":"ContainerStarted","Data":"37bac3c889aaf1d2a0605df7eb256ebab0d57b728b35a40507ee35c8d16d99d1"} Apr 22 17:35:31.846692 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:31.846669 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-p72t5" event={"ID":"c858366e-d830-4c42-b18e-1eab94ea3113","Type":"ContainerStarted","Data":"456fa4f7cd8a16d55e4be5472411d5c1d375decc0b21eec0cc2b22c29acbbe3f"} Apr 22 17:35:31.846797 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:31.846723 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:35:31.872152 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:31.872102 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-pnh8b" podStartSLOduration=6.281245803 podStartE2EDuration="39.872084077s" podCreationTimestamp="2026-04-22 17:34:52 +0000 UTC" firstStartedPulling="2026-04-22 17:34:55.194967435 +0000 UTC m=+3.087663137" lastFinishedPulling="2026-04-22 17:35:28.785805709 +0000 UTC m=+36.678501411" observedRunningTime="2026-04-22 17:35:31.8711156 +0000 UTC m=+39.763811326" watchObservedRunningTime="2026-04-22 17:35:31.872084077 +0000 UTC m=+39.764779803" Apr 22 17:35:31.890162 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:31.890110 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-diagnostics/network-check-target-p72t5" podStartSLOduration=37.046462361 podStartE2EDuration="39.890090938s" podCreationTimestamp="2026-04-22 17:34:52 +0000 UTC" firstStartedPulling="2026-04-22 17:35:28.762798449 +0000 UTC m=+36.655494151" lastFinishedPulling="2026-04-22 17:35:31.606427007 +0000 UTC m=+39.499122728" observedRunningTime="2026-04-22 17:35:31.889784359 +0000 UTC m=+39.782480083" watchObservedRunningTime="2026-04-22 17:35:31.890090938 +0000 UTC m=+39.782786665" Apr 22 17:35:33.236865 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:33.236825 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d6515c87-3bb8-42d3-be06-5bf26230f609-metrics-tls\") pod \"dns-default-nqmll\" (UID: \"d6515c87-3bb8-42d3-be06-5bf26230f609\") " pod="openshift-dns/dns-default-nqmll" Apr 22 17:35:33.237259 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:33.236988 2581 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 22 17:35:33.237259 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:33.237052 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d6515c87-3bb8-42d3-be06-5bf26230f609-metrics-tls podName:d6515c87-3bb8-42d3-be06-5bf26230f609 nodeName:}" failed. No retries permitted until 2026-04-22 17:35:41.237036302 +0000 UTC m=+49.129732003 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/d6515c87-3bb8-42d3-be06-5bf26230f609-metrics-tls") pod "dns-default-nqmll" (UID: "d6515c87-3bb8-42d3-be06-5bf26230f609") : secret "dns-default-metrics-tls" not found Apr 22 17:35:33.337411 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:33.337371 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/58576398-c4a1-4369-8c0c-3b936c2c0752-cert\") pod \"ingress-canary-jxlcn\" (UID: \"58576398-c4a1-4369-8c0c-3b936c2c0752\") " pod="openshift-ingress-canary/ingress-canary-jxlcn" Apr 22 17:35:33.337563 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:33.337516 2581 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 22 17:35:33.337608 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:33.337579 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58576398-c4a1-4369-8c0c-3b936c2c0752-cert podName:58576398-c4a1-4369-8c0c-3b936c2c0752 nodeName:}" failed. No retries permitted until 2026-04-22 17:35:41.337563805 +0000 UTC m=+49.230259508 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/58576398-c4a1-4369-8c0c-3b936c2c0752-cert") pod "ingress-canary-jxlcn" (UID: "58576398-c4a1-4369-8c0c-3b936c2c0752") : secret "canary-serving-cert" not found Apr 22 17:35:41.072374 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.072339 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp"] Apr 22 17:35:41.113671 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.113638 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp"] Apr 22 17:35:41.113827 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.113761 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" Apr 22 17:35:41.116201 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.116158 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"open-cluster-management-agent-addon\"/\"cluster-proxy-ca\"" Apr 22 17:35:41.116201 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.116179 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"open-cluster-management-agent-addon\"/\"cluster-proxy-open-cluster-management.io-proxy-agent-signer-client-cert\"" Apr 22 17:35:41.116201 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.116159 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"open-cluster-management-agent-addon\"/\"kube-root-ca.crt\"" Apr 22 17:35:41.116389 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.116159 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"open-cluster-management-agent-addon\"/\"openshift-service-ca.crt\"" Apr 22 17:35:41.116842 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.116824 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"open-cluster-management-agent-addon\"/\"cluster-proxy-service-proxy-server-certificates\"" Apr 22 17:35:41.116939 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.116839 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"open-cluster-management-agent-addon\"/\"cluster-proxy-hub-kubeconfig\"" Apr 22 17:35:41.116939 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.116843 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"open-cluster-management-agent-addon\"/\"open-cluster-management-image-pull-credentials\"" Apr 22 17:35:41.287081 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.287048 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hub-kubeconfig\" (UniqueName: \"kubernetes.io/secret/0c337636-6259-4b1c-a1b7-e948b21c03d6-hub-kubeconfig\") pod \"cluster-proxy-proxy-agent-5ccb5fcc49-d96zp\" (UID: \"0c337636-6259-4b1c-a1b7-e948b21c03d6\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" Apr 22 17:35:41.287081 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.287097 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-proxy-server-cert\" (UniqueName: \"kubernetes.io/secret/0c337636-6259-4b1c-a1b7-e948b21c03d6-service-proxy-server-cert\") pod \"cluster-proxy-proxy-agent-5ccb5fcc49-d96zp\" (UID: \"0c337636-6259-4b1c-a1b7-e948b21c03d6\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" Apr 22 17:35:41.287329 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.287165 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d6515c87-3bb8-42d3-be06-5bf26230f609-metrics-tls\") pod \"dns-default-nqmll\" (UID: \"d6515c87-3bb8-42d3-be06-5bf26230f609\") " pod="openshift-dns/dns-default-nqmll" Apr 22 17:35:41.287329 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.287208 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hub\" (UniqueName: \"kubernetes.io/secret/0c337636-6259-4b1c-a1b7-e948b21c03d6-hub\") pod \"cluster-proxy-proxy-agent-5ccb5fcc49-d96zp\" (UID: \"0c337636-6259-4b1c-a1b7-e948b21c03d6\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" Apr 22 17:35:41.287329 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.287235 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca\" (UniqueName: \"kubernetes.io/secret/0c337636-6259-4b1c-a1b7-e948b21c03d6-ca\") pod \"cluster-proxy-proxy-agent-5ccb5fcc49-d96zp\" (UID: \"0c337636-6259-4b1c-a1b7-e948b21c03d6\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" Apr 22 17:35:41.287329 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.287264 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ocpservice-ca\" (UniqueName: \"kubernetes.io/configmap/0c337636-6259-4b1c-a1b7-e948b21c03d6-ocpservice-ca\") pod \"cluster-proxy-proxy-agent-5ccb5fcc49-d96zp\" (UID: \"0c337636-6259-4b1c-a1b7-e948b21c03d6\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" Apr 22 17:35:41.287329 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.287286 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ffr8\" (UniqueName: \"kubernetes.io/projected/0c337636-6259-4b1c-a1b7-e948b21c03d6-kube-api-access-7ffr8\") pod \"cluster-proxy-proxy-agent-5ccb5fcc49-d96zp\" (UID: \"0c337636-6259-4b1c-a1b7-e948b21c03d6\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" Apr 22 17:35:41.287481 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:41.287349 2581 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 22 17:35:41.287481 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:41.287412 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d6515c87-3bb8-42d3-be06-5bf26230f609-metrics-tls podName:d6515c87-3bb8-42d3-be06-5bf26230f609 nodeName:}" failed. No retries permitted until 2026-04-22 17:35:57.287397642 +0000 UTC m=+65.180093343 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/d6515c87-3bb8-42d3-be06-5bf26230f609-metrics-tls") pod "dns-default-nqmll" (UID: "d6515c87-3bb8-42d3-be06-5bf26230f609") : secret "dns-default-metrics-tls" not found Apr 22 17:35:41.387992 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.387920 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ocpservice-ca\" (UniqueName: \"kubernetes.io/configmap/0c337636-6259-4b1c-a1b7-e948b21c03d6-ocpservice-ca\") pod \"cluster-proxy-proxy-agent-5ccb5fcc49-d96zp\" (UID: \"0c337636-6259-4b1c-a1b7-e948b21c03d6\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" Apr 22 17:35:41.387992 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.387958 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-7ffr8\" (UniqueName: \"kubernetes.io/projected/0c337636-6259-4b1c-a1b7-e948b21c03d6-kube-api-access-7ffr8\") pod \"cluster-proxy-proxy-agent-5ccb5fcc49-d96zp\" (UID: \"0c337636-6259-4b1c-a1b7-e948b21c03d6\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" Apr 22 17:35:41.387992 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.387978 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hub-kubeconfig\" (UniqueName: \"kubernetes.io/secret/0c337636-6259-4b1c-a1b7-e948b21c03d6-hub-kubeconfig\") pod \"cluster-proxy-proxy-agent-5ccb5fcc49-d96zp\" (UID: \"0c337636-6259-4b1c-a1b7-e948b21c03d6\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" Apr 22 17:35:41.388201 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.388113 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/58576398-c4a1-4369-8c0c-3b936c2c0752-cert\") pod \"ingress-canary-jxlcn\" (UID: \"58576398-c4a1-4369-8c0c-3b936c2c0752\") " pod="openshift-ingress-canary/ingress-canary-jxlcn" Apr 22 17:35:41.388201 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.388155 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-proxy-server-cert\" (UniqueName: \"kubernetes.io/secret/0c337636-6259-4b1c-a1b7-e948b21c03d6-service-proxy-server-cert\") pod \"cluster-proxy-proxy-agent-5ccb5fcc49-d96zp\" (UID: \"0c337636-6259-4b1c-a1b7-e948b21c03d6\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" Apr 22 17:35:41.388297 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.388231 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hub\" (UniqueName: \"kubernetes.io/secret/0c337636-6259-4b1c-a1b7-e948b21c03d6-hub\") pod \"cluster-proxy-proxy-agent-5ccb5fcc49-d96zp\" (UID: \"0c337636-6259-4b1c-a1b7-e948b21c03d6\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" Apr 22 17:35:41.388297 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:41.388265 2581 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 22 17:35:41.388297 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.388273 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca\" (UniqueName: \"kubernetes.io/secret/0c337636-6259-4b1c-a1b7-e948b21c03d6-ca\") pod \"cluster-proxy-proxy-agent-5ccb5fcc49-d96zp\" (UID: \"0c337636-6259-4b1c-a1b7-e948b21c03d6\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" Apr 22 17:35:41.388432 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:41.388349 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58576398-c4a1-4369-8c0c-3b936c2c0752-cert podName:58576398-c4a1-4369-8c0c-3b936c2c0752 nodeName:}" failed. No retries permitted until 2026-04-22 17:35:57.388324436 +0000 UTC m=+65.281020143 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/58576398-c4a1-4369-8c0c-3b936c2c0752-cert") pod "ingress-canary-jxlcn" (UID: "58576398-c4a1-4369-8c0c-3b936c2c0752") : secret "canary-serving-cert" not found Apr 22 17:35:41.388798 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.388765 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ocpservice-ca\" (UniqueName: \"kubernetes.io/configmap/0c337636-6259-4b1c-a1b7-e948b21c03d6-ocpservice-ca\") pod \"cluster-proxy-proxy-agent-5ccb5fcc49-d96zp\" (UID: \"0c337636-6259-4b1c-a1b7-e948b21c03d6\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" Apr 22 17:35:41.392164 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.392140 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca\" (UniqueName: \"kubernetes.io/secret/0c337636-6259-4b1c-a1b7-e948b21c03d6-ca\") pod \"cluster-proxy-proxy-agent-5ccb5fcc49-d96zp\" (UID: \"0c337636-6259-4b1c-a1b7-e948b21c03d6\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" Apr 22 17:35:41.392291 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.392273 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-proxy-server-cert\" (UniqueName: \"kubernetes.io/secret/0c337636-6259-4b1c-a1b7-e948b21c03d6-service-proxy-server-cert\") pod \"cluster-proxy-proxy-agent-5ccb5fcc49-d96zp\" (UID: \"0c337636-6259-4b1c-a1b7-e948b21c03d6\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" Apr 22 17:35:41.392340 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.392293 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hub\" (UniqueName: \"kubernetes.io/secret/0c337636-6259-4b1c-a1b7-e948b21c03d6-hub\") pod \"cluster-proxy-proxy-agent-5ccb5fcc49-d96zp\" (UID: \"0c337636-6259-4b1c-a1b7-e948b21c03d6\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" Apr 22 17:35:41.392340 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.392301 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hub-kubeconfig\" (UniqueName: \"kubernetes.io/secret/0c337636-6259-4b1c-a1b7-e948b21c03d6-hub-kubeconfig\") pod \"cluster-proxy-proxy-agent-5ccb5fcc49-d96zp\" (UID: \"0c337636-6259-4b1c-a1b7-e948b21c03d6\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" Apr 22 17:35:41.395513 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.395494 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ffr8\" (UniqueName: \"kubernetes.io/projected/0c337636-6259-4b1c-a1b7-e948b21c03d6-kube-api-access-7ffr8\") pod \"cluster-proxy-proxy-agent-5ccb5fcc49-d96zp\" (UID: \"0c337636-6259-4b1c-a1b7-e948b21c03d6\") " pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" Apr 22 17:35:41.431973 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.431937 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" Apr 22 17:35:41.548061 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.548029 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp"] Apr 22 17:35:41.551835 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:35:41.551802 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c337636_6259_4b1c_a1b7_e948b21c03d6.slice/crio-da699569b22d6d433abe962c5ab136c5d041ff34563ad9e23d45be1621b2e00e WatchSource:0}: Error finding container da699569b22d6d433abe962c5ab136c5d041ff34563ad9e23d45be1621b2e00e: Status 404 returned error can't find the container with id da699569b22d6d433abe962c5ab136c5d041ff34563ad9e23d45be1621b2e00e Apr 22 17:35:41.865818 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:41.865776 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" event={"ID":"0c337636-6259-4b1c-a1b7-e948b21c03d6","Type":"ContainerStarted","Data":"da699569b22d6d433abe962c5ab136c5d041ff34563ad9e23d45be1621b2e00e"} Apr 22 17:35:45.874982 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:45.874943 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" event={"ID":"0c337636-6259-4b1c-a1b7-e948b21c03d6","Type":"ContainerStarted","Data":"423bad0d072a5b4a2ec0a80165fbb497358f6c3aa5657c5797a6ca0ea5f10f14"} Apr 22 17:35:47.884125 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:47.884076 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" event={"ID":"0c337636-6259-4b1c-a1b7-e948b21c03d6","Type":"ContainerStarted","Data":"97fd3d639738d1a662180cd0f2e01591fb324a82a6c77dbd2562b47b14b9af74"} Apr 22 17:35:48.887653 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:48.887621 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" event={"ID":"0c337636-6259-4b1c-a1b7-e948b21c03d6","Type":"ContainerStarted","Data":"fd42bfaaed32d6e4b8c9b0c0c7b65fb007826baa9fb99ebf41b9b7f68e9184c7"} Apr 22 17:35:48.913591 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:48.913553 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" podStartSLOduration=1.7113440629999999 podStartE2EDuration="7.913541981s" podCreationTimestamp="2026-04-22 17:35:41 +0000 UTC" firstStartedPulling="2026-04-22 17:35:41.553614158 +0000 UTC m=+49.446309860" lastFinishedPulling="2026-04-22 17:35:47.755812074 +0000 UTC m=+55.648507778" observedRunningTime="2026-04-22 17:35:48.913079972 +0000 UTC m=+56.805775696" watchObservedRunningTime="2026-04-22 17:35:48.913541981 +0000 UTC m=+56.806237705" Apr 22 17:35:50.829054 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:50.829020 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tn68c" Apr 22 17:35:57.295652 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:57.295597 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d6515c87-3bb8-42d3-be06-5bf26230f609-metrics-tls\") pod \"dns-default-nqmll\" (UID: \"d6515c87-3bb8-42d3-be06-5bf26230f609\") " pod="openshift-dns/dns-default-nqmll" Apr 22 17:35:57.296172 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:57.295760 2581 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 22 17:35:57.296172 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:57.295827 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d6515c87-3bb8-42d3-be06-5bf26230f609-metrics-tls podName:d6515c87-3bb8-42d3-be06-5bf26230f609 nodeName:}" failed. No retries permitted until 2026-04-22 17:36:29.295812813 +0000 UTC m=+97.188508519 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/d6515c87-3bb8-42d3-be06-5bf26230f609-metrics-tls") pod "dns-default-nqmll" (UID: "d6515c87-3bb8-42d3-be06-5bf26230f609") : secret "dns-default-metrics-tls" not found Apr 22 17:35:57.396815 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:57.396783 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/58576398-c4a1-4369-8c0c-3b936c2c0752-cert\") pod \"ingress-canary-jxlcn\" (UID: \"58576398-c4a1-4369-8c0c-3b936c2c0752\") " pod="openshift-ingress-canary/ingress-canary-jxlcn" Apr 22 17:35:57.396958 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:57.396930 2581 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 22 17:35:57.397008 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:57.396983 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58576398-c4a1-4369-8c0c-3b936c2c0752-cert podName:58576398-c4a1-4369-8c0c-3b936c2c0752 nodeName:}" failed. No retries permitted until 2026-04-22 17:36:29.396968374 +0000 UTC m=+97.289664076 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/58576398-c4a1-4369-8c0c-3b936c2c0752-cert") pod "ingress-canary-jxlcn" (UID: "58576398-c4a1-4369-8c0c-3b936c2c0752") : secret "canary-serving-cert" not found Apr 22 17:35:58.302817 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:35:58.302768 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs\") pod \"network-metrics-daemon-zhl6m\" (UID: \"cff34b01-53b3-4374-b73d-d3146e3c558d\") " pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:35:58.303216 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:58.302937 2581 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: secret "metrics-daemon-secret" not found Apr 22 17:35:58.303216 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:35:58.303001 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs podName:cff34b01-53b3-4374-b73d-d3146e3c558d nodeName:}" failed. No retries permitted until 2026-04-22 17:37:02.302986889 +0000 UTC m=+130.195682595 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs") pod "network-metrics-daemon-zhl6m" (UID: "cff34b01-53b3-4374-b73d-d3146e3c558d") : secret "metrics-daemon-secret" not found Apr 22 17:36:02.850585 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:36:02.850543 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-p72t5" Apr 22 17:36:29.325664 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:36:29.325631 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d6515c87-3bb8-42d3-be06-5bf26230f609-metrics-tls\") pod \"dns-default-nqmll\" (UID: \"d6515c87-3bb8-42d3-be06-5bf26230f609\") " pod="openshift-dns/dns-default-nqmll" Apr 22 17:36:29.326051 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:36:29.325748 2581 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 22 17:36:29.326051 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:36:29.325799 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d6515c87-3bb8-42d3-be06-5bf26230f609-metrics-tls podName:d6515c87-3bb8-42d3-be06-5bf26230f609 nodeName:}" failed. No retries permitted until 2026-04-22 17:37:33.325786839 +0000 UTC m=+161.218482542 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/d6515c87-3bb8-42d3-be06-5bf26230f609-metrics-tls") pod "dns-default-nqmll" (UID: "d6515c87-3bb8-42d3-be06-5bf26230f609") : secret "dns-default-metrics-tls" not found Apr 22 17:36:29.426589 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:36:29.426555 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/58576398-c4a1-4369-8c0c-3b936c2c0752-cert\") pod \"ingress-canary-jxlcn\" (UID: \"58576398-c4a1-4369-8c0c-3b936c2c0752\") " pod="openshift-ingress-canary/ingress-canary-jxlcn" Apr 22 17:36:29.426719 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:36:29.426698 2581 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 22 17:36:29.426785 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:36:29.426776 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58576398-c4a1-4369-8c0c-3b936c2c0752-cert podName:58576398-c4a1-4369-8c0c-3b936c2c0752 nodeName:}" failed. No retries permitted until 2026-04-22 17:37:33.426760283 +0000 UTC m=+161.319455990 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/58576398-c4a1-4369-8c0c-3b936c2c0752-cert") pod "ingress-canary-jxlcn" (UID: "58576398-c4a1-4369-8c0c-3b936c2c0752") : secret "canary-serving-cert" not found Apr 22 17:37:02.361299 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:02.361251 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs\") pod \"network-metrics-daemon-zhl6m\" (UID: \"cff34b01-53b3-4374-b73d-d3146e3c558d\") " pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:37:02.361791 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:37:02.361426 2581 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: secret "metrics-daemon-secret" not found Apr 22 17:37:02.361791 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:37:02.361502 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs podName:cff34b01-53b3-4374-b73d-d3146e3c558d nodeName:}" failed. No retries permitted until 2026-04-22 17:39:04.361485693 +0000 UTC m=+252.254181395 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs") pod "network-metrics-daemon-zhl6m" (UID: "cff34b01-53b3-4374-b73d-d3146e3c558d") : secret "metrics-daemon-secret" not found Apr 22 17:37:14.084651 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:14.084615 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-d6fc45fc5-jf5fx"] Apr 22 17:37:14.087250 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:14.087231 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-jf5fx" Apr 22 17:37:14.089413 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:14.089388 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca-operator\"/\"openshift-service-ca.crt\"" Apr 22 17:37:14.089525 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:14.089453 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca-operator\"/\"service-ca-operator-config\"" Apr 22 17:37:14.089525 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:14.089469 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca-operator\"/\"service-ca-operator-dockercfg-5p2z5\"" Apr 22 17:37:14.090227 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:14.090211 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca-operator\"/\"kube-root-ca.crt\"" Apr 22 17:37:14.090280 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:14.090254 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca-operator\"/\"serving-cert\"" Apr 22 17:37:14.097304 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:14.097281 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-d6fc45fc5-jf5fx"] Apr 22 17:37:14.243301 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:14.243251 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbq5j\" (UniqueName: \"kubernetes.io/projected/18d358c9-5de4-4059-98a5-f3f648ff3d67-kube-api-access-jbq5j\") pod \"service-ca-operator-d6fc45fc5-jf5fx\" (UID: \"18d358c9-5de4-4059-98a5-f3f648ff3d67\") " pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-jf5fx" Apr 22 17:37:14.243481 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:14.243324 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18d358c9-5de4-4059-98a5-f3f648ff3d67-config\") pod \"service-ca-operator-d6fc45fc5-jf5fx\" (UID: \"18d358c9-5de4-4059-98a5-f3f648ff3d67\") " pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-jf5fx" Apr 22 17:37:14.243481 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:14.243367 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18d358c9-5de4-4059-98a5-f3f648ff3d67-serving-cert\") pod \"service-ca-operator-d6fc45fc5-jf5fx\" (UID: \"18d358c9-5de4-4059-98a5-f3f648ff3d67\") " pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-jf5fx" Apr 22 17:37:14.344591 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:14.344514 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18d358c9-5de4-4059-98a5-f3f648ff3d67-config\") pod \"service-ca-operator-d6fc45fc5-jf5fx\" (UID: \"18d358c9-5de4-4059-98a5-f3f648ff3d67\") " pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-jf5fx" Apr 22 17:37:14.344591 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:14.344554 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18d358c9-5de4-4059-98a5-f3f648ff3d67-serving-cert\") pod \"service-ca-operator-d6fc45fc5-jf5fx\" (UID: \"18d358c9-5de4-4059-98a5-f3f648ff3d67\") " pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-jf5fx" Apr 22 17:37:14.344741 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:14.344720 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-jbq5j\" (UniqueName: \"kubernetes.io/projected/18d358c9-5de4-4059-98a5-f3f648ff3d67-kube-api-access-jbq5j\") pod \"service-ca-operator-d6fc45fc5-jf5fx\" (UID: \"18d358c9-5de4-4059-98a5-f3f648ff3d67\") " pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-jf5fx" Apr 22 17:37:14.345167 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:14.345143 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18d358c9-5de4-4059-98a5-f3f648ff3d67-config\") pod \"service-ca-operator-d6fc45fc5-jf5fx\" (UID: \"18d358c9-5de4-4059-98a5-f3f648ff3d67\") " pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-jf5fx" Apr 22 17:37:14.346993 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:14.346973 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18d358c9-5de4-4059-98a5-f3f648ff3d67-serving-cert\") pod \"service-ca-operator-d6fc45fc5-jf5fx\" (UID: \"18d358c9-5de4-4059-98a5-f3f648ff3d67\") " pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-jf5fx" Apr 22 17:37:14.352991 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:14.352968 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbq5j\" (UniqueName: \"kubernetes.io/projected/18d358c9-5de4-4059-98a5-f3f648ff3d67-kube-api-access-jbq5j\") pod \"service-ca-operator-d6fc45fc5-jf5fx\" (UID: \"18d358c9-5de4-4059-98a5-f3f648ff3d67\") " pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-jf5fx" Apr 22 17:37:14.396070 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:14.396033 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-jf5fx" Apr 22 17:37:14.515183 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:14.515153 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-d6fc45fc5-jf5fx"] Apr 22 17:37:14.518826 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:37:14.518787 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18d358c9_5de4_4059_98a5_f3f648ff3d67.slice/crio-4b90d0a063969b187d9c8c478f07bc01efd655869ddc0acd679800db63ed72e1 WatchSource:0}: Error finding container 4b90d0a063969b187d9c8c478f07bc01efd655869ddc0acd679800db63ed72e1: Status 404 returned error can't find the container with id 4b90d0a063969b187d9c8c478f07bc01efd655869ddc0acd679800db63ed72e1 Apr 22 17:37:15.047064 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:15.047025 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-jf5fx" event={"ID":"18d358c9-5de4-4059-98a5-f3f648ff3d67","Type":"ContainerStarted","Data":"4b90d0a063969b187d9c8c478f07bc01efd655869ddc0acd679800db63ed72e1"} Apr 22 17:37:17.053242 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:17.053201 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-jf5fx" event={"ID":"18d358c9-5de4-4059-98a5-f3f648ff3d67","Type":"ContainerStarted","Data":"043e068f893e5a2e2880e20cd24ae29f793dce36bd0c13b823249ebbc6b53b28"} Apr 22 17:37:17.071311 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:17.071262 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-jf5fx" podStartSLOduration=1.259323616 podStartE2EDuration="3.071246749s" podCreationTimestamp="2026-04-22 17:37:14 +0000 UTC" firstStartedPulling="2026-04-22 17:37:14.521013555 +0000 UTC m=+142.413709258" lastFinishedPulling="2026-04-22 17:37:16.332936686 +0000 UTC m=+144.225632391" observedRunningTime="2026-04-22 17:37:17.070998175 +0000 UTC m=+144.963693929" watchObservedRunningTime="2026-04-22 17:37:17.071246749 +0000 UTC m=+144.963942473" Apr 22 17:37:18.426784 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:18.426757 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_node-resolver-4m4hg_5db20131-70d2-456b-a852-663399260efa/dns-node-resolver/0.log" Apr 22 17:37:19.425894 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:19.425854 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_node-ca-6zmdm_8dedc3a1-a7e9-4684-b53d-a1b087d23cdb/node-ca/0.log" Apr 22 17:37:20.327025 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:20.326989 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-865cb79987-qcr6z"] Apr 22 17:37:20.329859 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:20.329840 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-865cb79987-qcr6z" Apr 22 17:37:20.332327 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:20.332298 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"kube-root-ca.crt\"" Apr 22 17:37:20.332450 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:20.332422 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca\"/\"service-ca-dockercfg-rp7vc\"" Apr 22 17:37:20.333021 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:20.333004 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"signing-cabundle\"" Apr 22 17:37:20.333101 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:20.333034 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca\"/\"signing-key\"" Apr 22 17:37:20.333151 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:20.333037 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"openshift-service-ca.crt\"" Apr 22 17:37:20.340290 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:20.340268 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-865cb79987-qcr6z"] Apr 22 17:37:20.493132 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:20.493096 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/5d35b046-29e1-4ac1-9ee0-287e716df8bb-signing-key\") pod \"service-ca-865cb79987-qcr6z\" (UID: \"5d35b046-29e1-4ac1-9ee0-287e716df8bb\") " pod="openshift-service-ca/service-ca-865cb79987-qcr6z" Apr 22 17:37:20.493132 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:20.493129 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rksm\" (UniqueName: \"kubernetes.io/projected/5d35b046-29e1-4ac1-9ee0-287e716df8bb-kube-api-access-7rksm\") pod \"service-ca-865cb79987-qcr6z\" (UID: \"5d35b046-29e1-4ac1-9ee0-287e716df8bb\") " pod="openshift-service-ca/service-ca-865cb79987-qcr6z" Apr 22 17:37:20.493336 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:20.493154 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/5d35b046-29e1-4ac1-9ee0-287e716df8bb-signing-cabundle\") pod \"service-ca-865cb79987-qcr6z\" (UID: \"5d35b046-29e1-4ac1-9ee0-287e716df8bb\") " pod="openshift-service-ca/service-ca-865cb79987-qcr6z" Apr 22 17:37:20.594254 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:20.594166 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/5d35b046-29e1-4ac1-9ee0-287e716df8bb-signing-key\") pod \"service-ca-865cb79987-qcr6z\" (UID: \"5d35b046-29e1-4ac1-9ee0-287e716df8bb\") " pod="openshift-service-ca/service-ca-865cb79987-qcr6z" Apr 22 17:37:20.594254 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:20.594206 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-7rksm\" (UniqueName: \"kubernetes.io/projected/5d35b046-29e1-4ac1-9ee0-287e716df8bb-kube-api-access-7rksm\") pod \"service-ca-865cb79987-qcr6z\" (UID: \"5d35b046-29e1-4ac1-9ee0-287e716df8bb\") " pod="openshift-service-ca/service-ca-865cb79987-qcr6z" Apr 22 17:37:20.594254 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:20.594234 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/5d35b046-29e1-4ac1-9ee0-287e716df8bb-signing-cabundle\") pod \"service-ca-865cb79987-qcr6z\" (UID: \"5d35b046-29e1-4ac1-9ee0-287e716df8bb\") " pod="openshift-service-ca/service-ca-865cb79987-qcr6z" Apr 22 17:37:20.594860 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:20.594837 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/5d35b046-29e1-4ac1-9ee0-287e716df8bb-signing-cabundle\") pod \"service-ca-865cb79987-qcr6z\" (UID: \"5d35b046-29e1-4ac1-9ee0-287e716df8bb\") " pod="openshift-service-ca/service-ca-865cb79987-qcr6z" Apr 22 17:37:20.596816 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:20.596794 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/5d35b046-29e1-4ac1-9ee0-287e716df8bb-signing-key\") pod \"service-ca-865cb79987-qcr6z\" (UID: \"5d35b046-29e1-4ac1-9ee0-287e716df8bb\") " pod="openshift-service-ca/service-ca-865cb79987-qcr6z" Apr 22 17:37:20.602476 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:20.602455 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rksm\" (UniqueName: \"kubernetes.io/projected/5d35b046-29e1-4ac1-9ee0-287e716df8bb-kube-api-access-7rksm\") pod \"service-ca-865cb79987-qcr6z\" (UID: \"5d35b046-29e1-4ac1-9ee0-287e716df8bb\") " pod="openshift-service-ca/service-ca-865cb79987-qcr6z" Apr 22 17:37:20.638075 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:20.638044 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-865cb79987-qcr6z" Apr 22 17:37:20.754483 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:20.754444 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-865cb79987-qcr6z"] Apr 22 17:37:20.757910 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:37:20.757846 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d35b046_29e1_4ac1_9ee0_287e716df8bb.slice/crio-238b0748744e184f0a26c1209183dcaaf1e17641f6e1509b2af88cf41c202238 WatchSource:0}: Error finding container 238b0748744e184f0a26c1209183dcaaf1e17641f6e1509b2af88cf41c202238: Status 404 returned error can't find the container with id 238b0748744e184f0a26c1209183dcaaf1e17641f6e1509b2af88cf41c202238 Apr 22 17:37:21.064327 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:21.064290 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-865cb79987-qcr6z" event={"ID":"5d35b046-29e1-4ac1-9ee0-287e716df8bb","Type":"ContainerStarted","Data":"859b6cee47617f8b68deea030e906d42eb5cf6a5fb3e3b7c35e4fe7fd2f6481d"} Apr 22 17:37:21.064327 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:21.064330 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-865cb79987-qcr6z" event={"ID":"5d35b046-29e1-4ac1-9ee0-287e716df8bb","Type":"ContainerStarted","Data":"238b0748744e184f0a26c1209183dcaaf1e17641f6e1509b2af88cf41c202238"} Apr 22 17:37:21.083498 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:21.083441 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-865cb79987-qcr6z" podStartSLOduration=1.0834196249999999 podStartE2EDuration="1.083419625s" podCreationTimestamp="2026-04-22 17:37:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-22 17:37:21.082127331 +0000 UTC m=+148.974823048" watchObservedRunningTime="2026-04-22 17:37:21.083419625 +0000 UTC m=+148.976115350" Apr 22 17:37:28.510227 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:37:28.510179 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[metrics-tls], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-dns/dns-default-nqmll" podUID="d6515c87-3bb8-42d3-be06-5bf26230f609" Apr 22 17:37:28.524373 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:37:28.524344 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[cert], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-ingress-canary/ingress-canary-jxlcn" podUID="58576398-c4a1-4369-8c0c-3b936c2c0752" Apr 22 17:37:28.652014 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:37:28.651978 2581 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[metrics-certs], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-multus/network-metrics-daemon-zhl6m" podUID="cff34b01-53b3-4374-b73d-d3146e3c558d" Apr 22 17:37:29.087622 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:29.087589 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-nqmll" Apr 22 17:37:33.392780 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:33.392699 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d6515c87-3bb8-42d3-be06-5bf26230f609-metrics-tls\") pod \"dns-default-nqmll\" (UID: \"d6515c87-3bb8-42d3-be06-5bf26230f609\") " pod="openshift-dns/dns-default-nqmll" Apr 22 17:37:33.395124 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:33.395104 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/d6515c87-3bb8-42d3-be06-5bf26230f609-metrics-tls\") pod \"dns-default-nqmll\" (UID: \"d6515c87-3bb8-42d3-be06-5bf26230f609\") " pod="openshift-dns/dns-default-nqmll" Apr 22 17:37:33.493576 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:33.493539 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/58576398-c4a1-4369-8c0c-3b936c2c0752-cert\") pod \"ingress-canary-jxlcn\" (UID: \"58576398-c4a1-4369-8c0c-3b936c2c0752\") " pod="openshift-ingress-canary/ingress-canary-jxlcn" Apr 22 17:37:33.496090 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:33.496071 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/58576398-c4a1-4369-8c0c-3b936c2c0752-cert\") pod \"ingress-canary-jxlcn\" (UID: \"58576398-c4a1-4369-8c0c-3b936c2c0752\") " pod="openshift-ingress-canary/ingress-canary-jxlcn" Apr 22 17:37:33.591555 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:33.591522 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-dockercfg-m6rcd\"" Apr 22 17:37:33.599342 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:33.599316 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-nqmll" Apr 22 17:37:33.718679 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:33.718649 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-nqmll"] Apr 22 17:37:33.721903 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:37:33.721841 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6515c87_3bb8_42d3_be06_5bf26230f609.slice/crio-b33919c12470f4b74e5a46a64f2d18152e383f49c74c81c275986598cf14302c WatchSource:0}: Error finding container b33919c12470f4b74e5a46a64f2d18152e383f49c74c81c275986598cf14302c: Status 404 returned error can't find the container with id b33919c12470f4b74e5a46a64f2d18152e383f49c74c81c275986598cf14302c Apr 22 17:37:34.100201 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:34.100165 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-nqmll" event={"ID":"d6515c87-3bb8-42d3-be06-5bf26230f609","Type":"ContainerStarted","Data":"b33919c12470f4b74e5a46a64f2d18152e383f49c74c81c275986598cf14302c"} Apr 22 17:37:35.104559 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:35.104532 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-nqmll" event={"ID":"d6515c87-3bb8-42d3-be06-5bf26230f609","Type":"ContainerStarted","Data":"8bc2fb2bc23750177f60035e7fe30a5d53108fedef8310b1fb754bb29ee9ac02"} Apr 22 17:37:36.108297 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:36.108259 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-nqmll" event={"ID":"d6515c87-3bb8-42d3-be06-5bf26230f609","Type":"ContainerStarted","Data":"f544c9e7254be2bc4b9831e4cec832b0113743463b8fbcada4cd1b79c08fc561"} Apr 22 17:37:36.108690 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:36.108386 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-dns/dns-default-nqmll" Apr 22 17:37:36.126532 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:36.126483 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-nqmll" podStartSLOduration=129.947801156 podStartE2EDuration="2m11.126466476s" podCreationTimestamp="2026-04-22 17:35:25 +0000 UTC" firstStartedPulling="2026-04-22 17:37:33.723528416 +0000 UTC m=+161.616224118" lastFinishedPulling="2026-04-22 17:37:34.902193721 +0000 UTC m=+162.794889438" observedRunningTime="2026-04-22 17:37:36.125939477 +0000 UTC m=+164.018635204" watchObservedRunningTime="2026-04-22 17:37:36.126466476 +0000 UTC m=+164.019162199" Apr 22 17:37:42.018155 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.018121 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-69fdd76d7d-jtkk2"] Apr 22 17:37:42.021089 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.021073 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:37:42.023126 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.023102 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-private-configuration\"" Apr 22 17:37:42.023126 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.023118 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"registry-dockercfg-xzdz4\"" Apr 22 17:37:42.024478 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.024457 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"installation-pull-secrets\"" Apr 22 17:37:42.024589 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.024466 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-tls\"" Apr 22 17:37:42.031644 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.031627 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"trusted-ca\"" Apr 22 17:37:42.039052 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.039030 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-insights/insights-runtime-extractor-5bx7t"] Apr 22 17:37:42.041864 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.041850 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-runtime-extractor-5bx7t" Apr 22 17:37:42.047163 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.047142 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"insights-runtime-extractor-sa-dockercfg-tt9tl\"" Apr 22 17:37:42.047943 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.047918 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"kube-root-ca.crt\"" Apr 22 17:37:42.047943 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.047929 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"insights-runtime-extractor-tls\"" Apr 22 17:37:42.048313 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.048164 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"openshift-service-ca.crt\"" Apr 22 17:37:42.048313 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.048187 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"kube-rbac-proxy\"" Apr 22 17:37:42.048652 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.048632 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-69fdd76d7d-jtkk2"] Apr 22 17:37:42.053613 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.053595 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/01529b59-d931-40c9-9938-179fb72019e7-bound-sa-token\") pod \"image-registry-69fdd76d7d-jtkk2\" (UID: \"01529b59-d931-40c9-9938-179fb72019e7\") " pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:37:42.053708 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.053636 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/01529b59-d931-40c9-9938-179fb72019e7-installation-pull-secrets\") pod \"image-registry-69fdd76d7d-jtkk2\" (UID: \"01529b59-d931-40c9-9938-179fb72019e7\") " pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:37:42.053708 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.053657 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/01529b59-d931-40c9-9938-179fb72019e7-image-registry-private-configuration\") pod \"image-registry-69fdd76d7d-jtkk2\" (UID: \"01529b59-d931-40c9-9938-179fb72019e7\") " pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:37:42.053708 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.053672 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01529b59-d931-40c9-9938-179fb72019e7-trusted-ca\") pod \"image-registry-69fdd76d7d-jtkk2\" (UID: \"01529b59-d931-40c9-9938-179fb72019e7\") " pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:37:42.053845 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.053720 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/01529b59-d931-40c9-9938-179fb72019e7-registry-tls\") pod \"image-registry-69fdd76d7d-jtkk2\" (UID: \"01529b59-d931-40c9-9938-179fb72019e7\") " pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:37:42.053845 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.053741 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/01529b59-d931-40c9-9938-179fb72019e7-registry-certificates\") pod \"image-registry-69fdd76d7d-jtkk2\" (UID: \"01529b59-d931-40c9-9938-179fb72019e7\") " pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:37:42.053845 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.053761 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/01529b59-d931-40c9-9938-179fb72019e7-ca-trust-extracted\") pod \"image-registry-69fdd76d7d-jtkk2\" (UID: \"01529b59-d931-40c9-9938-179fb72019e7\") " pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:37:42.053845 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.053783 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwvwb\" (UniqueName: \"kubernetes.io/projected/01529b59-d931-40c9-9938-179fb72019e7-kube-api-access-nwvwb\") pod \"image-registry-69fdd76d7d-jtkk2\" (UID: \"01529b59-d931-40c9-9938-179fb72019e7\") " pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:37:42.088026 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.087999 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-runtime-extractor-5bx7t"] Apr 22 17:37:42.155008 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.154982 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/01529b59-d931-40c9-9938-179fb72019e7-bound-sa-token\") pod \"image-registry-69fdd76d7d-jtkk2\" (UID: \"01529b59-d931-40c9-9938-179fb72019e7\") " pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:37:42.155134 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.155012 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/19c6390d-ed21-45bc-ab1f-e2df9d1ee64b-crio-socket\") pod \"insights-runtime-extractor-5bx7t\" (UID: \"19c6390d-ed21-45bc-ab1f-e2df9d1ee64b\") " pod="openshift-insights/insights-runtime-extractor-5bx7t" Apr 22 17:37:42.155134 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.155071 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/01529b59-d931-40c9-9938-179fb72019e7-installation-pull-secrets\") pod \"image-registry-69fdd76d7d-jtkk2\" (UID: \"01529b59-d931-40c9-9938-179fb72019e7\") " pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:37:42.155134 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.155099 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/19c6390d-ed21-45bc-ab1f-e2df9d1ee64b-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-5bx7t\" (UID: \"19c6390d-ed21-45bc-ab1f-e2df9d1ee64b\") " pod="openshift-insights/insights-runtime-extractor-5bx7t" Apr 22 17:37:42.155228 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.155219 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/19c6390d-ed21-45bc-ab1f-e2df9d1ee64b-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-5bx7t\" (UID: \"19c6390d-ed21-45bc-ab1f-e2df9d1ee64b\") " pod="openshift-insights/insights-runtime-extractor-5bx7t" Apr 22 17:37:42.155263 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.155246 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/01529b59-d931-40c9-9938-179fb72019e7-ca-trust-extracted\") pod \"image-registry-69fdd76d7d-jtkk2\" (UID: \"01529b59-d931-40c9-9938-179fb72019e7\") " pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:37:42.155303 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.155278 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/01529b59-d931-40c9-9938-179fb72019e7-image-registry-private-configuration\") pod \"image-registry-69fdd76d7d-jtkk2\" (UID: \"01529b59-d931-40c9-9938-179fb72019e7\") " pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:37:42.155356 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.155313 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01529b59-d931-40c9-9938-179fb72019e7-trusted-ca\") pod \"image-registry-69fdd76d7d-jtkk2\" (UID: \"01529b59-d931-40c9-9938-179fb72019e7\") " pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:37:42.155415 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.155393 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/01529b59-d931-40c9-9938-179fb72019e7-registry-tls\") pod \"image-registry-69fdd76d7d-jtkk2\" (UID: \"01529b59-d931-40c9-9938-179fb72019e7\") " pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:37:42.155468 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.155432 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/19c6390d-ed21-45bc-ab1f-e2df9d1ee64b-data-volume\") pod \"insights-runtime-extractor-5bx7t\" (UID: \"19c6390d-ed21-45bc-ab1f-e2df9d1ee64b\") " pod="openshift-insights/insights-runtime-extractor-5bx7t" Apr 22 17:37:42.155468 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.155462 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/01529b59-d931-40c9-9938-179fb72019e7-registry-certificates\") pod \"image-registry-69fdd76d7d-jtkk2\" (UID: \"01529b59-d931-40c9-9938-179fb72019e7\") " pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:37:42.155576 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.155496 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nwvwb\" (UniqueName: \"kubernetes.io/projected/01529b59-d931-40c9-9938-179fb72019e7-kube-api-access-nwvwb\") pod \"image-registry-69fdd76d7d-jtkk2\" (UID: \"01529b59-d931-40c9-9938-179fb72019e7\") " pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:37:42.155576 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.155550 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfpss\" (UniqueName: \"kubernetes.io/projected/19c6390d-ed21-45bc-ab1f-e2df9d1ee64b-kube-api-access-sfpss\") pod \"insights-runtime-extractor-5bx7t\" (UID: \"19c6390d-ed21-45bc-ab1f-e2df9d1ee64b\") " pod="openshift-insights/insights-runtime-extractor-5bx7t" Apr 22 17:37:42.155691 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.155670 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/01529b59-d931-40c9-9938-179fb72019e7-ca-trust-extracted\") pod \"image-registry-69fdd76d7d-jtkk2\" (UID: \"01529b59-d931-40c9-9938-179fb72019e7\") " pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:37:42.156272 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.156251 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/01529b59-d931-40c9-9938-179fb72019e7-registry-certificates\") pod \"image-registry-69fdd76d7d-jtkk2\" (UID: \"01529b59-d931-40c9-9938-179fb72019e7\") " pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:37:42.156792 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.156772 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/01529b59-d931-40c9-9938-179fb72019e7-trusted-ca\") pod \"image-registry-69fdd76d7d-jtkk2\" (UID: \"01529b59-d931-40c9-9938-179fb72019e7\") " pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:37:42.157694 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.157675 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/01529b59-d931-40c9-9938-179fb72019e7-installation-pull-secrets\") pod \"image-registry-69fdd76d7d-jtkk2\" (UID: \"01529b59-d931-40c9-9938-179fb72019e7\") " pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:37:42.158072 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.158057 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/01529b59-d931-40c9-9938-179fb72019e7-registry-tls\") pod \"image-registry-69fdd76d7d-jtkk2\" (UID: \"01529b59-d931-40c9-9938-179fb72019e7\") " pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:37:42.158430 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.158410 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/01529b59-d931-40c9-9938-179fb72019e7-image-registry-private-configuration\") pod \"image-registry-69fdd76d7d-jtkk2\" (UID: \"01529b59-d931-40c9-9938-179fb72019e7\") " pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:37:42.162804 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.162781 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/01529b59-d931-40c9-9938-179fb72019e7-bound-sa-token\") pod \"image-registry-69fdd76d7d-jtkk2\" (UID: \"01529b59-d931-40c9-9938-179fb72019e7\") " pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:37:42.164745 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.164724 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwvwb\" (UniqueName: \"kubernetes.io/projected/01529b59-d931-40c9-9938-179fb72019e7-kube-api-access-nwvwb\") pod \"image-registry-69fdd76d7d-jtkk2\" (UID: \"01529b59-d931-40c9-9938-179fb72019e7\") " pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:37:42.256866 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.256832 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/19c6390d-ed21-45bc-ab1f-e2df9d1ee64b-crio-socket\") pod \"insights-runtime-extractor-5bx7t\" (UID: \"19c6390d-ed21-45bc-ab1f-e2df9d1ee64b\") " pod="openshift-insights/insights-runtime-extractor-5bx7t" Apr 22 17:37:42.257016 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.256905 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/19c6390d-ed21-45bc-ab1f-e2df9d1ee64b-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-5bx7t\" (UID: \"19c6390d-ed21-45bc-ab1f-e2df9d1ee64b\") " pod="openshift-insights/insights-runtime-extractor-5bx7t" Apr 22 17:37:42.257016 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.256926 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/19c6390d-ed21-45bc-ab1f-e2df9d1ee64b-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-5bx7t\" (UID: \"19c6390d-ed21-45bc-ab1f-e2df9d1ee64b\") " pod="openshift-insights/insights-runtime-extractor-5bx7t" Apr 22 17:37:42.257016 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.256937 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/19c6390d-ed21-45bc-ab1f-e2df9d1ee64b-crio-socket\") pod \"insights-runtime-extractor-5bx7t\" (UID: \"19c6390d-ed21-45bc-ab1f-e2df9d1ee64b\") " pod="openshift-insights/insights-runtime-extractor-5bx7t" Apr 22 17:37:42.257016 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.256978 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/19c6390d-ed21-45bc-ab1f-e2df9d1ee64b-data-volume\") pod \"insights-runtime-extractor-5bx7t\" (UID: \"19c6390d-ed21-45bc-ab1f-e2df9d1ee64b\") " pod="openshift-insights/insights-runtime-extractor-5bx7t" Apr 22 17:37:42.257016 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.257008 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-sfpss\" (UniqueName: \"kubernetes.io/projected/19c6390d-ed21-45bc-ab1f-e2df9d1ee64b-kube-api-access-sfpss\") pod \"insights-runtime-extractor-5bx7t\" (UID: \"19c6390d-ed21-45bc-ab1f-e2df9d1ee64b\") " pod="openshift-insights/insights-runtime-extractor-5bx7t" Apr 22 17:37:42.257422 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.257406 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/19c6390d-ed21-45bc-ab1f-e2df9d1ee64b-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-5bx7t\" (UID: \"19c6390d-ed21-45bc-ab1f-e2df9d1ee64b\") " pod="openshift-insights/insights-runtime-extractor-5bx7t" Apr 22 17:37:42.257836 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.257817 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/19c6390d-ed21-45bc-ab1f-e2df9d1ee64b-data-volume\") pod \"insights-runtime-extractor-5bx7t\" (UID: \"19c6390d-ed21-45bc-ab1f-e2df9d1ee64b\") " pod="openshift-insights/insights-runtime-extractor-5bx7t" Apr 22 17:37:42.259699 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.259678 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/19c6390d-ed21-45bc-ab1f-e2df9d1ee64b-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-5bx7t\" (UID: \"19c6390d-ed21-45bc-ab1f-e2df9d1ee64b\") " pod="openshift-insights/insights-runtime-extractor-5bx7t" Apr 22 17:37:42.267256 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.267234 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfpss\" (UniqueName: \"kubernetes.io/projected/19c6390d-ed21-45bc-ab1f-e2df9d1ee64b-kube-api-access-sfpss\") pod \"insights-runtime-extractor-5bx7t\" (UID: \"19c6390d-ed21-45bc-ab1f-e2df9d1ee64b\") " pod="openshift-insights/insights-runtime-extractor-5bx7t" Apr 22 17:37:42.329315 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.329236 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:37:42.350165 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.350141 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-runtime-extractor-5bx7t" Apr 22 17:37:42.457466 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.457439 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-69fdd76d7d-jtkk2"] Apr 22 17:37:42.460601 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:37:42.460575 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01529b59_d931_40c9_9938_179fb72019e7.slice/crio-507678f2f596f2c63b9c556069c447705c8aeae5ebff21ddfea73cb63b153e74 WatchSource:0}: Error finding container 507678f2f596f2c63b9c556069c447705c8aeae5ebff21ddfea73cb63b153e74: Status 404 returned error can't find the container with id 507678f2f596f2c63b9c556069c447705c8aeae5ebff21ddfea73cb63b153e74 Apr 22 17:37:42.477860 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:42.477839 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-runtime-extractor-5bx7t"] Apr 22 17:37:42.498936 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:37:42.498907 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19c6390d_ed21_45bc_ab1f_e2df9d1ee64b.slice/crio-d175b0887f3121680fb809c3783b396ca090f979655a9780066de40e17c7bfd7 WatchSource:0}: Error finding container d175b0887f3121680fb809c3783b396ca090f979655a9780066de40e17c7bfd7: Status 404 returned error can't find the container with id d175b0887f3121680fb809c3783b396ca090f979655a9780066de40e17c7bfd7 Apr 22 17:37:43.127702 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:43.127662 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-5bx7t" event={"ID":"19c6390d-ed21-45bc-ab1f-e2df9d1ee64b","Type":"ContainerStarted","Data":"c712af6672ba399111f79b22795008d307ad72b51de42d928426f6baca5974c9"} Apr 22 17:37:43.128143 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:43.127715 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-5bx7t" event={"ID":"19c6390d-ed21-45bc-ab1f-e2df9d1ee64b","Type":"ContainerStarted","Data":"d175b0887f3121680fb809c3783b396ca090f979655a9780066de40e17c7bfd7"} Apr 22 17:37:43.128900 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:43.128860 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" event={"ID":"01529b59-d931-40c9-9938-179fb72019e7","Type":"ContainerStarted","Data":"3b83c3ee7794d1d3ee1ddf864a7c4a7f26aa931cfcc42a9144b392817dccbb1a"} Apr 22 17:37:43.128986 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:43.128910 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" event={"ID":"01529b59-d931-40c9-9938-179fb72019e7","Type":"ContainerStarted","Data":"507678f2f596f2c63b9c556069c447705c8aeae5ebff21ddfea73cb63b153e74"} Apr 22 17:37:43.129030 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:43.129013 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:37:43.147789 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:43.147741 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" podStartSLOduration=2.147723809 podStartE2EDuration="2.147723809s" podCreationTimestamp="2026-04-22 17:37:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-22 17:37:43.146760816 +0000 UTC m=+171.039456542" watchObservedRunningTime="2026-04-22 17:37:43.147723809 +0000 UTC m=+171.040419532" Apr 22 17:37:43.635470 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:43.635435 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-jxlcn" Apr 22 17:37:43.635470 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:43.635435 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:37:43.637914 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:43.637895 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"default-dockercfg-dt854\"" Apr 22 17:37:43.646814 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:43.646782 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-jxlcn" Apr 22 17:37:43.779374 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:43.779343 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-jxlcn"] Apr 22 17:37:43.782839 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:37:43.782805 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58576398_c4a1_4369_8c0c_3b936c2c0752.slice/crio-73bcfeda6bb049ea33a2557fd71c4dfd5896e4c8a0d3ef53a3f7b285e1412848 WatchSource:0}: Error finding container 73bcfeda6bb049ea33a2557fd71c4dfd5896e4c8a0d3ef53a3f7b285e1412848: Status 404 returned error can't find the container with id 73bcfeda6bb049ea33a2557fd71c4dfd5896e4c8a0d3ef53a3f7b285e1412848 Apr 22 17:37:44.133863 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:44.133808 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-5bx7t" event={"ID":"19c6390d-ed21-45bc-ab1f-e2df9d1ee64b","Type":"ContainerStarted","Data":"fc648ef66824cb119612f36c3c58518e45c42bdd2b4b3e7b560604837f891441"} Apr 22 17:37:44.135086 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:44.135040 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-jxlcn" event={"ID":"58576398-c4a1-4369-8c0c-3b936c2c0752","Type":"ContainerStarted","Data":"73bcfeda6bb049ea33a2557fd71c4dfd5896e4c8a0d3ef53a3f7b285e1412848"} Apr 22 17:37:45.140196 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:45.140153 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-5bx7t" event={"ID":"19c6390d-ed21-45bc-ab1f-e2df9d1ee64b","Type":"ContainerStarted","Data":"391b31d008eb90002f8be8b0b6b2ece9dfa0d2613885d1daa90629801a64c06c"} Apr 22 17:37:45.158058 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:45.158004 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-insights/insights-runtime-extractor-5bx7t" podStartSLOduration=1.047820407 podStartE2EDuration="3.157989774s" podCreationTimestamp="2026-04-22 17:37:42 +0000 UTC" firstStartedPulling="2026-04-22 17:37:42.542812939 +0000 UTC m=+170.435508648" lastFinishedPulling="2026-04-22 17:37:44.652982299 +0000 UTC m=+172.545678015" observedRunningTime="2026-04-22 17:37:45.156981119 +0000 UTC m=+173.049676843" watchObservedRunningTime="2026-04-22 17:37:45.157989774 +0000 UTC m=+173.050685497" Apr 22 17:37:46.112737 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:46.112707 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-nqmll" Apr 22 17:37:46.144101 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:46.144064 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-jxlcn" event={"ID":"58576398-c4a1-4369-8c0c-3b936c2c0752","Type":"ContainerStarted","Data":"f329a17d8685a58c623c27e0b1d8b3f74fe878bb675a607e5b87dfa030ce3d3a"} Apr 22 17:37:46.160221 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:46.160173 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-jxlcn" podStartSLOduration=139.648067572 podStartE2EDuration="2m21.160156386s" podCreationTimestamp="2026-04-22 17:35:25 +0000 UTC" firstStartedPulling="2026-04-22 17:37:43.785067126 +0000 UTC m=+171.677762828" lastFinishedPulling="2026-04-22 17:37:45.29715594 +0000 UTC m=+173.189851642" observedRunningTime="2026-04-22 17:37:46.159169335 +0000 UTC m=+174.051865085" watchObservedRunningTime="2026-04-22 17:37:46.160156386 +0000 UTC m=+174.052852101" Apr 22 17:37:48.546051 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:48.546015 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-5676c8c784-mghw5"] Apr 22 17:37:48.550836 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:48.550814 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-5676c8c784-mghw5" Apr 22 17:37:48.553803 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:48.553778 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-operator-tls\"" Apr 22 17:37:48.553803 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:48.553800 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"metrics-client-ca\"" Apr 22 17:37:48.553987 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:48.553848 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"openshift-service-ca.crt\"" Apr 22 17:37:48.553987 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:48.553861 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-operator-kube-rbac-proxy-config\"" Apr 22 17:37:48.553987 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:48.553864 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-operator-dockercfg-d5fhr\"" Apr 22 17:37:48.554154 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:48.554081 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"kube-root-ca.crt\"" Apr 22 17:37:48.558320 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:48.558300 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-5676c8c784-mghw5"] Apr 22 17:37:48.608010 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:48.607985 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/fa6843bd-5517-4faa-851b-99232e1b972e-metrics-client-ca\") pod \"prometheus-operator-5676c8c784-mghw5\" (UID: \"fa6843bd-5517-4faa-851b-99232e1b972e\") " pod="openshift-monitoring/prometheus-operator-5676c8c784-mghw5" Apr 22 17:37:48.608128 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:48.608013 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/fa6843bd-5517-4faa-851b-99232e1b972e-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-5676c8c784-mghw5\" (UID: \"fa6843bd-5517-4faa-851b-99232e1b972e\") " pod="openshift-monitoring/prometheus-operator-5676c8c784-mghw5" Apr 22 17:37:48.608128 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:48.608033 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/fa6843bd-5517-4faa-851b-99232e1b972e-prometheus-operator-tls\") pod \"prometheus-operator-5676c8c784-mghw5\" (UID: \"fa6843bd-5517-4faa-851b-99232e1b972e\") " pod="openshift-monitoring/prometheus-operator-5676c8c784-mghw5" Apr 22 17:37:48.608195 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:48.608145 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjtx4\" (UniqueName: \"kubernetes.io/projected/fa6843bd-5517-4faa-851b-99232e1b972e-kube-api-access-qjtx4\") pod \"prometheus-operator-5676c8c784-mghw5\" (UID: \"fa6843bd-5517-4faa-851b-99232e1b972e\") " pod="openshift-monitoring/prometheus-operator-5676c8c784-mghw5" Apr 22 17:37:48.708736 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:48.708709 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-qjtx4\" (UniqueName: \"kubernetes.io/projected/fa6843bd-5517-4faa-851b-99232e1b972e-kube-api-access-qjtx4\") pod \"prometheus-operator-5676c8c784-mghw5\" (UID: \"fa6843bd-5517-4faa-851b-99232e1b972e\") " pod="openshift-monitoring/prometheus-operator-5676c8c784-mghw5" Apr 22 17:37:48.708927 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:48.708764 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/fa6843bd-5517-4faa-851b-99232e1b972e-metrics-client-ca\") pod \"prometheus-operator-5676c8c784-mghw5\" (UID: \"fa6843bd-5517-4faa-851b-99232e1b972e\") " pod="openshift-monitoring/prometheus-operator-5676c8c784-mghw5" Apr 22 17:37:48.708927 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:48.708828 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/fa6843bd-5517-4faa-851b-99232e1b972e-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-5676c8c784-mghw5\" (UID: \"fa6843bd-5517-4faa-851b-99232e1b972e\") " pod="openshift-monitoring/prometheus-operator-5676c8c784-mghw5" Apr 22 17:37:48.708927 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:48.708865 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/fa6843bd-5517-4faa-851b-99232e1b972e-prometheus-operator-tls\") pod \"prometheus-operator-5676c8c784-mghw5\" (UID: \"fa6843bd-5517-4faa-851b-99232e1b972e\") " pod="openshift-monitoring/prometheus-operator-5676c8c784-mghw5" Apr 22 17:37:48.709465 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:48.709441 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/fa6843bd-5517-4faa-851b-99232e1b972e-metrics-client-ca\") pod \"prometheus-operator-5676c8c784-mghw5\" (UID: \"fa6843bd-5517-4faa-851b-99232e1b972e\") " pod="openshift-monitoring/prometheus-operator-5676c8c784-mghw5" Apr 22 17:37:48.711357 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:48.711337 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/fa6843bd-5517-4faa-851b-99232e1b972e-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-5676c8c784-mghw5\" (UID: \"fa6843bd-5517-4faa-851b-99232e1b972e\") " pod="openshift-monitoring/prometheus-operator-5676c8c784-mghw5" Apr 22 17:37:48.711432 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:48.711407 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/fa6843bd-5517-4faa-851b-99232e1b972e-prometheus-operator-tls\") pod \"prometheus-operator-5676c8c784-mghw5\" (UID: \"fa6843bd-5517-4faa-851b-99232e1b972e\") " pod="openshift-monitoring/prometheus-operator-5676c8c784-mghw5" Apr 22 17:37:48.716517 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:48.716494 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjtx4\" (UniqueName: \"kubernetes.io/projected/fa6843bd-5517-4faa-851b-99232e1b972e-kube-api-access-qjtx4\") pod \"prometheus-operator-5676c8c784-mghw5\" (UID: \"fa6843bd-5517-4faa-851b-99232e1b972e\") " pod="openshift-monitoring/prometheus-operator-5676c8c784-mghw5" Apr 22 17:37:48.859979 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:48.859898 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-5676c8c784-mghw5" Apr 22 17:37:48.976403 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:48.976374 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-5676c8c784-mghw5"] Apr 22 17:37:48.979592 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:37:48.979556 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa6843bd_5517_4faa_851b_99232e1b972e.slice/crio-ec72ee0ef918300dfb7ea221b68b7d3bea8ecd2e9fff7a3992f3d0bcc3fbe37e WatchSource:0}: Error finding container ec72ee0ef918300dfb7ea221b68b7d3bea8ecd2e9fff7a3992f3d0bcc3fbe37e: Status 404 returned error can't find the container with id ec72ee0ef918300dfb7ea221b68b7d3bea8ecd2e9fff7a3992f3d0bcc3fbe37e Apr 22 17:37:49.153219 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:49.153132 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-5676c8c784-mghw5" event={"ID":"fa6843bd-5517-4faa-851b-99232e1b972e","Type":"ContainerStarted","Data":"ec72ee0ef918300dfb7ea221b68b7d3bea8ecd2e9fff7a3992f3d0bcc3fbe37e"} Apr 22 17:37:51.159553 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:51.159518 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-5676c8c784-mghw5" event={"ID":"fa6843bd-5517-4faa-851b-99232e1b972e","Type":"ContainerStarted","Data":"f969d7ea32909075d9edcf5782b4563b1274cd986b77587810e3d33881487906"} Apr 22 17:37:51.159553 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:51.159554 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-5676c8c784-mghw5" event={"ID":"fa6843bd-5517-4faa-851b-99232e1b972e","Type":"ContainerStarted","Data":"67000a220f66361fe19bbc2b8328517963642679e992bd1f7531417a86b45c4b"} Apr 22 17:37:51.182512 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:51.182455 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-5676c8c784-mghw5" podStartSLOduration=2.041409628 podStartE2EDuration="3.182439106s" podCreationTimestamp="2026-04-22 17:37:48 +0000 UTC" firstStartedPulling="2026-04-22 17:37:48.981499304 +0000 UTC m=+176.874195008" lastFinishedPulling="2026-04-22 17:37:50.122528781 +0000 UTC m=+178.015224486" observedRunningTime="2026-04-22 17:37:51.182271288 +0000 UTC m=+179.074967013" watchObservedRunningTime="2026-04-22 17:37:51.182439106 +0000 UTC m=+179.075134830" Apr 22 17:37:52.083039 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.083004 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-758964cdfb-8wtwd"] Apr 22 17:37:52.085773 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.085757 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:37:52.087937 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.087916 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"service-ca\"" Apr 22 17:37:52.088056 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.087944 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"oauth-serving-cert\"" Apr 22 17:37:52.088678 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.088609 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"console-config\"" Apr 22 17:37:52.088784 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.088719 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-oauth-config\"" Apr 22 17:37:52.088784 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.088725 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"openshift-service-ca.crt\"" Apr 22 17:37:52.088893 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.088835 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-serving-cert\"" Apr 22 17:37:52.088949 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.088924 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-dockercfg-m79jb\"" Apr 22 17:37:52.089040 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.089025 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"kube-root-ca.crt\"" Apr 22 17:37:52.094077 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.094058 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"trusted-ca-bundle\"" Apr 22 17:37:52.096619 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.096597 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-758964cdfb-8wtwd"] Apr 22 17:37:52.134396 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.134361 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5723ab46-57a5-4f9c-8391-cbb6e58fb651-console-oauth-config\") pod \"console-758964cdfb-8wtwd\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:37:52.134396 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.134400 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5723ab46-57a5-4f9c-8391-cbb6e58fb651-service-ca\") pod \"console-758964cdfb-8wtwd\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:37:52.134539 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.134417 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l94q6\" (UniqueName: \"kubernetes.io/projected/5723ab46-57a5-4f9c-8391-cbb6e58fb651-kube-api-access-l94q6\") pod \"console-758964cdfb-8wtwd\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:37:52.134539 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.134443 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5723ab46-57a5-4f9c-8391-cbb6e58fb651-oauth-serving-cert\") pod \"console-758964cdfb-8wtwd\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:37:52.134539 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.134484 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5723ab46-57a5-4f9c-8391-cbb6e58fb651-console-serving-cert\") pod \"console-758964cdfb-8wtwd\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:37:52.134539 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.134521 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5723ab46-57a5-4f9c-8391-cbb6e58fb651-console-config\") pod \"console-758964cdfb-8wtwd\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:37:52.134669 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.134546 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5723ab46-57a5-4f9c-8391-cbb6e58fb651-trusted-ca-bundle\") pod \"console-758964cdfb-8wtwd\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:37:52.234923 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.234866 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5723ab46-57a5-4f9c-8391-cbb6e58fb651-service-ca\") pod \"console-758964cdfb-8wtwd\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:37:52.235346 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.234935 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l94q6\" (UniqueName: \"kubernetes.io/projected/5723ab46-57a5-4f9c-8391-cbb6e58fb651-kube-api-access-l94q6\") pod \"console-758964cdfb-8wtwd\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:37:52.235346 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.234962 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5723ab46-57a5-4f9c-8391-cbb6e58fb651-oauth-serving-cert\") pod \"console-758964cdfb-8wtwd\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:37:52.235346 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.235223 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5723ab46-57a5-4f9c-8391-cbb6e58fb651-console-serving-cert\") pod \"console-758964cdfb-8wtwd\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:37:52.235514 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.235359 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5723ab46-57a5-4f9c-8391-cbb6e58fb651-console-config\") pod \"console-758964cdfb-8wtwd\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:37:52.235514 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.235397 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5723ab46-57a5-4f9c-8391-cbb6e58fb651-trusted-ca-bundle\") pod \"console-758964cdfb-8wtwd\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:37:52.235514 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.235444 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5723ab46-57a5-4f9c-8391-cbb6e58fb651-console-oauth-config\") pod \"console-758964cdfb-8wtwd\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:37:52.235733 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.235680 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5723ab46-57a5-4f9c-8391-cbb6e58fb651-service-ca\") pod \"console-758964cdfb-8wtwd\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:37:52.235929 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.235909 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5723ab46-57a5-4f9c-8391-cbb6e58fb651-oauth-serving-cert\") pod \"console-758964cdfb-8wtwd\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:37:52.236608 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.236584 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5723ab46-57a5-4f9c-8391-cbb6e58fb651-console-config\") pod \"console-758964cdfb-8wtwd\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:37:52.236958 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.236939 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5723ab46-57a5-4f9c-8391-cbb6e58fb651-trusted-ca-bundle\") pod \"console-758964cdfb-8wtwd\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:37:52.237943 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.237924 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5723ab46-57a5-4f9c-8391-cbb6e58fb651-console-serving-cert\") pod \"console-758964cdfb-8wtwd\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:37:52.238252 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.238228 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5723ab46-57a5-4f9c-8391-cbb6e58fb651-console-oauth-config\") pod \"console-758964cdfb-8wtwd\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:37:52.242207 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.242185 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-l94q6\" (UniqueName: \"kubernetes.io/projected/5723ab46-57a5-4f9c-8391-cbb6e58fb651-kube-api-access-l94q6\") pod \"console-758964cdfb-8wtwd\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:37:52.396508 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.396419 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:37:52.517897 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.517845 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-758964cdfb-8wtwd"] Apr 22 17:37:52.521088 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:37:52.521058 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5723ab46_57a5_4f9c_8391_cbb6e58fb651.slice/crio-d258a8a92bd1da3731f31bc7c55cda6fac47cdd68ef04ec0c3fc36c2729edbde WatchSource:0}: Error finding container d258a8a92bd1da3731f31bc7c55cda6fac47cdd68ef04ec0c3fc36c2729edbde: Status 404 returned error can't find the container with id d258a8a92bd1da3731f31bc7c55cda6fac47cdd68ef04ec0c3fc36c2729edbde Apr 22 17:37:52.912698 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.912660 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/node-exporter-sf6bh"] Apr 22 17:37:52.917814 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.917785 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:52.920731 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.920702 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-dockercfg-8rhq9\"" Apr 22 17:37:52.920979 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.920961 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"node-exporter-accelerators-collector-config\"" Apr 22 17:37:52.921157 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.921144 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-tls\"" Apr 22 17:37:52.921449 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:52.921426 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-kube-rbac-proxy-config\"" Apr 22 17:37:53.041017 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.040942 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/b9de5d49-8aba-4723-8f93-a5cf469b3ea8-node-exporter-wtmp\") pod \"node-exporter-sf6bh\" (UID: \"b9de5d49-8aba-4723-8f93-a5cf469b3ea8\") " pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.041199 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.041064 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/b9de5d49-8aba-4723-8f93-a5cf469b3ea8-node-exporter-accelerators-collector-config\") pod \"node-exporter-sf6bh\" (UID: \"b9de5d49-8aba-4723-8f93-a5cf469b3ea8\") " pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.041199 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.041097 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/b9de5d49-8aba-4723-8f93-a5cf469b3ea8-root\") pod \"node-exporter-sf6bh\" (UID: \"b9de5d49-8aba-4723-8f93-a5cf469b3ea8\") " pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.041312 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.041230 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b9de5d49-8aba-4723-8f93-a5cf469b3ea8-sys\") pod \"node-exporter-sf6bh\" (UID: \"b9de5d49-8aba-4723-8f93-a5cf469b3ea8\") " pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.041312 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.041303 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/b9de5d49-8aba-4723-8f93-a5cf469b3ea8-node-exporter-textfile\") pod \"node-exporter-sf6bh\" (UID: \"b9de5d49-8aba-4723-8f93-a5cf469b3ea8\") " pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.041409 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.041330 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/b9de5d49-8aba-4723-8f93-a5cf469b3ea8-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-sf6bh\" (UID: \"b9de5d49-8aba-4723-8f93-a5cf469b3ea8\") " pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.041409 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.041363 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/b9de5d49-8aba-4723-8f93-a5cf469b3ea8-node-exporter-tls\") pod \"node-exporter-sf6bh\" (UID: \"b9de5d49-8aba-4723-8f93-a5cf469b3ea8\") " pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.041409 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.041399 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b9de5d49-8aba-4723-8f93-a5cf469b3ea8-metrics-client-ca\") pod \"node-exporter-sf6bh\" (UID: \"b9de5d49-8aba-4723-8f93-a5cf469b3ea8\") " pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.041555 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.041425 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnjvd\" (UniqueName: \"kubernetes.io/projected/b9de5d49-8aba-4723-8f93-a5cf469b3ea8-kube-api-access-gnjvd\") pod \"node-exporter-sf6bh\" (UID: \"b9de5d49-8aba-4723-8f93-a5cf469b3ea8\") " pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.142580 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.142541 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b9de5d49-8aba-4723-8f93-a5cf469b3ea8-sys\") pod \"node-exporter-sf6bh\" (UID: \"b9de5d49-8aba-4723-8f93-a5cf469b3ea8\") " pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.142752 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.142608 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/b9de5d49-8aba-4723-8f93-a5cf469b3ea8-node-exporter-textfile\") pod \"node-exporter-sf6bh\" (UID: \"b9de5d49-8aba-4723-8f93-a5cf469b3ea8\") " pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.142752 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.142636 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/b9de5d49-8aba-4723-8f93-a5cf469b3ea8-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-sf6bh\" (UID: \"b9de5d49-8aba-4723-8f93-a5cf469b3ea8\") " pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.142752 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.142670 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/b9de5d49-8aba-4723-8f93-a5cf469b3ea8-node-exporter-tls\") pod \"node-exporter-sf6bh\" (UID: \"b9de5d49-8aba-4723-8f93-a5cf469b3ea8\") " pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.142752 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.142710 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b9de5d49-8aba-4723-8f93-a5cf469b3ea8-metrics-client-ca\") pod \"node-exporter-sf6bh\" (UID: \"b9de5d49-8aba-4723-8f93-a5cf469b3ea8\") " pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.142752 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.142711 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b9de5d49-8aba-4723-8f93-a5cf469b3ea8-sys\") pod \"node-exporter-sf6bh\" (UID: \"b9de5d49-8aba-4723-8f93-a5cf469b3ea8\") " pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.142752 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.142737 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gnjvd\" (UniqueName: \"kubernetes.io/projected/b9de5d49-8aba-4723-8f93-a5cf469b3ea8-kube-api-access-gnjvd\") pod \"node-exporter-sf6bh\" (UID: \"b9de5d49-8aba-4723-8f93-a5cf469b3ea8\") " pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.143064 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.142822 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/b9de5d49-8aba-4723-8f93-a5cf469b3ea8-node-exporter-wtmp\") pod \"node-exporter-sf6bh\" (UID: \"b9de5d49-8aba-4723-8f93-a5cf469b3ea8\") " pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.143064 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.142853 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/b9de5d49-8aba-4723-8f93-a5cf469b3ea8-node-exporter-accelerators-collector-config\") pod \"node-exporter-sf6bh\" (UID: \"b9de5d49-8aba-4723-8f93-a5cf469b3ea8\") " pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.143064 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.142907 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/b9de5d49-8aba-4723-8f93-a5cf469b3ea8-root\") pod \"node-exporter-sf6bh\" (UID: \"b9de5d49-8aba-4723-8f93-a5cf469b3ea8\") " pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.143064 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.143026 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/b9de5d49-8aba-4723-8f93-a5cf469b3ea8-root\") pod \"node-exporter-sf6bh\" (UID: \"b9de5d49-8aba-4723-8f93-a5cf469b3ea8\") " pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.143252 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.143151 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/b9de5d49-8aba-4723-8f93-a5cf469b3ea8-node-exporter-wtmp\") pod \"node-exporter-sf6bh\" (UID: \"b9de5d49-8aba-4723-8f93-a5cf469b3ea8\") " pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.143411 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.143388 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/b9de5d49-8aba-4723-8f93-a5cf469b3ea8-node-exporter-textfile\") pod \"node-exporter-sf6bh\" (UID: \"b9de5d49-8aba-4723-8f93-a5cf469b3ea8\") " pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.143686 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.143660 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/b9de5d49-8aba-4723-8f93-a5cf469b3ea8-node-exporter-accelerators-collector-config\") pod \"node-exporter-sf6bh\" (UID: \"b9de5d49-8aba-4723-8f93-a5cf469b3ea8\") " pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.144327 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.144308 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b9de5d49-8aba-4723-8f93-a5cf469b3ea8-metrics-client-ca\") pod \"node-exporter-sf6bh\" (UID: \"b9de5d49-8aba-4723-8f93-a5cf469b3ea8\") " pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.149723 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.149694 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/b9de5d49-8aba-4723-8f93-a5cf469b3ea8-node-exporter-tls\") pod \"node-exporter-sf6bh\" (UID: \"b9de5d49-8aba-4723-8f93-a5cf469b3ea8\") " pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.156781 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.156752 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/b9de5d49-8aba-4723-8f93-a5cf469b3ea8-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-sf6bh\" (UID: \"b9de5d49-8aba-4723-8f93-a5cf469b3ea8\") " pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.157920 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.157858 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnjvd\" (UniqueName: \"kubernetes.io/projected/b9de5d49-8aba-4723-8f93-a5cf469b3ea8-kube-api-access-gnjvd\") pod \"node-exporter-sf6bh\" (UID: \"b9de5d49-8aba-4723-8f93-a5cf469b3ea8\") " pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.168256 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.168173 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-758964cdfb-8wtwd" event={"ID":"5723ab46-57a5-4f9c-8391-cbb6e58fb651","Type":"ContainerStarted","Data":"d258a8a92bd1da3731f31bc7c55cda6fac47cdd68ef04ec0c3fc36c2729edbde"} Apr 22 17:37:53.230859 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:53.230822 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-sf6bh" Apr 22 17:37:53.241163 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:37:53.241116 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9de5d49_8aba_4723_8f93_a5cf469b3ea8.slice/crio-5f5b56c3d1bc1630f99b3b6d39d4287798ba9427316e253254ef894baf25e23e WatchSource:0}: Error finding container 5f5b56c3d1bc1630f99b3b6d39d4287798ba9427316e253254ef894baf25e23e: Status 404 returned error can't find the container with id 5f5b56c3d1bc1630f99b3b6d39d4287798ba9427316e253254ef894baf25e23e Apr 22 17:37:54.178481 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:54.178180 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-sf6bh" event={"ID":"b9de5d49-8aba-4723-8f93-a5cf469b3ea8","Type":"ContainerStarted","Data":"5f5b56c3d1bc1630f99b3b6d39d4287798ba9427316e253254ef894baf25e23e"} Apr 22 17:37:55.182959 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:55.182913 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-sf6bh" event={"ID":"b9de5d49-8aba-4723-8f93-a5cf469b3ea8","Type":"ContainerStarted","Data":"eca006c98b9da254c737367644c6cc30ada0e1d995124e8e67893ac253b0801e"} Apr 22 17:37:56.186811 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:56.186772 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-758964cdfb-8wtwd" event={"ID":"5723ab46-57a5-4f9c-8391-cbb6e58fb651","Type":"ContainerStarted","Data":"3108944edd58e9918c18af817ec69f1e8da6927bdfb1771c30f195f7f7523bc6"} Apr 22 17:37:56.188210 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:56.188183 2581 generic.go:358] "Generic (PLEG): container finished" podID="b9de5d49-8aba-4723-8f93-a5cf469b3ea8" containerID="eca006c98b9da254c737367644c6cc30ada0e1d995124e8e67893ac253b0801e" exitCode=0 Apr 22 17:37:56.188322 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:56.188221 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-sf6bh" event={"ID":"b9de5d49-8aba-4723-8f93-a5cf469b3ea8","Type":"ContainerDied","Data":"eca006c98b9da254c737367644c6cc30ada0e1d995124e8e67893ac253b0801e"} Apr 22 17:37:56.206031 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:56.205984 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-758964cdfb-8wtwd" podStartSLOduration=1.382295151 podStartE2EDuration="4.205967653s" podCreationTimestamp="2026-04-22 17:37:52 +0000 UTC" firstStartedPulling="2026-04-22 17:37:52.523025437 +0000 UTC m=+180.415721139" lastFinishedPulling="2026-04-22 17:37:55.346697931 +0000 UTC m=+183.239393641" observedRunningTime="2026-04-22 17:37:56.204821768 +0000 UTC m=+184.097517528" watchObservedRunningTime="2026-04-22 17:37:56.205967653 +0000 UTC m=+184.098663376" Apr 22 17:37:57.192938 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:57.192897 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-sf6bh" event={"ID":"b9de5d49-8aba-4723-8f93-a5cf469b3ea8","Type":"ContainerStarted","Data":"65274480f4746df090da8c960115bc08175eacd99a663f15c33f509a490644d6"} Apr 22 17:37:57.193405 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:57.192941 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-sf6bh" event={"ID":"b9de5d49-8aba-4723-8f93-a5cf469b3ea8","Type":"ContainerStarted","Data":"6b166b7165622969ebd5a4dea29c6278937cfaf248974731e9535c0e6a24807b"} Apr 22 17:37:57.212385 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:37:57.212331 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/node-exporter-sf6bh" podStartSLOduration=4.3293663989999995 podStartE2EDuration="5.212314575s" podCreationTimestamp="2026-04-22 17:37:52 +0000 UTC" firstStartedPulling="2026-04-22 17:37:53.243438038 +0000 UTC m=+181.136133740" lastFinishedPulling="2026-04-22 17:37:54.126386204 +0000 UTC m=+182.019081916" observedRunningTime="2026-04-22 17:37:57.211454262 +0000 UTC m=+185.104149996" watchObservedRunningTime="2026-04-22 17:37:57.212314575 +0000 UTC m=+185.105010298" Apr 22 17:38:02.396672 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:02.396630 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:38:02.397147 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:02.396743 2581 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:38:02.402756 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:02.402731 2581 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:38:02.690080 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:02.689998 2581 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-758964cdfb-8wtwd"] Apr 22 17:38:03.212240 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:03.212206 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:38:04.138960 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:04.138933 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-69fdd76d7d-jtkk2" Apr 22 17:38:29.229479 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:29.229416 2581 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-console/console-758964cdfb-8wtwd" podUID="5723ab46-57a5-4f9c-8391-cbb6e58fb651" containerName="console" containerID="cri-o://3108944edd58e9918c18af817ec69f1e8da6927bdfb1771c30f195f7f7523bc6" gracePeriod=15 Apr 22 17:38:29.459954 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:29.459932 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-758964cdfb-8wtwd_5723ab46-57a5-4f9c-8391-cbb6e58fb651/console/0.log" Apr 22 17:38:29.460083 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:29.460004 2581 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:38:29.546489 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:29.546453 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5723ab46-57a5-4f9c-8391-cbb6e58fb651-oauth-serving-cert\") pod \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " Apr 22 17:38:29.546697 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:29.546517 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5723ab46-57a5-4f9c-8391-cbb6e58fb651-service-ca\") pod \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " Apr 22 17:38:29.546697 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:29.546538 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5723ab46-57a5-4f9c-8391-cbb6e58fb651-console-oauth-config\") pod \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " Apr 22 17:38:29.546697 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:29.546569 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l94q6\" (UniqueName: \"kubernetes.io/projected/5723ab46-57a5-4f9c-8391-cbb6e58fb651-kube-api-access-l94q6\") pod \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " Apr 22 17:38:29.546697 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:29.546590 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5723ab46-57a5-4f9c-8391-cbb6e58fb651-console-config\") pod \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " Apr 22 17:38:29.546697 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:29.546607 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5723ab46-57a5-4f9c-8391-cbb6e58fb651-trusted-ca-bundle\") pod \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " Apr 22 17:38:29.546697 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:29.546626 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5723ab46-57a5-4f9c-8391-cbb6e58fb651-console-serving-cert\") pod \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\" (UID: \"5723ab46-57a5-4f9c-8391-cbb6e58fb651\") " Apr 22 17:38:29.547092 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:29.547000 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5723ab46-57a5-4f9c-8391-cbb6e58fb651-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "5723ab46-57a5-4f9c-8391-cbb6e58fb651" (UID: "5723ab46-57a5-4f9c-8391-cbb6e58fb651"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 17:38:29.547092 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:29.547039 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5723ab46-57a5-4f9c-8391-cbb6e58fb651-console-config" (OuterVolumeSpecName: "console-config") pod "5723ab46-57a5-4f9c-8391-cbb6e58fb651" (UID: "5723ab46-57a5-4f9c-8391-cbb6e58fb651"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 17:38:29.547182 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:29.547123 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5723ab46-57a5-4f9c-8391-cbb6e58fb651-service-ca" (OuterVolumeSpecName: "service-ca") pod "5723ab46-57a5-4f9c-8391-cbb6e58fb651" (UID: "5723ab46-57a5-4f9c-8391-cbb6e58fb651"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 17:38:29.547227 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:29.547206 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5723ab46-57a5-4f9c-8391-cbb6e58fb651-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "5723ab46-57a5-4f9c-8391-cbb6e58fb651" (UID: "5723ab46-57a5-4f9c-8391-cbb6e58fb651"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 22 17:38:29.549176 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:29.549149 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5723ab46-57a5-4f9c-8391-cbb6e58fb651-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "5723ab46-57a5-4f9c-8391-cbb6e58fb651" (UID: "5723ab46-57a5-4f9c-8391-cbb6e58fb651"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 17:38:29.549270 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:29.549193 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5723ab46-57a5-4f9c-8391-cbb6e58fb651-kube-api-access-l94q6" (OuterVolumeSpecName: "kube-api-access-l94q6") pod "5723ab46-57a5-4f9c-8391-cbb6e58fb651" (UID: "5723ab46-57a5-4f9c-8391-cbb6e58fb651"). InnerVolumeSpecName "kube-api-access-l94q6". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 22 17:38:29.549270 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:29.549208 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5723ab46-57a5-4f9c-8391-cbb6e58fb651-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "5723ab46-57a5-4f9c-8391-cbb6e58fb651" (UID: "5723ab46-57a5-4f9c-8391-cbb6e58fb651"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 22 17:38:29.648081 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:29.648025 2581 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-l94q6\" (UniqueName: \"kubernetes.io/projected/5723ab46-57a5-4f9c-8391-cbb6e58fb651-kube-api-access-l94q6\") on node \"ip-10-0-138-57.ec2.internal\" DevicePath \"\"" Apr 22 17:38:29.648081 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:29.648071 2581 reconciler_common.go:299] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5723ab46-57a5-4f9c-8391-cbb6e58fb651-console-config\") on node \"ip-10-0-138-57.ec2.internal\" DevicePath \"\"" Apr 22 17:38:29.648081 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:29.648081 2581 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5723ab46-57a5-4f9c-8391-cbb6e58fb651-trusted-ca-bundle\") on node \"ip-10-0-138-57.ec2.internal\" DevicePath \"\"" Apr 22 17:38:29.648081 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:29.648092 2581 reconciler_common.go:299] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5723ab46-57a5-4f9c-8391-cbb6e58fb651-console-serving-cert\") on node \"ip-10-0-138-57.ec2.internal\" DevicePath \"\"" Apr 22 17:38:29.648081 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:29.648101 2581 reconciler_common.go:299] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5723ab46-57a5-4f9c-8391-cbb6e58fb651-oauth-serving-cert\") on node \"ip-10-0-138-57.ec2.internal\" DevicePath \"\"" Apr 22 17:38:29.648385 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:29.648109 2581 reconciler_common.go:299] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5723ab46-57a5-4f9c-8391-cbb6e58fb651-service-ca\") on node \"ip-10-0-138-57.ec2.internal\" DevicePath \"\"" Apr 22 17:38:29.648385 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:29.648120 2581 reconciler_common.go:299] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5723ab46-57a5-4f9c-8391-cbb6e58fb651-console-oauth-config\") on node \"ip-10-0-138-57.ec2.internal\" DevicePath \"\"" Apr 22 17:38:30.279489 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:30.279459 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-758964cdfb-8wtwd_5723ab46-57a5-4f9c-8391-cbb6e58fb651/console/0.log" Apr 22 17:38:30.279972 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:30.279501 2581 generic.go:358] "Generic (PLEG): container finished" podID="5723ab46-57a5-4f9c-8391-cbb6e58fb651" containerID="3108944edd58e9918c18af817ec69f1e8da6927bdfb1771c30f195f7f7523bc6" exitCode=2 Apr 22 17:38:30.279972 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:30.279541 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-758964cdfb-8wtwd" event={"ID":"5723ab46-57a5-4f9c-8391-cbb6e58fb651","Type":"ContainerDied","Data":"3108944edd58e9918c18af817ec69f1e8da6927bdfb1771c30f195f7f7523bc6"} Apr 22 17:38:30.279972 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:30.279571 2581 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-758964cdfb-8wtwd" Apr 22 17:38:30.279972 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:30.279591 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-758964cdfb-8wtwd" event={"ID":"5723ab46-57a5-4f9c-8391-cbb6e58fb651","Type":"ContainerDied","Data":"d258a8a92bd1da3731f31bc7c55cda6fac47cdd68ef04ec0c3fc36c2729edbde"} Apr 22 17:38:30.279972 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:30.279614 2581 scope.go:117] "RemoveContainer" containerID="3108944edd58e9918c18af817ec69f1e8da6927bdfb1771c30f195f7f7523bc6" Apr 22 17:38:30.287783 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:30.287764 2581 scope.go:117] "RemoveContainer" containerID="3108944edd58e9918c18af817ec69f1e8da6927bdfb1771c30f195f7f7523bc6" Apr 22 17:38:30.288094 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:38:30.288073 2581 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3108944edd58e9918c18af817ec69f1e8da6927bdfb1771c30f195f7f7523bc6\": container with ID starting with 3108944edd58e9918c18af817ec69f1e8da6927bdfb1771c30f195f7f7523bc6 not found: ID does not exist" containerID="3108944edd58e9918c18af817ec69f1e8da6927bdfb1771c30f195f7f7523bc6" Apr 22 17:38:30.288180 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:30.288106 2581 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3108944edd58e9918c18af817ec69f1e8da6927bdfb1771c30f195f7f7523bc6"} err="failed to get container status \"3108944edd58e9918c18af817ec69f1e8da6927bdfb1771c30f195f7f7523bc6\": rpc error: code = NotFound desc = could not find container \"3108944edd58e9918c18af817ec69f1e8da6927bdfb1771c30f195f7f7523bc6\": container with ID starting with 3108944edd58e9918c18af817ec69f1e8da6927bdfb1771c30f195f7f7523bc6 not found: ID does not exist" Apr 22 17:38:30.300496 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:30.300468 2581 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-console/console-758964cdfb-8wtwd"] Apr 22 17:38:30.304033 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:30.303988 2581 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-758964cdfb-8wtwd"] Apr 22 17:38:30.639089 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:30.639006 2581 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5723ab46-57a5-4f9c-8391-cbb6e58fb651" path="/var/lib/kubelet/pods/5723ab46-57a5-4f9c-8391-cbb6e58fb651/volumes" Apr 22 17:38:41.433535 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:41.433466 2581 prober.go:120] "Probe failed" probeType="Liveness" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" podUID="0c337636-6259-4b1c-a1b7-e948b21c03d6" containerName="service-proxy" probeResult="failure" output="HTTP probe failed with statuscode: 500" Apr 22 17:38:42.315513 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:42.315479 2581 generic.go:358] "Generic (PLEG): container finished" podID="18d358c9-5de4-4059-98a5-f3f648ff3d67" containerID="043e068f893e5a2e2880e20cd24ae29f793dce36bd0c13b823249ebbc6b53b28" exitCode=0 Apr 22 17:38:42.315682 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:42.315549 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-jf5fx" event={"ID":"18d358c9-5de4-4059-98a5-f3f648ff3d67","Type":"ContainerDied","Data":"043e068f893e5a2e2880e20cd24ae29f793dce36bd0c13b823249ebbc6b53b28"} Apr 22 17:38:42.315955 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:42.315935 2581 scope.go:117] "RemoveContainer" containerID="043e068f893e5a2e2880e20cd24ae29f793dce36bd0c13b823249ebbc6b53b28" Apr 22 17:38:43.328680 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:43.328649 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-jf5fx" event={"ID":"18d358c9-5de4-4059-98a5-f3f648ff3d67","Type":"ContainerStarted","Data":"34433f95e7e067de5d65f43a138ab3ccc93d87151c5053d21dde308b4220090a"} Apr 22 17:38:51.433246 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:38:51.433204 2581 prober.go:120] "Probe failed" probeType="Liveness" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" podUID="0c337636-6259-4b1c-a1b7-e948b21c03d6" containerName="service-proxy" probeResult="failure" output="HTTP probe failed with statuscode: 500" Apr 22 17:39:01.433231 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:01.433193 2581 prober.go:120] "Probe failed" probeType="Liveness" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" podUID="0c337636-6259-4b1c-a1b7-e948b21c03d6" containerName="service-proxy" probeResult="failure" output="HTTP probe failed with statuscode: 500" Apr 22 17:39:01.433589 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:01.433270 2581 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" Apr 22 17:39:01.433711 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:01.433696 2581 kuberuntime_manager.go:1107] "Message for Container of pod" containerName="service-proxy" containerStatusID={"Type":"cri-o","ID":"fd42bfaaed32d6e4b8c9b0c0c7b65fb007826baa9fb99ebf41b9b7f68e9184c7"} pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" containerMessage="Container service-proxy failed liveness probe, will be restarted" Apr 22 17:39:01.433748 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:01.433730 2581 kuberuntime_container.go:864] "Killing container with a grace period" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" podUID="0c337636-6259-4b1c-a1b7-e948b21c03d6" containerName="service-proxy" containerID="cri-o://fd42bfaaed32d6e4b8c9b0c0c7b65fb007826baa9fb99ebf41b9b7f68e9184c7" gracePeriod=30 Apr 22 17:39:02.378719 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:02.378688 2581 generic.go:358] "Generic (PLEG): container finished" podID="0c337636-6259-4b1c-a1b7-e948b21c03d6" containerID="fd42bfaaed32d6e4b8c9b0c0c7b65fb007826baa9fb99ebf41b9b7f68e9184c7" exitCode=2 Apr 22 17:39:02.378903 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:02.378760 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" event={"ID":"0c337636-6259-4b1c-a1b7-e948b21c03d6","Type":"ContainerDied","Data":"fd42bfaaed32d6e4b8c9b0c0c7b65fb007826baa9fb99ebf41b9b7f68e9184c7"} Apr 22 17:39:02.378903 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:02.378798 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="open-cluster-management-agent-addon/cluster-proxy-proxy-agent-5ccb5fcc49-d96zp" event={"ID":"0c337636-6259-4b1c-a1b7-e948b21c03d6","Type":"ContainerStarted","Data":"ef75ddc56cf813f758095ee802c6c8da32178061dc20fbf5ceb68f38091809df"} Apr 22 17:39:04.402666 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:04.402575 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs\") pod \"network-metrics-daemon-zhl6m\" (UID: \"cff34b01-53b3-4374-b73d-d3146e3c558d\") " pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:39:04.405015 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:04.404994 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cff34b01-53b3-4374-b73d-d3146e3c558d-metrics-certs\") pod \"network-metrics-daemon-zhl6m\" (UID: \"cff34b01-53b3-4374-b73d-d3146e3c558d\") " pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:39:04.638540 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:04.638507 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-qs9jr\"" Apr 22 17:39:04.646412 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:04.646391 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-zhl6m" Apr 22 17:39:04.762005 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:04.761966 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-zhl6m"] Apr 22 17:39:04.766494 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:39:04.766465 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcff34b01_53b3_4374_b73d_d3146e3c558d.slice/crio-a1b8b9d796cea1832b1f39a18585cf94091c2a3e73a4fd098c5df275252ecec9 WatchSource:0}: Error finding container a1b8b9d796cea1832b1f39a18585cf94091c2a3e73a4fd098c5df275252ecec9: Status 404 returned error can't find the container with id a1b8b9d796cea1832b1f39a18585cf94091c2a3e73a4fd098c5df275252ecec9 Apr 22 17:39:05.387792 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:05.387746 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zhl6m" event={"ID":"cff34b01-53b3-4374-b73d-d3146e3c558d","Type":"ContainerStarted","Data":"a1b8b9d796cea1832b1f39a18585cf94091c2a3e73a4fd098c5df275252ecec9"} Apr 22 17:39:06.394384 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:06.394349 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zhl6m" event={"ID":"cff34b01-53b3-4374-b73d-d3146e3c558d","Type":"ContainerStarted","Data":"c8f546f7af47381f29027ddbc7877e829c8e013cec04b0dbd4165bb2b180d667"} Apr 22 17:39:06.394384 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:06.394386 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-zhl6m" event={"ID":"cff34b01-53b3-4374-b73d-d3146e3c558d","Type":"ContainerStarted","Data":"edf45d73a42b1c2dfa4a8c9448f94a48538efdab8879b2a1dbf1ee058cbd6a11"} Apr 22 17:39:06.413319 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:06.413276 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-zhl6m" podStartSLOduration=253.481590516 podStartE2EDuration="4m14.413263318s" podCreationTimestamp="2026-04-22 17:34:52 +0000 UTC" firstStartedPulling="2026-04-22 17:39:04.76907105 +0000 UTC m=+252.661766754" lastFinishedPulling="2026-04-22 17:39:05.700743843 +0000 UTC m=+253.593439556" observedRunningTime="2026-04-22 17:39:06.41178084 +0000 UTC m=+254.304476563" watchObservedRunningTime="2026-04-22 17:39:06.413263318 +0000 UTC m=+254.305959075" Apr 22 17:39:48.179512 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:48.179477 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kube-system/global-pull-secret-syncer-drtgm"] Apr 22 17:39:48.180043 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:48.179827 2581 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="5723ab46-57a5-4f9c-8391-cbb6e58fb651" containerName="console" Apr 22 17:39:48.180043 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:48.179845 2581 state_mem.go:107] "Deleted CPUSet assignment" podUID="5723ab46-57a5-4f9c-8391-cbb6e58fb651" containerName="console" Apr 22 17:39:48.180043 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:48.179927 2581 memory_manager.go:356] "RemoveStaleState removing state" podUID="5723ab46-57a5-4f9c-8391-cbb6e58fb651" containerName="console" Apr 22 17:39:48.184081 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:48.184064 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-drtgm" Apr 22 17:39:48.186076 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:48.186056 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"original-pull-secret\"" Apr 22 17:39:48.190023 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:48.190000 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kube-system/global-pull-secret-syncer-drtgm"] Apr 22 17:39:48.290735 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:48.290701 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/8ab355fc-d1b4-424a-94f9-5d7cad6d032f-original-pull-secret\") pod \"global-pull-secret-syncer-drtgm\" (UID: \"8ab355fc-d1b4-424a-94f9-5d7cad6d032f\") " pod="kube-system/global-pull-secret-syncer-drtgm" Apr 22 17:39:48.290943 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:48.290746 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/8ab355fc-d1b4-424a-94f9-5d7cad6d032f-kubelet-config\") pod \"global-pull-secret-syncer-drtgm\" (UID: \"8ab355fc-d1b4-424a-94f9-5d7cad6d032f\") " pod="kube-system/global-pull-secret-syncer-drtgm" Apr 22 17:39:48.290943 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:48.290826 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/8ab355fc-d1b4-424a-94f9-5d7cad6d032f-dbus\") pod \"global-pull-secret-syncer-drtgm\" (UID: \"8ab355fc-d1b4-424a-94f9-5d7cad6d032f\") " pod="kube-system/global-pull-secret-syncer-drtgm" Apr 22 17:39:48.391959 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:48.391927 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/8ab355fc-d1b4-424a-94f9-5d7cad6d032f-kubelet-config\") pod \"global-pull-secret-syncer-drtgm\" (UID: \"8ab355fc-d1b4-424a-94f9-5d7cad6d032f\") " pod="kube-system/global-pull-secret-syncer-drtgm" Apr 22 17:39:48.392094 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:48.391978 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/8ab355fc-d1b4-424a-94f9-5d7cad6d032f-dbus\") pod \"global-pull-secret-syncer-drtgm\" (UID: \"8ab355fc-d1b4-424a-94f9-5d7cad6d032f\") " pod="kube-system/global-pull-secret-syncer-drtgm" Apr 22 17:39:48.392094 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:48.392012 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/8ab355fc-d1b4-424a-94f9-5d7cad6d032f-original-pull-secret\") pod \"global-pull-secret-syncer-drtgm\" (UID: \"8ab355fc-d1b4-424a-94f9-5d7cad6d032f\") " pod="kube-system/global-pull-secret-syncer-drtgm" Apr 22 17:39:48.392094 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:48.392063 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/8ab355fc-d1b4-424a-94f9-5d7cad6d032f-kubelet-config\") pod \"global-pull-secret-syncer-drtgm\" (UID: \"8ab355fc-d1b4-424a-94f9-5d7cad6d032f\") " pod="kube-system/global-pull-secret-syncer-drtgm" Apr 22 17:39:48.392217 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:48.392197 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/8ab355fc-d1b4-424a-94f9-5d7cad6d032f-dbus\") pod \"global-pull-secret-syncer-drtgm\" (UID: \"8ab355fc-d1b4-424a-94f9-5d7cad6d032f\") " pod="kube-system/global-pull-secret-syncer-drtgm" Apr 22 17:39:48.394416 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:48.394397 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/8ab355fc-d1b4-424a-94f9-5d7cad6d032f-original-pull-secret\") pod \"global-pull-secret-syncer-drtgm\" (UID: \"8ab355fc-d1b4-424a-94f9-5d7cad6d032f\") " pod="kube-system/global-pull-secret-syncer-drtgm" Apr 22 17:39:48.493464 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:48.493390 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-drtgm" Apr 22 17:39:48.610064 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:48.610034 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kube-system/global-pull-secret-syncer-drtgm"] Apr 22 17:39:48.613170 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:39:48.613139 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ab355fc_d1b4_424a_94f9_5d7cad6d032f.slice/crio-f61f6436bbf86ac8618bd401c5df9fb9f1cd124ae6b91167f92519fec02925a3 WatchSource:0}: Error finding container f61f6436bbf86ac8618bd401c5df9fb9f1cd124ae6b91167f92519fec02925a3: Status 404 returned error can't find the container with id f61f6436bbf86ac8618bd401c5df9fb9f1cd124ae6b91167f92519fec02925a3 Apr 22 17:39:49.512619 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:49.512582 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/global-pull-secret-syncer-drtgm" event={"ID":"8ab355fc-d1b4-424a-94f9-5d7cad6d032f","Type":"ContainerStarted","Data":"f61f6436bbf86ac8618bd401c5df9fb9f1cd124ae6b91167f92519fec02925a3"} Apr 22 17:39:52.547967 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:52.547855 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 17:39:52.548952 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:52.548893 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 17:39:52.550515 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:52.550492 2581 kubelet.go:1628] "Image garbage collection succeeded" Apr 22 17:39:53.523607 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:53.523567 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/global-pull-secret-syncer-drtgm" event={"ID":"8ab355fc-d1b4-424a-94f9-5d7cad6d032f","Type":"ContainerStarted","Data":"3dc6bfbdcb88cfdd95e4f9a6aa8e0d8cf1d1246b0fe0b013426729c8958f0e20"} Apr 22 17:39:53.539461 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:39:53.539413 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/global-pull-secret-syncer-drtgm" podStartSLOduration=1.6490585 podStartE2EDuration="5.539398462s" podCreationTimestamp="2026-04-22 17:39:48 +0000 UTC" firstStartedPulling="2026-04-22 17:39:48.614682521 +0000 UTC m=+296.507378223" lastFinishedPulling="2026-04-22 17:39:52.505022481 +0000 UTC m=+300.397718185" observedRunningTime="2026-04-22 17:39:53.53866367 +0000 UTC m=+301.431359406" watchObservedRunningTime="2026-04-22 17:39:53.539398462 +0000 UTC m=+301.432094186" Apr 22 17:40:10.626110 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:10.626072 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm"] Apr 22 17:40:10.629216 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:10.629198 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm" Apr 22 17:40:10.631316 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:10.631292 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"kube-root-ca.crt\"" Apr 22 17:40:10.631913 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:10.631891 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"openshift-service-ca.crt\"" Apr 22 17:40:10.631913 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:10.631901 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"default-dockercfg-kv2np\"" Apr 22 17:40:10.640012 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:10.639991 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm"] Apr 22 17:40:10.760654 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:10.760612 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72llm\" (UniqueName: \"kubernetes.io/projected/ec0b5dff-4c93-4293-8d19-c70164d2b0da-kube-api-access-72llm\") pod \"59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm\" (UID: \"ec0b5dff-4c93-4293-8d19-c70164d2b0da\") " pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm" Apr 22 17:40:10.760845 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:10.760690 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ec0b5dff-4c93-4293-8d19-c70164d2b0da-bundle\") pod \"59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm\" (UID: \"ec0b5dff-4c93-4293-8d19-c70164d2b0da\") " pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm" Apr 22 17:40:10.760845 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:10.760724 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ec0b5dff-4c93-4293-8d19-c70164d2b0da-util\") pod \"59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm\" (UID: \"ec0b5dff-4c93-4293-8d19-c70164d2b0da\") " pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm" Apr 22 17:40:10.861305 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:10.861252 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ec0b5dff-4c93-4293-8d19-c70164d2b0da-bundle\") pod \"59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm\" (UID: \"ec0b5dff-4c93-4293-8d19-c70164d2b0da\") " pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm" Apr 22 17:40:10.861476 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:10.861333 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ec0b5dff-4c93-4293-8d19-c70164d2b0da-util\") pod \"59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm\" (UID: \"ec0b5dff-4c93-4293-8d19-c70164d2b0da\") " pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm" Apr 22 17:40:10.861476 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:10.861365 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-72llm\" (UniqueName: \"kubernetes.io/projected/ec0b5dff-4c93-4293-8d19-c70164d2b0da-kube-api-access-72llm\") pod \"59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm\" (UID: \"ec0b5dff-4c93-4293-8d19-c70164d2b0da\") " pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm" Apr 22 17:40:10.861642 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:10.861622 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ec0b5dff-4c93-4293-8d19-c70164d2b0da-bundle\") pod \"59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm\" (UID: \"ec0b5dff-4c93-4293-8d19-c70164d2b0da\") " pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm" Apr 22 17:40:10.861740 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:10.861719 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ec0b5dff-4c93-4293-8d19-c70164d2b0da-util\") pod \"59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm\" (UID: \"ec0b5dff-4c93-4293-8d19-c70164d2b0da\") " pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm" Apr 22 17:40:10.870199 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:10.870163 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-72llm\" (UniqueName: \"kubernetes.io/projected/ec0b5dff-4c93-4293-8d19-c70164d2b0da-kube-api-access-72llm\") pod \"59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm\" (UID: \"ec0b5dff-4c93-4293-8d19-c70164d2b0da\") " pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm" Apr 22 17:40:10.939189 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:10.939088 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm" Apr 22 17:40:11.061853 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:11.061814 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm"] Apr 22 17:40:11.065031 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:40:11.065001 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podec0b5dff_4c93_4293_8d19_c70164d2b0da.slice/crio-c1f8f785883d52d59ac0b6ba2f31b452671a9ac1b5c627d8632cdffb2c555688 WatchSource:0}: Error finding container c1f8f785883d52d59ac0b6ba2f31b452671a9ac1b5c627d8632cdffb2c555688: Status 404 returned error can't find the container with id c1f8f785883d52d59ac0b6ba2f31b452671a9ac1b5c627d8632cdffb2c555688 Apr 22 17:40:11.066993 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:11.066978 2581 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 22 17:40:11.578276 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:11.578236 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm" event={"ID":"ec0b5dff-4c93-4293-8d19-c70164d2b0da","Type":"ContainerStarted","Data":"c1f8f785883d52d59ac0b6ba2f31b452671a9ac1b5c627d8632cdffb2c555688"} Apr 22 17:40:16.593732 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:16.593696 2581 generic.go:358] "Generic (PLEG): container finished" podID="ec0b5dff-4c93-4293-8d19-c70164d2b0da" containerID="e34e6d55484fd0f878a3b5775aec6ef7dc95d6db2b498a8881e6ec0ad2747609" exitCode=0 Apr 22 17:40:16.594097 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:16.593748 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm" event={"ID":"ec0b5dff-4c93-4293-8d19-c70164d2b0da","Type":"ContainerDied","Data":"e34e6d55484fd0f878a3b5775aec6ef7dc95d6db2b498a8881e6ec0ad2747609"} Apr 22 17:40:18.601335 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:18.601305 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm" event={"ID":"ec0b5dff-4c93-4293-8d19-c70164d2b0da","Type":"ContainerStarted","Data":"905baccd6692315ff0408de6a85ca85db13a7b18ef4cbcb3bf3f521243b15d58"} Apr 22 17:40:19.605929 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:19.605869 2581 generic.go:358] "Generic (PLEG): container finished" podID="ec0b5dff-4c93-4293-8d19-c70164d2b0da" containerID="905baccd6692315ff0408de6a85ca85db13a7b18ef4cbcb3bf3f521243b15d58" exitCode=0 Apr 22 17:40:19.605929 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:19.605917 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm" event={"ID":"ec0b5dff-4c93-4293-8d19-c70164d2b0da","Type":"ContainerDied","Data":"905baccd6692315ff0408de6a85ca85db13a7b18ef4cbcb3bf3f521243b15d58"} Apr 22 17:40:26.626268 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:26.626236 2581 generic.go:358] "Generic (PLEG): container finished" podID="ec0b5dff-4c93-4293-8d19-c70164d2b0da" containerID="f4d5c3fcbbd32cfc6adefb8c2293c1140fb975bf10a6ec74df0a99c5b3ea13b3" exitCode=0 Apr 22 17:40:26.626646 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:26.626328 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm" event={"ID":"ec0b5dff-4c93-4293-8d19-c70164d2b0da","Type":"ContainerDied","Data":"f4d5c3fcbbd32cfc6adefb8c2293c1140fb975bf10a6ec74df0a99c5b3ea13b3"} Apr 22 17:40:27.744246 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:27.744228 2581 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm" Apr 22 17:40:27.790128 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:27.790101 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ec0b5dff-4c93-4293-8d19-c70164d2b0da-bundle\") pod \"ec0b5dff-4c93-4293-8d19-c70164d2b0da\" (UID: \"ec0b5dff-4c93-4293-8d19-c70164d2b0da\") " Apr 22 17:40:27.790247 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:27.790173 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72llm\" (UniqueName: \"kubernetes.io/projected/ec0b5dff-4c93-4293-8d19-c70164d2b0da-kube-api-access-72llm\") pod \"ec0b5dff-4c93-4293-8d19-c70164d2b0da\" (UID: \"ec0b5dff-4c93-4293-8d19-c70164d2b0da\") " Apr 22 17:40:27.790247 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:27.790213 2581 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ec0b5dff-4c93-4293-8d19-c70164d2b0da-util\") pod \"ec0b5dff-4c93-4293-8d19-c70164d2b0da\" (UID: \"ec0b5dff-4c93-4293-8d19-c70164d2b0da\") " Apr 22 17:40:27.790682 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:27.790656 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec0b5dff-4c93-4293-8d19-c70164d2b0da-bundle" (OuterVolumeSpecName: "bundle") pod "ec0b5dff-4c93-4293-8d19-c70164d2b0da" (UID: "ec0b5dff-4c93-4293-8d19-c70164d2b0da"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 22 17:40:27.792450 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:27.792428 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec0b5dff-4c93-4293-8d19-c70164d2b0da-kube-api-access-72llm" (OuterVolumeSpecName: "kube-api-access-72llm") pod "ec0b5dff-4c93-4293-8d19-c70164d2b0da" (UID: "ec0b5dff-4c93-4293-8d19-c70164d2b0da"). InnerVolumeSpecName "kube-api-access-72llm". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 22 17:40:27.794966 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:27.794946 2581 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec0b5dff-4c93-4293-8d19-c70164d2b0da-util" (OuterVolumeSpecName: "util") pod "ec0b5dff-4c93-4293-8d19-c70164d2b0da" (UID: "ec0b5dff-4c93-4293-8d19-c70164d2b0da"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 22 17:40:27.891430 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:27.891359 2581 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-72llm\" (UniqueName: \"kubernetes.io/projected/ec0b5dff-4c93-4293-8d19-c70164d2b0da-kube-api-access-72llm\") on node \"ip-10-0-138-57.ec2.internal\" DevicePath \"\"" Apr 22 17:40:27.891430 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:27.891385 2581 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ec0b5dff-4c93-4293-8d19-c70164d2b0da-util\") on node \"ip-10-0-138-57.ec2.internal\" DevicePath \"\"" Apr 22 17:40:27.891430 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:27.891395 2581 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ec0b5dff-4c93-4293-8d19-c70164d2b0da-bundle\") on node \"ip-10-0-138-57.ec2.internal\" DevicePath \"\"" Apr 22 17:40:28.632551 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:28.632514 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm" event={"ID":"ec0b5dff-4c93-4293-8d19-c70164d2b0da","Type":"ContainerDied","Data":"c1f8f785883d52d59ac0b6ba2f31b452671a9ac1b5c627d8632cdffb2c555688"} Apr 22 17:40:28.632551 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:28.632553 2581 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1f8f785883d52d59ac0b6ba2f31b452671a9ac1b5c627d8632cdffb2c555688" Apr 22 17:40:28.632746 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:28.632571 2581 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/59039e319e11338a40c6b6f1054d265f40bb50ceac6068d5c59955d29cp8gkm" Apr 22 17:40:37.030158 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.030063 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-keda/keda-operator-ffbb595cb-hhwgx"] Apr 22 17:40:37.030498 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.030319 2581 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ec0b5dff-4c93-4293-8d19-c70164d2b0da" containerName="util" Apr 22 17:40:37.030498 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.030329 2581 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec0b5dff-4c93-4293-8d19-c70164d2b0da" containerName="util" Apr 22 17:40:37.030498 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.030337 2581 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ec0b5dff-4c93-4293-8d19-c70164d2b0da" containerName="pull" Apr 22 17:40:37.030498 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.030342 2581 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec0b5dff-4c93-4293-8d19-c70164d2b0da" containerName="pull" Apr 22 17:40:37.030498 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.030350 2581 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ec0b5dff-4c93-4293-8d19-c70164d2b0da" containerName="extract" Apr 22 17:40:37.030498 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.030356 2581 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec0b5dff-4c93-4293-8d19-c70164d2b0da" containerName="extract" Apr 22 17:40:37.030498 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.030407 2581 memory_manager.go:356] "RemoveStaleState removing state" podUID="ec0b5dff-4c93-4293-8d19-c70164d2b0da" containerName="extract" Apr 22 17:40:37.033078 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.033060 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-keda/keda-operator-ffbb595cb-hhwgx" Apr 22 17:40:37.036890 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.036839 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-keda\"/\"openshift-service-ca.crt\"" Apr 22 17:40:37.036890 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.036870 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-keda\"/\"keda-operator-certs\"" Apr 22 17:40:37.037133 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.036834 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-keda\"/\"kube-root-ca.crt\"" Apr 22 17:40:37.037590 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.037567 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-keda\"/\"kedaorg-certs\"" Apr 22 17:40:37.037709 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.037593 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-keda\"/\"keda-ocp-cabundle\"" Apr 22 17:40:37.037709 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.037568 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-keda\"/\"keda-operator-dockercfg-jc8ss\"" Apr 22 17:40:37.047272 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.047251 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-keda/keda-operator-ffbb595cb-hhwgx"] Apr 22 17:40:37.155952 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.155919 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/8ebdcb73-11e8-487c-b4ae-7b55b0644382-certificates\") pod \"keda-operator-ffbb595cb-hhwgx\" (UID: \"8ebdcb73-11e8-487c-b4ae-7b55b0644382\") " pod="openshift-keda/keda-operator-ffbb595cb-hhwgx" Apr 22 17:40:37.155952 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.155965 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rgd8\" (UniqueName: \"kubernetes.io/projected/8ebdcb73-11e8-487c-b4ae-7b55b0644382-kube-api-access-9rgd8\") pod \"keda-operator-ffbb595cb-hhwgx\" (UID: \"8ebdcb73-11e8-487c-b4ae-7b55b0644382\") " pod="openshift-keda/keda-operator-ffbb595cb-hhwgx" Apr 22 17:40:37.156169 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.155995 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cabundle0\" (UniqueName: \"kubernetes.io/configmap/8ebdcb73-11e8-487c-b4ae-7b55b0644382-cabundle0\") pod \"keda-operator-ffbb595cb-hhwgx\" (UID: \"8ebdcb73-11e8-487c-b4ae-7b55b0644382\") " pod="openshift-keda/keda-operator-ffbb595cb-hhwgx" Apr 22 17:40:37.256646 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.256604 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/8ebdcb73-11e8-487c-b4ae-7b55b0644382-certificates\") pod \"keda-operator-ffbb595cb-hhwgx\" (UID: \"8ebdcb73-11e8-487c-b4ae-7b55b0644382\") " pod="openshift-keda/keda-operator-ffbb595cb-hhwgx" Apr 22 17:40:37.256941 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.256657 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-9rgd8\" (UniqueName: \"kubernetes.io/projected/8ebdcb73-11e8-487c-b4ae-7b55b0644382-kube-api-access-9rgd8\") pod \"keda-operator-ffbb595cb-hhwgx\" (UID: \"8ebdcb73-11e8-487c-b4ae-7b55b0644382\") " pod="openshift-keda/keda-operator-ffbb595cb-hhwgx" Apr 22 17:40:37.256941 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.256698 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cabundle0\" (UniqueName: \"kubernetes.io/configmap/8ebdcb73-11e8-487c-b4ae-7b55b0644382-cabundle0\") pod \"keda-operator-ffbb595cb-hhwgx\" (UID: \"8ebdcb73-11e8-487c-b4ae-7b55b0644382\") " pod="openshift-keda/keda-operator-ffbb595cb-hhwgx" Apr 22 17:40:37.256941 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:37.256773 2581 projected.go:264] Couldn't get secret openshift-keda/keda-operator-certs: secret "keda-operator-certs" not found Apr 22 17:40:37.256941 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:37.256804 2581 secret.go:281] references non-existent secret key: ca.crt Apr 22 17:40:37.256941 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:37.256814 2581 projected.go:277] Couldn't get secret payload openshift-keda/kedaorg-certs: references non-existent secret key: ca.crt Apr 22 17:40:37.256941 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:37.256829 2581 projected.go:194] Error preparing data for projected volume certificates for pod openshift-keda/keda-operator-ffbb595cb-hhwgx: [secret "keda-operator-certs" not found, references non-existent secret key: ca.crt] Apr 22 17:40:37.256941 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:37.256917 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8ebdcb73-11e8-487c-b4ae-7b55b0644382-certificates podName:8ebdcb73-11e8-487c-b4ae-7b55b0644382 nodeName:}" failed. No retries permitted until 2026-04-22 17:40:37.756869512 +0000 UTC m=+345.649565227 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certificates" (UniqueName: "kubernetes.io/projected/8ebdcb73-11e8-487c-b4ae-7b55b0644382-certificates") pod "keda-operator-ffbb595cb-hhwgx" (UID: "8ebdcb73-11e8-487c-b4ae-7b55b0644382") : [secret "keda-operator-certs" not found, references non-existent secret key: ca.crt] Apr 22 17:40:37.257489 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.257467 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cabundle0\" (UniqueName: \"kubernetes.io/configmap/8ebdcb73-11e8-487c-b4ae-7b55b0644382-cabundle0\") pod \"keda-operator-ffbb595cb-hhwgx\" (UID: \"8ebdcb73-11e8-487c-b4ae-7b55b0644382\") " pod="openshift-keda/keda-operator-ffbb595cb-hhwgx" Apr 22 17:40:37.266998 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.266970 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rgd8\" (UniqueName: \"kubernetes.io/projected/8ebdcb73-11e8-487c-b4ae-7b55b0644382-kube-api-access-9rgd8\") pod \"keda-operator-ffbb595cb-hhwgx\" (UID: \"8ebdcb73-11e8-487c-b4ae-7b55b0644382\") " pod="openshift-keda/keda-operator-ffbb595cb-hhwgx" Apr 22 17:40:37.387024 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.386950 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-keda/keda-metrics-apiserver-7c9f485588-dtj8s"] Apr 22 17:40:37.390256 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.390237 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-keda/keda-metrics-apiserver-7c9f485588-dtj8s" Apr 22 17:40:37.392438 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.392413 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-keda\"/\"keda-metrics-apiserver-certs\"" Apr 22 17:40:37.398165 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.398138 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-keda/keda-metrics-apiserver-7c9f485588-dtj8s"] Apr 22 17:40:37.458928 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.458891 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/62d65290-70ab-4113-850d-23cbbf2b72b6-certificates\") pod \"keda-metrics-apiserver-7c9f485588-dtj8s\" (UID: \"62d65290-70ab-4113-850d-23cbbf2b72b6\") " pod="openshift-keda/keda-metrics-apiserver-7c9f485588-dtj8s" Apr 22 17:40:37.459106 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.459004 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcc7w\" (UniqueName: \"kubernetes.io/projected/62d65290-70ab-4113-850d-23cbbf2b72b6-kube-api-access-vcc7w\") pod \"keda-metrics-apiserver-7c9f485588-dtj8s\" (UID: \"62d65290-70ab-4113-850d-23cbbf2b72b6\") " pod="openshift-keda/keda-metrics-apiserver-7c9f485588-dtj8s" Apr 22 17:40:37.459106 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.459068 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"temp-vol\" (UniqueName: \"kubernetes.io/empty-dir/62d65290-70ab-4113-850d-23cbbf2b72b6-temp-vol\") pod \"keda-metrics-apiserver-7c9f485588-dtj8s\" (UID: \"62d65290-70ab-4113-850d-23cbbf2b72b6\") " pod="openshift-keda/keda-metrics-apiserver-7c9f485588-dtj8s" Apr 22 17:40:37.560071 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.560031 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"temp-vol\" (UniqueName: \"kubernetes.io/empty-dir/62d65290-70ab-4113-850d-23cbbf2b72b6-temp-vol\") pod \"keda-metrics-apiserver-7c9f485588-dtj8s\" (UID: \"62d65290-70ab-4113-850d-23cbbf2b72b6\") " pod="openshift-keda/keda-metrics-apiserver-7c9f485588-dtj8s" Apr 22 17:40:37.560251 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.560093 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/62d65290-70ab-4113-850d-23cbbf2b72b6-certificates\") pod \"keda-metrics-apiserver-7c9f485588-dtj8s\" (UID: \"62d65290-70ab-4113-850d-23cbbf2b72b6\") " pod="openshift-keda/keda-metrics-apiserver-7c9f485588-dtj8s" Apr 22 17:40:37.560251 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.560152 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vcc7w\" (UniqueName: \"kubernetes.io/projected/62d65290-70ab-4113-850d-23cbbf2b72b6-kube-api-access-vcc7w\") pod \"keda-metrics-apiserver-7c9f485588-dtj8s\" (UID: \"62d65290-70ab-4113-850d-23cbbf2b72b6\") " pod="openshift-keda/keda-metrics-apiserver-7c9f485588-dtj8s" Apr 22 17:40:37.560251 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:37.560202 2581 secret.go:281] references non-existent secret key: tls.crt Apr 22 17:40:37.560251 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:37.560226 2581 projected.go:277] Couldn't get secret payload openshift-keda/kedaorg-certs: references non-existent secret key: tls.crt Apr 22 17:40:37.560251 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:37.560250 2581 projected.go:194] Error preparing data for projected volume certificates for pod openshift-keda/keda-metrics-apiserver-7c9f485588-dtj8s: references non-existent secret key: tls.crt Apr 22 17:40:37.560435 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:37.560331 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/62d65290-70ab-4113-850d-23cbbf2b72b6-certificates podName:62d65290-70ab-4113-850d-23cbbf2b72b6 nodeName:}" failed. No retries permitted until 2026-04-22 17:40:38.060308853 +0000 UTC m=+345.953004572 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certificates" (UniqueName: "kubernetes.io/projected/62d65290-70ab-4113-850d-23cbbf2b72b6-certificates") pod "keda-metrics-apiserver-7c9f485588-dtj8s" (UID: "62d65290-70ab-4113-850d-23cbbf2b72b6") : references non-existent secret key: tls.crt Apr 22 17:40:37.560582 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.560565 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"temp-vol\" (UniqueName: \"kubernetes.io/empty-dir/62d65290-70ab-4113-850d-23cbbf2b72b6-temp-vol\") pod \"keda-metrics-apiserver-7c9f485588-dtj8s\" (UID: \"62d65290-70ab-4113-850d-23cbbf2b72b6\") " pod="openshift-keda/keda-metrics-apiserver-7c9f485588-dtj8s" Apr 22 17:40:37.569261 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.569158 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcc7w\" (UniqueName: \"kubernetes.io/projected/62d65290-70ab-4113-850d-23cbbf2b72b6-kube-api-access-vcc7w\") pod \"keda-metrics-apiserver-7c9f485588-dtj8s\" (UID: \"62d65290-70ab-4113-850d-23cbbf2b72b6\") " pod="openshift-keda/keda-metrics-apiserver-7c9f485588-dtj8s" Apr 22 17:40:37.761651 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:37.761543 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/8ebdcb73-11e8-487c-b4ae-7b55b0644382-certificates\") pod \"keda-operator-ffbb595cb-hhwgx\" (UID: \"8ebdcb73-11e8-487c-b4ae-7b55b0644382\") " pod="openshift-keda/keda-operator-ffbb595cb-hhwgx" Apr 22 17:40:37.761837 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:37.761690 2581 secret.go:281] references non-existent secret key: ca.crt Apr 22 17:40:37.761837 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:37.761714 2581 projected.go:277] Couldn't get secret payload openshift-keda/kedaorg-certs: references non-existent secret key: ca.crt Apr 22 17:40:37.761837 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:37.761724 2581 projected.go:194] Error preparing data for projected volume certificates for pod openshift-keda/keda-operator-ffbb595cb-hhwgx: references non-existent secret key: ca.crt Apr 22 17:40:37.761837 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:37.761777 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8ebdcb73-11e8-487c-b4ae-7b55b0644382-certificates podName:8ebdcb73-11e8-487c-b4ae-7b55b0644382 nodeName:}" failed. No retries permitted until 2026-04-22 17:40:38.761758433 +0000 UTC m=+346.654454148 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "certificates" (UniqueName: "kubernetes.io/projected/8ebdcb73-11e8-487c-b4ae-7b55b0644382-certificates") pod "keda-operator-ffbb595cb-hhwgx" (UID: "8ebdcb73-11e8-487c-b4ae-7b55b0644382") : references non-existent secret key: ca.crt Apr 22 17:40:38.065515 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:38.065460 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/62d65290-70ab-4113-850d-23cbbf2b72b6-certificates\") pod \"keda-metrics-apiserver-7c9f485588-dtj8s\" (UID: \"62d65290-70ab-4113-850d-23cbbf2b72b6\") " pod="openshift-keda/keda-metrics-apiserver-7c9f485588-dtj8s" Apr 22 17:40:38.066016 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:38.065638 2581 secret.go:281] references non-existent secret key: tls.crt Apr 22 17:40:38.066016 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:38.065663 2581 projected.go:277] Couldn't get secret payload openshift-keda/kedaorg-certs: references non-existent secret key: tls.crt Apr 22 17:40:38.066016 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:38.065688 2581 projected.go:194] Error preparing data for projected volume certificates for pod openshift-keda/keda-metrics-apiserver-7c9f485588-dtj8s: references non-existent secret key: tls.crt Apr 22 17:40:38.066016 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:38.065769 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/62d65290-70ab-4113-850d-23cbbf2b72b6-certificates podName:62d65290-70ab-4113-850d-23cbbf2b72b6 nodeName:}" failed. No retries permitted until 2026-04-22 17:40:39.065748978 +0000 UTC m=+346.958444697 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "certificates" (UniqueName: "kubernetes.io/projected/62d65290-70ab-4113-850d-23cbbf2b72b6-certificates") pod "keda-metrics-apiserver-7c9f485588-dtj8s" (UID: "62d65290-70ab-4113-850d-23cbbf2b72b6") : references non-existent secret key: tls.crt Apr 22 17:40:38.771449 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:38.771414 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/8ebdcb73-11e8-487c-b4ae-7b55b0644382-certificates\") pod \"keda-operator-ffbb595cb-hhwgx\" (UID: \"8ebdcb73-11e8-487c-b4ae-7b55b0644382\") " pod="openshift-keda/keda-operator-ffbb595cb-hhwgx" Apr 22 17:40:38.771633 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:38.771551 2581 secret.go:281] references non-existent secret key: ca.crt Apr 22 17:40:38.771633 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:38.771569 2581 projected.go:277] Couldn't get secret payload openshift-keda/kedaorg-certs: references non-existent secret key: ca.crt Apr 22 17:40:38.771633 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:38.771577 2581 projected.go:194] Error preparing data for projected volume certificates for pod openshift-keda/keda-operator-ffbb595cb-hhwgx: references non-existent secret key: ca.crt Apr 22 17:40:38.771633 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:38.771625 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8ebdcb73-11e8-487c-b4ae-7b55b0644382-certificates podName:8ebdcb73-11e8-487c-b4ae-7b55b0644382 nodeName:}" failed. No retries permitted until 2026-04-22 17:40:40.771612951 +0000 UTC m=+348.664308654 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "certificates" (UniqueName: "kubernetes.io/projected/8ebdcb73-11e8-487c-b4ae-7b55b0644382-certificates") pod "keda-operator-ffbb595cb-hhwgx" (UID: "8ebdcb73-11e8-487c-b4ae-7b55b0644382") : references non-existent secret key: ca.crt Apr 22 17:40:39.074665 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:39.074572 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/62d65290-70ab-4113-850d-23cbbf2b72b6-certificates\") pod \"keda-metrics-apiserver-7c9f485588-dtj8s\" (UID: \"62d65290-70ab-4113-850d-23cbbf2b72b6\") " pod="openshift-keda/keda-metrics-apiserver-7c9f485588-dtj8s" Apr 22 17:40:39.075079 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:39.074689 2581 secret.go:281] references non-existent secret key: tls.crt Apr 22 17:40:39.075079 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:39.074701 2581 projected.go:277] Couldn't get secret payload openshift-keda/kedaorg-certs: references non-existent secret key: tls.crt Apr 22 17:40:39.075079 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:39.074718 2581 projected.go:194] Error preparing data for projected volume certificates for pod openshift-keda/keda-metrics-apiserver-7c9f485588-dtj8s: references non-existent secret key: tls.crt Apr 22 17:40:39.075079 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:39.074775 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/62d65290-70ab-4113-850d-23cbbf2b72b6-certificates podName:62d65290-70ab-4113-850d-23cbbf2b72b6 nodeName:}" failed. No retries permitted until 2026-04-22 17:40:41.074761474 +0000 UTC m=+348.967457176 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "certificates" (UniqueName: "kubernetes.io/projected/62d65290-70ab-4113-850d-23cbbf2b72b6-certificates") pod "keda-metrics-apiserver-7c9f485588-dtj8s" (UID: "62d65290-70ab-4113-850d-23cbbf2b72b6") : references non-existent secret key: tls.crt Apr 22 17:40:40.789726 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:40.789674 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/8ebdcb73-11e8-487c-b4ae-7b55b0644382-certificates\") pod \"keda-operator-ffbb595cb-hhwgx\" (UID: \"8ebdcb73-11e8-487c-b4ae-7b55b0644382\") " pod="openshift-keda/keda-operator-ffbb595cb-hhwgx" Apr 22 17:40:40.790166 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:40.789833 2581 secret.go:281] references non-existent secret key: ca.crt Apr 22 17:40:40.790166 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:40.789855 2581 projected.go:277] Couldn't get secret payload openshift-keda/kedaorg-certs: references non-existent secret key: ca.crt Apr 22 17:40:40.790166 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:40.789864 2581 projected.go:194] Error preparing data for projected volume certificates for pod openshift-keda/keda-operator-ffbb595cb-hhwgx: references non-existent secret key: ca.crt Apr 22 17:40:40.790166 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:40.789940 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8ebdcb73-11e8-487c-b4ae-7b55b0644382-certificates podName:8ebdcb73-11e8-487c-b4ae-7b55b0644382 nodeName:}" failed. No retries permitted until 2026-04-22 17:40:44.789926178 +0000 UTC m=+352.682621879 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "certificates" (UniqueName: "kubernetes.io/projected/8ebdcb73-11e8-487c-b4ae-7b55b0644382-certificates") pod "keda-operator-ffbb595cb-hhwgx" (UID: "8ebdcb73-11e8-487c-b4ae-7b55b0644382") : references non-existent secret key: ca.crt Apr 22 17:40:41.092266 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:41.092150 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/62d65290-70ab-4113-850d-23cbbf2b72b6-certificates\") pod \"keda-metrics-apiserver-7c9f485588-dtj8s\" (UID: \"62d65290-70ab-4113-850d-23cbbf2b72b6\") " pod="openshift-keda/keda-metrics-apiserver-7c9f485588-dtj8s" Apr 22 17:40:41.092428 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:41.092318 2581 secret.go:281] references non-existent secret key: tls.crt Apr 22 17:40:41.092428 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:41.092342 2581 projected.go:277] Couldn't get secret payload openshift-keda/kedaorg-certs: references non-existent secret key: tls.crt Apr 22 17:40:41.092428 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:41.092360 2581 projected.go:194] Error preparing data for projected volume certificates for pod openshift-keda/keda-metrics-apiserver-7c9f485588-dtj8s: references non-existent secret key: tls.crt Apr 22 17:40:41.092428 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:40:41.092428 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/62d65290-70ab-4113-850d-23cbbf2b72b6-certificates podName:62d65290-70ab-4113-850d-23cbbf2b72b6 nodeName:}" failed. No retries permitted until 2026-04-22 17:40:45.092413085 +0000 UTC m=+352.985108790 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "certificates" (UniqueName: "kubernetes.io/projected/62d65290-70ab-4113-850d-23cbbf2b72b6-certificates") pod "keda-metrics-apiserver-7c9f485588-dtj8s" (UID: "62d65290-70ab-4113-850d-23cbbf2b72b6") : references non-existent secret key: tls.crt Apr 22 17:40:44.822234 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:44.822195 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/8ebdcb73-11e8-487c-b4ae-7b55b0644382-certificates\") pod \"keda-operator-ffbb595cb-hhwgx\" (UID: \"8ebdcb73-11e8-487c-b4ae-7b55b0644382\") " pod="openshift-keda/keda-operator-ffbb595cb-hhwgx" Apr 22 17:40:44.824897 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:44.824841 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/8ebdcb73-11e8-487c-b4ae-7b55b0644382-certificates\") pod \"keda-operator-ffbb595cb-hhwgx\" (UID: \"8ebdcb73-11e8-487c-b4ae-7b55b0644382\") " pod="openshift-keda/keda-operator-ffbb595cb-hhwgx" Apr 22 17:40:44.842681 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:44.842642 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-keda/keda-operator-ffbb595cb-hhwgx" Apr 22 17:40:44.966711 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:44.966663 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-keda/keda-operator-ffbb595cb-hhwgx"] Apr 22 17:40:44.970468 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:40:44.970440 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ebdcb73_11e8_487c_b4ae_7b55b0644382.slice/crio-55a224945d2a3ae7a0cd1800b5184384452dd756575519371200e24dc4e8445d WatchSource:0}: Error finding container 55a224945d2a3ae7a0cd1800b5184384452dd756575519371200e24dc4e8445d: Status 404 returned error can't find the container with id 55a224945d2a3ae7a0cd1800b5184384452dd756575519371200e24dc4e8445d Apr 22 17:40:45.124897 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:45.124782 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/62d65290-70ab-4113-850d-23cbbf2b72b6-certificates\") pod \"keda-metrics-apiserver-7c9f485588-dtj8s\" (UID: \"62d65290-70ab-4113-850d-23cbbf2b72b6\") " pod="openshift-keda/keda-metrics-apiserver-7c9f485588-dtj8s" Apr 22 17:40:45.127409 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:45.127384 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"certificates\" (UniqueName: \"kubernetes.io/projected/62d65290-70ab-4113-850d-23cbbf2b72b6-certificates\") pod \"keda-metrics-apiserver-7c9f485588-dtj8s\" (UID: \"62d65290-70ab-4113-850d-23cbbf2b72b6\") " pod="openshift-keda/keda-metrics-apiserver-7c9f485588-dtj8s" Apr 22 17:40:45.201437 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:45.201388 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-keda/keda-metrics-apiserver-7c9f485588-dtj8s" Apr 22 17:40:45.323733 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:45.323696 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-keda/keda-metrics-apiserver-7c9f485588-dtj8s"] Apr 22 17:40:45.326647 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:40:45.326619 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62d65290_70ab_4113_850d_23cbbf2b72b6.slice/crio-a92743a4b5e6545c3dcb06122b09f909b838d76e26363200f13f63d1c9b2b87b WatchSource:0}: Error finding container a92743a4b5e6545c3dcb06122b09f909b838d76e26363200f13f63d1c9b2b87b: Status 404 returned error can't find the container with id a92743a4b5e6545c3dcb06122b09f909b838d76e26363200f13f63d1c9b2b87b Apr 22 17:40:45.682203 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:45.682166 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-keda/keda-metrics-apiserver-7c9f485588-dtj8s" event={"ID":"62d65290-70ab-4113-850d-23cbbf2b72b6","Type":"ContainerStarted","Data":"a92743a4b5e6545c3dcb06122b09f909b838d76e26363200f13f63d1c9b2b87b"} Apr 22 17:40:45.683032 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:45.683008 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-keda/keda-operator-ffbb595cb-hhwgx" event={"ID":"8ebdcb73-11e8-487c-b4ae-7b55b0644382","Type":"ContainerStarted","Data":"55a224945d2a3ae7a0cd1800b5184384452dd756575519371200e24dc4e8445d"} Apr 22 17:40:50.699403 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:50.699368 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-keda/keda-operator-ffbb595cb-hhwgx" event={"ID":"8ebdcb73-11e8-487c-b4ae-7b55b0644382","Type":"ContainerStarted","Data":"401983d0528b5f4acf6f8cd7d564fb6329b44ac6c9b195be365858c4d7c3e54f"} Apr 22 17:40:50.700135 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:50.699530 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-keda/keda-operator-ffbb595cb-hhwgx" Apr 22 17:40:50.700684 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:50.700661 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-keda/keda-metrics-apiserver-7c9f485588-dtj8s" event={"ID":"62d65290-70ab-4113-850d-23cbbf2b72b6","Type":"ContainerStarted","Data":"6f4b8f6859252228644d9e1fa2f72c07981f7243a59c67485f9f8e794637c7ae"} Apr 22 17:40:50.700810 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:50.700797 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-keda/keda-metrics-apiserver-7c9f485588-dtj8s" Apr 22 17:40:50.714802 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:50.714742 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-keda/keda-operator-ffbb595cb-hhwgx" podStartSLOduration=8.868013506 podStartE2EDuration="13.714725036s" podCreationTimestamp="2026-04-22 17:40:37 +0000 UTC" firstStartedPulling="2026-04-22 17:40:44.971700186 +0000 UTC m=+352.864395892" lastFinishedPulling="2026-04-22 17:40:49.81841171 +0000 UTC m=+357.711107422" observedRunningTime="2026-04-22 17:40:50.714529158 +0000 UTC m=+358.607224882" watchObservedRunningTime="2026-04-22 17:40:50.714725036 +0000 UTC m=+358.607420799" Apr 22 17:40:50.732214 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:40:50.732151 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-keda/keda-metrics-apiserver-7c9f485588-dtj8s" podStartSLOduration=9.250698947 podStartE2EDuration="13.732133846s" podCreationTimestamp="2026-04-22 17:40:37 +0000 UTC" firstStartedPulling="2026-04-22 17:40:45.328013907 +0000 UTC m=+353.220709608" lastFinishedPulling="2026-04-22 17:40:49.809448794 +0000 UTC m=+357.702144507" observedRunningTime="2026-04-22 17:40:50.730213678 +0000 UTC m=+358.622909402" watchObservedRunningTime="2026-04-22 17:40:50.732133846 +0000 UTC m=+358.624829570" Apr 22 17:41:01.708095 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:41:01.708066 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-keda/keda-metrics-apiserver-7c9f485588-dtj8s" Apr 22 17:41:11.705771 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:41:11.705739 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-keda/keda-operator-ffbb595cb-hhwgx" Apr 22 17:41:45.163080 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:41:45.163041 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve/seaweedfs-86cc847c5c-hjjbq"] Apr 22 17:41:45.166393 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:41:45.166375 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve/seaweedfs-86cc847c5c-hjjbq" Apr 22 17:41:45.168779 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:41:45.168756 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kserve\"/\"default-dockercfg-nd6qt\"" Apr 22 17:41:45.168779 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:41:45.168770 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kserve\"/\"openshift-service-ca.crt\"" Apr 22 17:41:45.168962 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:41:45.168756 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kserve\"/\"mlpipeline-s3-artifact\"" Apr 22 17:41:45.169510 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:41:45.169490 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kserve\"/\"kube-root-ca.crt\"" Apr 22 17:41:45.169679 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:41:45.169655 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data\" (UniqueName: \"kubernetes.io/empty-dir/a553d1b9-68b1-4f40-b5b8-f115369d1a55-data\") pod \"seaweedfs-86cc847c5c-hjjbq\" (UID: \"a553d1b9-68b1-4f40-b5b8-f115369d1a55\") " pod="kserve/seaweedfs-86cc847c5c-hjjbq" Apr 22 17:41:45.169799 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:41:45.169725 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bczb\" (UniqueName: \"kubernetes.io/projected/a553d1b9-68b1-4f40-b5b8-f115369d1a55-kube-api-access-8bczb\") pod \"seaweedfs-86cc847c5c-hjjbq\" (UID: \"a553d1b9-68b1-4f40-b5b8-f115369d1a55\") " pod="kserve/seaweedfs-86cc847c5c-hjjbq" Apr 22 17:41:45.177834 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:41:45.177803 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve/seaweedfs-86cc847c5c-hjjbq"] Apr 22 17:41:45.270232 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:41:45.270189 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"data\" (UniqueName: \"kubernetes.io/empty-dir/a553d1b9-68b1-4f40-b5b8-f115369d1a55-data\") pod \"seaweedfs-86cc847c5c-hjjbq\" (UID: \"a553d1b9-68b1-4f40-b5b8-f115369d1a55\") " pod="kserve/seaweedfs-86cc847c5c-hjjbq" Apr 22 17:41:45.270232 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:41:45.270245 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8bczb\" (UniqueName: \"kubernetes.io/projected/a553d1b9-68b1-4f40-b5b8-f115369d1a55-kube-api-access-8bczb\") pod \"seaweedfs-86cc847c5c-hjjbq\" (UID: \"a553d1b9-68b1-4f40-b5b8-f115369d1a55\") " pod="kserve/seaweedfs-86cc847c5c-hjjbq" Apr 22 17:41:45.270578 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:41:45.270556 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"data\" (UniqueName: \"kubernetes.io/empty-dir/a553d1b9-68b1-4f40-b5b8-f115369d1a55-data\") pod \"seaweedfs-86cc847c5c-hjjbq\" (UID: \"a553d1b9-68b1-4f40-b5b8-f115369d1a55\") " pod="kserve/seaweedfs-86cc847c5c-hjjbq" Apr 22 17:41:45.278463 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:41:45.278436 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bczb\" (UniqueName: \"kubernetes.io/projected/a553d1b9-68b1-4f40-b5b8-f115369d1a55-kube-api-access-8bczb\") pod \"seaweedfs-86cc847c5c-hjjbq\" (UID: \"a553d1b9-68b1-4f40-b5b8-f115369d1a55\") " pod="kserve/seaweedfs-86cc847c5c-hjjbq" Apr 22 17:41:45.480032 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:41:45.479937 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve/seaweedfs-86cc847c5c-hjjbq" Apr 22 17:41:45.600050 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:41:45.600018 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve/seaweedfs-86cc847c5c-hjjbq"] Apr 22 17:41:45.604305 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:41:45.604269 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda553d1b9_68b1_4f40_b5b8_f115369d1a55.slice/crio-e42b0a71f539f558989045a7c9ce522659faf42eb18c3f8d7c61b13d28682538 WatchSource:0}: Error finding container e42b0a71f539f558989045a7c9ce522659faf42eb18c3f8d7c61b13d28682538: Status 404 returned error can't find the container with id e42b0a71f539f558989045a7c9ce522659faf42eb18c3f8d7c61b13d28682538 Apr 22 17:41:45.842360 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:41:45.842329 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve/seaweedfs-86cc847c5c-hjjbq" event={"ID":"a553d1b9-68b1-4f40-b5b8-f115369d1a55","Type":"ContainerStarted","Data":"e42b0a71f539f558989045a7c9ce522659faf42eb18c3f8d7c61b13d28682538"} Apr 22 17:41:48.851125 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:41:48.851089 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve/seaweedfs-86cc847c5c-hjjbq" event={"ID":"a553d1b9-68b1-4f40-b5b8-f115369d1a55","Type":"ContainerStarted","Data":"ebb04496ac492e83f6119497e61e4218010d64c877d2b2bd35c4e22b17f49664"} Apr 22 17:41:48.851537 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:41:48.851223 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve/seaweedfs-86cc847c5c-hjjbq" Apr 22 17:41:48.867819 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:41:48.867766 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve/seaweedfs-86cc847c5c-hjjbq" podStartSLOduration=1.5013839949999999 podStartE2EDuration="3.867750159s" podCreationTimestamp="2026-04-22 17:41:45 +0000 UTC" firstStartedPulling="2026-04-22 17:41:45.60542888 +0000 UTC m=+413.498124582" lastFinishedPulling="2026-04-22 17:41:47.971795042 +0000 UTC m=+415.864490746" observedRunningTime="2026-04-22 17:41:48.866824809 +0000 UTC m=+416.759520532" watchObservedRunningTime="2026-04-22 17:41:48.867750159 +0000 UTC m=+416.760445881" Apr 22 17:41:54.856525 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:41:54.856496 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve/seaweedfs-86cc847c5c-hjjbq" Apr 22 17:42:55.211372 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:42:55.211334 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve/model-serving-api-86f7b4b499-z45jv"] Apr 22 17:42:55.214433 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:42:55.214413 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve/model-serving-api-86f7b4b499-z45jv" Apr 22 17:42:55.219126 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:42:55.219103 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kserve\"/\"model-serving-api-dockercfg-7bqnc\"" Apr 22 17:42:55.219247 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:42:55.219112 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kserve\"/\"model-serving-api-tls\"" Apr 22 17:42:55.226707 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:42:55.226686 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve/model-serving-api-86f7b4b499-z45jv"] Apr 22 17:42:55.280749 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:42:55.280708 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vw8z\" (UniqueName: \"kubernetes.io/projected/aa87536a-7da2-4334-b300-c89609b7e02c-kube-api-access-6vw8z\") pod \"model-serving-api-86f7b4b499-z45jv\" (UID: \"aa87536a-7da2-4334-b300-c89609b7e02c\") " pod="kserve/model-serving-api-86f7b4b499-z45jv" Apr 22 17:42:55.280923 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:42:55.280804 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certs\" (UniqueName: \"kubernetes.io/secret/aa87536a-7da2-4334-b300-c89609b7e02c-tls-certs\") pod \"model-serving-api-86f7b4b499-z45jv\" (UID: \"aa87536a-7da2-4334-b300-c89609b7e02c\") " pod="kserve/model-serving-api-86f7b4b499-z45jv" Apr 22 17:42:55.382001 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:42:55.381959 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-certs\" (UniqueName: \"kubernetes.io/secret/aa87536a-7da2-4334-b300-c89609b7e02c-tls-certs\") pod \"model-serving-api-86f7b4b499-z45jv\" (UID: \"aa87536a-7da2-4334-b300-c89609b7e02c\") " pod="kserve/model-serving-api-86f7b4b499-z45jv" Apr 22 17:42:55.382187 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:42:55.382014 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6vw8z\" (UniqueName: \"kubernetes.io/projected/aa87536a-7da2-4334-b300-c89609b7e02c-kube-api-access-6vw8z\") pod \"model-serving-api-86f7b4b499-z45jv\" (UID: \"aa87536a-7da2-4334-b300-c89609b7e02c\") " pod="kserve/model-serving-api-86f7b4b499-z45jv" Apr 22 17:42:55.382187 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:42:55.382121 2581 secret.go:189] Couldn't get secret kserve/model-serving-api-tls: secret "model-serving-api-tls" not found Apr 22 17:42:55.382295 ip-10-0-138-57 kubenswrapper[2581]: E0422 17:42:55.382212 2581 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aa87536a-7da2-4334-b300-c89609b7e02c-tls-certs podName:aa87536a-7da2-4334-b300-c89609b7e02c nodeName:}" failed. No retries permitted until 2026-04-22 17:42:55.882189307 +0000 UTC m=+483.774885018 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-certs" (UniqueName: "kubernetes.io/secret/aa87536a-7da2-4334-b300-c89609b7e02c-tls-certs") pod "model-serving-api-86f7b4b499-z45jv" (UID: "aa87536a-7da2-4334-b300-c89609b7e02c") : secret "model-serving-api-tls" not found Apr 22 17:42:55.392270 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:42:55.392247 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vw8z\" (UniqueName: \"kubernetes.io/projected/aa87536a-7da2-4334-b300-c89609b7e02c-kube-api-access-6vw8z\") pod \"model-serving-api-86f7b4b499-z45jv\" (UID: \"aa87536a-7da2-4334-b300-c89609b7e02c\") " pod="kserve/model-serving-api-86f7b4b499-z45jv" Apr 22 17:42:55.886954 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:42:55.886901 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-certs\" (UniqueName: \"kubernetes.io/secret/aa87536a-7da2-4334-b300-c89609b7e02c-tls-certs\") pod \"model-serving-api-86f7b4b499-z45jv\" (UID: \"aa87536a-7da2-4334-b300-c89609b7e02c\") " pod="kserve/model-serving-api-86f7b4b499-z45jv" Apr 22 17:42:55.889390 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:42:55.889372 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-certs\" (UniqueName: \"kubernetes.io/secret/aa87536a-7da2-4334-b300-c89609b7e02c-tls-certs\") pod \"model-serving-api-86f7b4b499-z45jv\" (UID: \"aa87536a-7da2-4334-b300-c89609b7e02c\") " pod="kserve/model-serving-api-86f7b4b499-z45jv" Apr 22 17:42:56.126761 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:42:56.126710 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve/model-serving-api-86f7b4b499-z45jv" Apr 22 17:42:56.247744 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:42:56.247708 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve/model-serving-api-86f7b4b499-z45jv"] Apr 22 17:42:56.250866 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:42:56.250832 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa87536a_7da2_4334_b300_c89609b7e02c.slice/crio-4bbf53b8d2e93f2ce65d2c4b7d5fb1aabd4781b79b90dab8dd8bcbf9ea35ae9f WatchSource:0}: Error finding container 4bbf53b8d2e93f2ce65d2c4b7d5fb1aabd4781b79b90dab8dd8bcbf9ea35ae9f: Status 404 returned error can't find the container with id 4bbf53b8d2e93f2ce65d2c4b7d5fb1aabd4781b79b90dab8dd8bcbf9ea35ae9f Apr 22 17:42:57.038430 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:42:57.038373 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve/model-serving-api-86f7b4b499-z45jv" event={"ID":"aa87536a-7da2-4334-b300-c89609b7e02c","Type":"ContainerStarted","Data":"4bbf53b8d2e93f2ce65d2c4b7d5fb1aabd4781b79b90dab8dd8bcbf9ea35ae9f"} Apr 22 17:42:59.044648 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:42:59.044607 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve/model-serving-api-86f7b4b499-z45jv" event={"ID":"aa87536a-7da2-4334-b300-c89609b7e02c","Type":"ContainerStarted","Data":"58a67d70f61a20bbcac6a4fe7924a02c65fbb0d48759776f10fe23d95098e084"} Apr 22 17:42:59.045066 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:42:59.044750 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve/model-serving-api-86f7b4b499-z45jv" Apr 22 17:42:59.062297 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:42:59.062252 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve/model-serving-api-86f7b4b499-z45jv" podStartSLOduration=1.883343013 podStartE2EDuration="4.062237245s" podCreationTimestamp="2026-04-22 17:42:55 +0000 UTC" firstStartedPulling="2026-04-22 17:42:56.252671775 +0000 UTC m=+484.145367477" lastFinishedPulling="2026-04-22 17:42:58.431566002 +0000 UTC m=+486.324261709" observedRunningTime="2026-04-22 17:42:59.060961983 +0000 UTC m=+486.953657718" watchObservedRunningTime="2026-04-22 17:42:59.062237245 +0000 UTC m=+486.954932970" Apr 22 17:43:10.051485 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:43:10.051450 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve/model-serving-api-86f7b4b499-z45jv" Apr 22 17:43:30.758652 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:43:30.758623 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-a1771-predictor-5f659f66bd-hn25n"] Apr 22 17:43:30.761774 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:43:30.761758 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-a1771-predictor-5f659f66bd-hn25n" Apr 22 17:43:30.763840 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:43:30.763821 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kserve-ci-e2e-test\"/\"default-dockercfg-22fdj\"" Apr 22 17:43:30.771013 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:43:30.770993 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-a1771-predictor-5f659f66bd-hn25n"] Apr 22 17:43:30.771558 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:43:30.771542 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kserve-ci-e2e-test/success-200-isvc-a1771-predictor-5f659f66bd-hn25n" Apr 22 17:43:30.889101 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:43:30.889068 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kserve-ci-e2e-test/success-200-isvc-a1771-predictor-5f659f66bd-hn25n"] Apr 22 17:43:30.892240 ip-10-0-138-57 kubenswrapper[2581]: W0422 17:43:30.892214 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8bbd723_f094_4e1d_8792_5ba352e66fad.slice/crio-de15b1499c61348a5622109861318394ee1d8b297937217af3306509357afd1e WatchSource:0}: Error finding container de15b1499c61348a5622109861318394ee1d8b297937217af3306509357afd1e: Status 404 returned error can't find the container with id de15b1499c61348a5622109861318394ee1d8b297937217af3306509357afd1e Apr 22 17:43:31.132965 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:43:31.132933 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-a1771-predictor-5f659f66bd-hn25n" event={"ID":"b8bbd723-f094-4e1d-8792-5ba352e66fad","Type":"ContainerStarted","Data":"de15b1499c61348a5622109861318394ee1d8b297937217af3306509357afd1e"} Apr 22 17:43:43.173402 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:43:43.173363 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kserve-ci-e2e-test/success-200-isvc-a1771-predictor-5f659f66bd-hn25n" event={"ID":"b8bbd723-f094-4e1d-8792-5ba352e66fad","Type":"ContainerStarted","Data":"fb5e3fbf70cb47285957437cfadfa2474aa9e0067c4d222130abee96f9aaf356"} Apr 22 17:43:43.173912 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:43:43.173543 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kserve-ci-e2e-test/success-200-isvc-a1771-predictor-5f659f66bd-hn25n" Apr 22 17:43:43.174754 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:43:43.174723 2581 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-a1771-predictor-5f659f66bd-hn25n" podUID="b8bbd723-f094-4e1d-8792-5ba352e66fad" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.20:8080: connect: connection refused" Apr 22 17:43:44.176508 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:43:44.176469 2581 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-a1771-predictor-5f659f66bd-hn25n" podUID="b8bbd723-f094-4e1d-8792-5ba352e66fad" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.20:8080: connect: connection refused" Apr 22 17:43:54.176596 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:43:54.176548 2581 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-a1771-predictor-5f659f66bd-hn25n" podUID="b8bbd723-f094-4e1d-8792-5ba352e66fad" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.20:8080: connect: connection refused" Apr 22 17:44:04.177297 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:44:04.177245 2581 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-a1771-predictor-5f659f66bd-hn25n" podUID="b8bbd723-f094-4e1d-8792-5ba352e66fad" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.20:8080: connect: connection refused" Apr 22 17:44:14.177344 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:44:14.177291 2581 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-a1771-predictor-5f659f66bd-hn25n" podUID="b8bbd723-f094-4e1d-8792-5ba352e66fad" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.20:8080: connect: connection refused" Apr 22 17:44:24.176655 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:44:24.176602 2581 prober.go:120] "Probe failed" probeType="Readiness" pod="kserve-ci-e2e-test/success-200-isvc-a1771-predictor-5f659f66bd-hn25n" podUID="b8bbd723-f094-4e1d-8792-5ba352e66fad" containerName="kserve-container" probeResult="failure" output="dial tcp 10.134.0.20:8080: connect: connection refused" Apr 22 17:44:34.178356 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:44:34.178325 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kserve-ci-e2e-test/success-200-isvc-a1771-predictor-5f659f66bd-hn25n" Apr 22 17:44:34.196046 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:44:34.195995 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kserve-ci-e2e-test/success-200-isvc-a1771-predictor-5f659f66bd-hn25n" podStartSLOduration=52.430434544 podStartE2EDuration="1m4.195979858s" podCreationTimestamp="2026-04-22 17:43:30 +0000 UTC" firstStartedPulling="2026-04-22 17:43:30.894432697 +0000 UTC m=+518.787128402" lastFinishedPulling="2026-04-22 17:43:42.659978 +0000 UTC m=+530.552673716" observedRunningTime="2026-04-22 17:43:43.187651339 +0000 UTC m=+531.080347063" watchObservedRunningTime="2026-04-22 17:44:34.195979858 +0000 UTC m=+582.088675583" Apr 22 17:44:52.568358 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:44:52.568318 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 17:44:52.568972 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:44:52.568633 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 17:49:52.587217 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:49:52.587119 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 17:49:52.588974 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:49:52.588951 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 17:54:52.607112 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:54:52.607078 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 17:54:52.609441 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:54:52.609422 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 17:59:52.626226 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:59:52.626115 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 17:59:52.630081 ip-10-0-138-57 kubenswrapper[2581]: I0422 17:59:52.628288 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 18:04:52.646135 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:04:52.646027 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 18:04:52.652805 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:04:52.648444 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 18:09:52.665055 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:09:52.664952 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 18:09:52.669014 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:09:52.667752 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 18:14:52.684717 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:14:52.684609 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 18:14:52.688743 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:14:52.687415 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 18:19:52.703136 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:19:52.703031 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 18:19:52.707328 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:19:52.707311 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 18:24:52.721420 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:24:52.721291 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 18:24:52.726137 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:24:52.726101 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 18:29:52.738974 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:29:52.738858 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 18:29:52.744637 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:29:52.744620 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 18:34:52.757752 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:34:52.757630 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 18:34:52.763353 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:34:52.763335 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 18:39:52.776261 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:39:52.776231 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 18:39:52.782945 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:39:52.782921 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 18:43:48.968169 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:43:48.968085 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:43:49.421212 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:43:49.421183 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:43:49.854014 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:43:49.853985 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:43:50.270704 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:43:50.270674 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:43:50.691228 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:43:50.691140 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:43:51.121072 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:43:51.121045 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:43:51.565300 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:43:51.565275 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:43:51.987630 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:43:51.987552 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:43:52.410637 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:43:52.410607 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:43:52.837691 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:43:52.837657 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:43:53.262540 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:43:53.262454 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:43:53.707319 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:43:53.707282 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:43:54.152479 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:43:54.152452 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:43:54.582767 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:43:54.582743 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:43:55.023363 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:43:55.023330 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:43:55.459828 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:43:55.459747 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:43:55.925127 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:43:55.925099 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:43:56.438525 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:43:56.438499 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:43:56.986928 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:43:56.986898 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:43:57.503414 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:43:57.503382 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:43:57.983598 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:43:57.983572 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:43:58.438680 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:43:58.438644 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:43:58.901202 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:43:58.901170 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:43:59.374281 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:43:59.374253 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:43:59.844243 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:43:59.844214 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:44:00.325929 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:00.325897 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:44:00.771920 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:00.771894 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:44:01.242329 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:01.242248 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:44:01.690397 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:01.690368 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:44:02.147226 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:02.147184 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:44:02.592113 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:02.592080 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:44:03.034418 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:03.034388 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:44:03.465930 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:03.465824 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:44:03.911534 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:03.911496 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:44:04.360408 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:04.360368 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve-ci-e2e-test_success-200-isvc-a1771-predictor-5f659f66bd-hn25n_b8bbd723-f094-4e1d-8792-5ba352e66fad/kserve-container/0.log" Apr 22 18:44:09.069512 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:09.069472 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_global-pull-secret-syncer-drtgm_8ab355fc-d1b4-424a-94f9-5d7cad6d032f/global-pull-secret-syncer/0.log" Apr 22 18:44:09.375178 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:09.375098 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_konnectivity-agent-vtsck_0288dec2-ebf5-4de0-82a8-4785af694080/konnectivity-agent/0.log" Apr 22 18:44:09.460465 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:09.460431 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_kube-apiserver-proxy-ip-10-0-138-57.ec2.internal_8df1f5cf5626886b4ad3778d0dba9ac9/haproxy/0.log" Apr 22 18:44:13.631226 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:13.631196 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-sf6bh_b9de5d49-8aba-4723-8f93-a5cf469b3ea8/node-exporter/0.log" Apr 22 18:44:13.651797 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:13.651773 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-sf6bh_b9de5d49-8aba-4723-8f93-a5cf469b3ea8/kube-rbac-proxy/0.log" Apr 22 18:44:13.677914 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:13.677897 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-sf6bh_b9de5d49-8aba-4723-8f93-a5cf469b3ea8/init-textfile/0.log" Apr 22 18:44:14.063198 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:14.063164 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-operator-5676c8c784-mghw5_fa6843bd-5517-4faa-851b-99232e1b972e/prometheus-operator/0.log" Apr 22 18:44:14.083458 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:14.083422 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-operator-5676c8c784-mghw5_fa6843bd-5517-4faa-851b-99232e1b972e/kube-rbac-proxy/0.log" Apr 22 18:44:16.148550 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:16.148516 2581 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-must-gather-cfxdp/perf-node-gather-daemonset-s7fwp"] Apr 22 18:44:16.151556 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:16.151534 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cfxdp/perf-node-gather-daemonset-s7fwp" Apr 22 18:44:16.154228 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:16.154201 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-cfxdp\"/\"openshift-service-ca.crt\"" Apr 22 18:44:16.154600 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:16.154577 2581 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-must-gather-cfxdp\"/\"default-dockercfg-m8sx6\"" Apr 22 18:44:16.154866 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:16.154840 2581 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-cfxdp\"/\"kube-root-ca.crt\"" Apr 22 18:44:16.165241 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:16.165217 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-cfxdp/perf-node-gather-daemonset-s7fwp"] Apr 22 18:44:16.243099 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:16.243064 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5f6zm\" (UniqueName: \"kubernetes.io/projected/8b0fc8af-1824-4097-a34a-4c4138f1c51a-kube-api-access-5f6zm\") pod \"perf-node-gather-daemonset-s7fwp\" (UID: \"8b0fc8af-1824-4097-a34a-4c4138f1c51a\") " pod="openshift-must-gather-cfxdp/perf-node-gather-daemonset-s7fwp" Apr 22 18:44:16.243276 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:16.243108 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/8b0fc8af-1824-4097-a34a-4c4138f1c51a-podres\") pod \"perf-node-gather-daemonset-s7fwp\" (UID: \"8b0fc8af-1824-4097-a34a-4c4138f1c51a\") " pod="openshift-must-gather-cfxdp/perf-node-gather-daemonset-s7fwp" Apr 22 18:44:16.243276 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:16.243135 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8b0fc8af-1824-4097-a34a-4c4138f1c51a-sys\") pod \"perf-node-gather-daemonset-s7fwp\" (UID: \"8b0fc8af-1824-4097-a34a-4c4138f1c51a\") " pod="openshift-must-gather-cfxdp/perf-node-gather-daemonset-s7fwp" Apr 22 18:44:16.243276 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:16.243150 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/8b0fc8af-1824-4097-a34a-4c4138f1c51a-proc\") pod \"perf-node-gather-daemonset-s7fwp\" (UID: \"8b0fc8af-1824-4097-a34a-4c4138f1c51a\") " pod="openshift-must-gather-cfxdp/perf-node-gather-daemonset-s7fwp" Apr 22 18:44:16.243276 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:16.243200 2581 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8b0fc8af-1824-4097-a34a-4c4138f1c51a-lib-modules\") pod \"perf-node-gather-daemonset-s7fwp\" (UID: \"8b0fc8af-1824-4097-a34a-4c4138f1c51a\") " pod="openshift-must-gather-cfxdp/perf-node-gather-daemonset-s7fwp" Apr 22 18:44:16.343836 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:16.343806 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8b0fc8af-1824-4097-a34a-4c4138f1c51a-lib-modules\") pod \"perf-node-gather-daemonset-s7fwp\" (UID: \"8b0fc8af-1824-4097-a34a-4c4138f1c51a\") " pod="openshift-must-gather-cfxdp/perf-node-gather-daemonset-s7fwp" Apr 22 18:44:16.343975 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:16.343845 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-5f6zm\" (UniqueName: \"kubernetes.io/projected/8b0fc8af-1824-4097-a34a-4c4138f1c51a-kube-api-access-5f6zm\") pod \"perf-node-gather-daemonset-s7fwp\" (UID: \"8b0fc8af-1824-4097-a34a-4c4138f1c51a\") " pod="openshift-must-gather-cfxdp/perf-node-gather-daemonset-s7fwp" Apr 22 18:44:16.343975 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:16.343872 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/8b0fc8af-1824-4097-a34a-4c4138f1c51a-podres\") pod \"perf-node-gather-daemonset-s7fwp\" (UID: \"8b0fc8af-1824-4097-a34a-4c4138f1c51a\") " pod="openshift-must-gather-cfxdp/perf-node-gather-daemonset-s7fwp" Apr 22 18:44:16.343975 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:16.343917 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8b0fc8af-1824-4097-a34a-4c4138f1c51a-sys\") pod \"perf-node-gather-daemonset-s7fwp\" (UID: \"8b0fc8af-1824-4097-a34a-4c4138f1c51a\") " pod="openshift-must-gather-cfxdp/perf-node-gather-daemonset-s7fwp" Apr 22 18:44:16.343975 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:16.343933 2581 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/8b0fc8af-1824-4097-a34a-4c4138f1c51a-proc\") pod \"perf-node-gather-daemonset-s7fwp\" (UID: \"8b0fc8af-1824-4097-a34a-4c4138f1c51a\") " pod="openshift-must-gather-cfxdp/perf-node-gather-daemonset-s7fwp" Apr 22 18:44:16.344144 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:16.344010 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/8b0fc8af-1824-4097-a34a-4c4138f1c51a-proc\") pod \"perf-node-gather-daemonset-s7fwp\" (UID: \"8b0fc8af-1824-4097-a34a-4c4138f1c51a\") " pod="openshift-must-gather-cfxdp/perf-node-gather-daemonset-s7fwp" Apr 22 18:44:16.344144 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:16.344035 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8b0fc8af-1824-4097-a34a-4c4138f1c51a-sys\") pod \"perf-node-gather-daemonset-s7fwp\" (UID: \"8b0fc8af-1824-4097-a34a-4c4138f1c51a\") " pod="openshift-must-gather-cfxdp/perf-node-gather-daemonset-s7fwp" Apr 22 18:44:16.344144 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:16.344050 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8b0fc8af-1824-4097-a34a-4c4138f1c51a-lib-modules\") pod \"perf-node-gather-daemonset-s7fwp\" (UID: \"8b0fc8af-1824-4097-a34a-4c4138f1c51a\") " pod="openshift-must-gather-cfxdp/perf-node-gather-daemonset-s7fwp" Apr 22 18:44:16.344144 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:16.344069 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/8b0fc8af-1824-4097-a34a-4c4138f1c51a-podres\") pod \"perf-node-gather-daemonset-s7fwp\" (UID: \"8b0fc8af-1824-4097-a34a-4c4138f1c51a\") " pod="openshift-must-gather-cfxdp/perf-node-gather-daemonset-s7fwp" Apr 22 18:44:16.353001 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:16.352968 2581 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-5f6zm\" (UniqueName: \"kubernetes.io/projected/8b0fc8af-1824-4097-a34a-4c4138f1c51a-kube-api-access-5f6zm\") pod \"perf-node-gather-daemonset-s7fwp\" (UID: \"8b0fc8af-1824-4097-a34a-4c4138f1c51a\") " pod="openshift-must-gather-cfxdp/perf-node-gather-daemonset-s7fwp" Apr 22 18:44:16.465248 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:16.465160 2581 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-cfxdp/perf-node-gather-daemonset-s7fwp" Apr 22 18:44:16.585374 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:16.585352 2581 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-cfxdp/perf-node-gather-daemonset-s7fwp"] Apr 22 18:44:16.587521 ip-10-0-138-57 kubenswrapper[2581]: W0422 18:44:16.587498 2581 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod8b0fc8af_1824_4097_a34a_4c4138f1c51a.slice/crio-15484c32ec2bfc9e80f6e0f784ec72d45c3288bf7004c0ac77c3b3fa5bae15b7 WatchSource:0}: Error finding container 15484c32ec2bfc9e80f6e0f784ec72d45c3288bf7004c0ac77c3b3fa5bae15b7: Status 404 returned error can't find the container with id 15484c32ec2bfc9e80f6e0f784ec72d45c3288bf7004c0ac77c3b3fa5bae15b7 Apr 22 18:44:16.589224 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:16.589210 2581 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 22 18:44:17.199732 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:17.199698 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cfxdp/perf-node-gather-daemonset-s7fwp" event={"ID":"8b0fc8af-1824-4097-a34a-4c4138f1c51a","Type":"ContainerStarted","Data":"725b44ff8c5122a1d384e137fc68559e70c3645b3414dd68397b2e38c89c4139"} Apr 22 18:44:17.200098 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:17.199737 2581 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-cfxdp/perf-node-gather-daemonset-s7fwp" event={"ID":"8b0fc8af-1824-4097-a34a-4c4138f1c51a","Type":"ContainerStarted","Data":"15484c32ec2bfc9e80f6e0f784ec72d45c3288bf7004c0ac77c3b3fa5bae15b7"} Apr 22 18:44:17.200098 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:17.199756 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-must-gather-cfxdp/perf-node-gather-daemonset-s7fwp" Apr 22 18:44:17.218731 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:17.218685 2581 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-cfxdp/perf-node-gather-daemonset-s7fwp" podStartSLOduration=1.218662173 podStartE2EDuration="1.218662173s" podCreationTimestamp="2026-04-22 18:44:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-22 18:44:17.217811553 +0000 UTC m=+4165.110507277" watchObservedRunningTime="2026-04-22 18:44:17.218662173 +0000 UTC m=+4165.111357898" Apr 22 18:44:17.482848 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:17.482744 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_dns-default-nqmll_d6515c87-3bb8-42d3-be06-5bf26230f609/dns/0.log" Apr 22 18:44:17.537950 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:17.537927 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_dns-default-nqmll_d6515c87-3bb8-42d3-be06-5bf26230f609/kube-rbac-proxy/0.log" Apr 22 18:44:17.578855 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:17.578832 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_node-resolver-4m4hg_5db20131-70d2-456b-a852-663399260efa/dns-node-resolver/0.log" Apr 22 18:44:18.184130 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:18.184103 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_image-registry-69fdd76d7d-jtkk2_01529b59-d931-40c9-9938-179fb72019e7/registry/0.log" Apr 22 18:44:18.265810 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:18.265784 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_node-ca-6zmdm_8dedc3a1-a7e9-4684-b53d-a1b087d23cdb/node-ca/0.log" Apr 22 18:44:19.648999 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:19.648971 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-canary_ingress-canary-jxlcn_58576398-c4a1-4369-8c0c-3b936c2c0752/serve-healthcheck-canary/0.log" Apr 22 18:44:20.085175 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:20.085150 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-5bx7t_19c6390d-ed21-45bc-ab1f-e2df9d1ee64b/kube-rbac-proxy/0.log" Apr 22 18:44:20.105550 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:20.105521 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-5bx7t_19c6390d-ed21-45bc-ab1f-e2df9d1ee64b/exporter/0.log" Apr 22 18:44:20.127927 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:20.127904 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-5bx7t_19c6390d-ed21-45bc-ab1f-e2df9d1ee64b/extractor/0.log" Apr 22 18:44:22.420371 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:22.420339 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve_model-serving-api-86f7b4b499-z45jv_aa87536a-7da2-4334-b300-c89609b7e02c/server/0.log" Apr 22 18:44:22.955394 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:22.955357 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/kserve_seaweedfs-86cc847c5c-hjjbq_a553d1b9-68b1-4f40-b5b8-f115369d1a55/seaweedfs/0.log" Apr 22 18:44:23.212608 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:23.212534 2581 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-must-gather-cfxdp/perf-node-gather-daemonset-s7fwp" Apr 22 18:44:29.123488 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:29.123457 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-pnh8b_78313ad4-d7f9-4788-9293-7cc067465f5a/kube-multus-additional-cni-plugins/0.log" Apr 22 18:44:29.156841 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:29.156817 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-pnh8b_78313ad4-d7f9-4788-9293-7cc067465f5a/egress-router-binary-copy/0.log" Apr 22 18:44:29.188308 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:29.188285 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-pnh8b_78313ad4-d7f9-4788-9293-7cc067465f5a/cni-plugins/0.log" Apr 22 18:44:29.223129 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:29.223109 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-pnh8b_78313ad4-d7f9-4788-9293-7cc067465f5a/bond-cni-plugin/0.log" Apr 22 18:44:29.270091 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:29.270068 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-pnh8b_78313ad4-d7f9-4788-9293-7cc067465f5a/routeoverride-cni/0.log" Apr 22 18:44:29.299896 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:29.299858 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-pnh8b_78313ad4-d7f9-4788-9293-7cc067465f5a/whereabouts-cni-bincopy/0.log" Apr 22 18:44:29.336468 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:29.336441 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-pnh8b_78313ad4-d7f9-4788-9293-7cc067465f5a/whereabouts-cni/0.log" Apr 22 18:44:29.426892 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:29.426806 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-nwhzz_29e69548-2984-41bc-b4bd-8554ba75d702/kube-multus/0.log" Apr 22 18:44:29.639424 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:29.639400 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-zhl6m_cff34b01-53b3-4374-b73d-d3146e3c558d/network-metrics-daemon/0.log" Apr 22 18:44:29.685567 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:29.685491 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-zhl6m_cff34b01-53b3-4374-b73d-d3146e3c558d/kube-rbac-proxy/0.log" Apr 22 18:44:31.058184 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:31.058154 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-controller/0.log" Apr 22 18:44:31.088711 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:31.088684 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/0.log" Apr 22 18:44:31.124280 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:31.124240 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovn-acl-logging/1.log" Apr 22 18:44:31.160611 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:31.160577 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/kube-rbac-proxy-node/0.log" Apr 22 18:44:31.199758 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:31.199731 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/kube-rbac-proxy-ovn-metrics/0.log" Apr 22 18:44:31.255015 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:31.254975 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/northd/0.log" Apr 22 18:44:31.293409 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:31.293384 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/nbdb/0.log" Apr 22 18:44:31.349081 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:31.349005 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/sbdb/0.log" Apr 22 18:44:31.514735 ip-10-0-138-57 kubenswrapper[2581]: I0422 18:44:31.514695 2581 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-tn68c_24c19adf-c002-4c22-9098-f1f3be79bb45/ovnkube-controller/0.log"