Apr 17 20:44:34.400050 ip-10-0-137-110 systemd[1]: Starting Kubernetes Kubelet... Apr 17 20:44:34.860395 ip-10-0-137-110 kubenswrapper[2577]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 17 20:44:34.860395 ip-10-0-137-110 kubenswrapper[2577]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Apr 17 20:44:34.860395 ip-10-0-137-110 kubenswrapper[2577]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 17 20:44:34.860395 ip-10-0-137-110 kubenswrapper[2577]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Apr 17 20:44:34.860395 ip-10-0-137-110 kubenswrapper[2577]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 17 20:44:34.863841 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.863726 2577 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 17 20:44:34.870206 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870192 2577 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 17 20:44:34.870206 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870207 2577 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 17 20:44:34.870266 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870210 2577 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 17 20:44:34.870266 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870215 2577 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 17 20:44:34.870266 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870218 2577 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 17 20:44:34.870266 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870221 2577 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 17 20:44:34.870266 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870229 2577 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 17 20:44:34.870266 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870232 2577 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 17 20:44:34.870266 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870235 2577 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 17 20:44:34.870266 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870237 2577 feature_gate.go:328] unrecognized feature gate: Example2 Apr 17 20:44:34.870266 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870240 2577 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 17 20:44:34.870266 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870242 2577 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 17 20:44:34.870266 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870245 2577 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 17 20:44:34.870266 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870247 2577 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 17 20:44:34.870266 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870250 2577 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 17 20:44:34.870266 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870252 2577 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 17 20:44:34.870266 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870255 2577 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 17 20:44:34.870266 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870257 2577 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 17 20:44:34.870266 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870260 2577 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 17 20:44:34.870266 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870262 2577 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 17 20:44:34.870266 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870265 2577 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 17 20:44:34.870266 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870267 2577 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 17 20:44:34.870746 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870270 2577 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 17 20:44:34.870746 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870273 2577 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 17 20:44:34.870746 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870275 2577 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 17 20:44:34.870746 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870279 2577 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 17 20:44:34.870746 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870282 2577 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 17 20:44:34.870746 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870284 2577 feature_gate.go:328] unrecognized feature gate: Example Apr 17 20:44:34.870746 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870288 2577 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 17 20:44:34.870746 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870291 2577 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 17 20:44:34.870746 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870293 2577 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 17 20:44:34.870746 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870295 2577 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 17 20:44:34.870746 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870298 2577 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 17 20:44:34.870746 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870301 2577 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 17 20:44:34.870746 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870303 2577 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 17 20:44:34.870746 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870306 2577 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 17 20:44:34.870746 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870309 2577 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 17 20:44:34.870746 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870312 2577 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 17 20:44:34.870746 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870314 2577 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 17 20:44:34.870746 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870316 2577 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 17 20:44:34.870746 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870319 2577 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 17 20:44:34.870746 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870321 2577 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 17 20:44:34.871405 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870323 2577 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 17 20:44:34.871405 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870326 2577 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 17 20:44:34.871405 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870328 2577 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 17 20:44:34.871405 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870330 2577 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 17 20:44:34.871405 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870333 2577 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 17 20:44:34.871405 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870335 2577 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 17 20:44:34.871405 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870337 2577 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 17 20:44:34.871405 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870340 2577 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 17 20:44:34.871405 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870342 2577 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 17 20:44:34.871405 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870344 2577 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 17 20:44:34.871405 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870347 2577 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 17 20:44:34.871405 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870349 2577 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 17 20:44:34.871405 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870353 2577 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 17 20:44:34.871405 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870356 2577 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 17 20:44:34.871405 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870358 2577 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 17 20:44:34.871405 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870361 2577 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 17 20:44:34.871405 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870365 2577 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 17 20:44:34.871405 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870369 2577 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 17 20:44:34.871405 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870373 2577 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 17 20:44:34.871970 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870377 2577 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 17 20:44:34.871970 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870380 2577 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 17 20:44:34.871970 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870383 2577 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 17 20:44:34.871970 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870386 2577 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 17 20:44:34.871970 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870388 2577 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 17 20:44:34.871970 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870391 2577 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 17 20:44:34.871970 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870393 2577 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 17 20:44:34.871970 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870395 2577 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 17 20:44:34.871970 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870398 2577 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 17 20:44:34.871970 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870401 2577 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 17 20:44:34.871970 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870403 2577 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 17 20:44:34.871970 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870406 2577 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 17 20:44:34.871970 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870408 2577 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 17 20:44:34.871970 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870410 2577 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 17 20:44:34.871970 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870413 2577 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 17 20:44:34.871970 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870415 2577 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 17 20:44:34.871970 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870417 2577 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 17 20:44:34.871970 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870420 2577 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 17 20:44:34.871970 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870422 2577 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 17 20:44:34.872433 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870426 2577 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 17 20:44:34.872433 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870428 2577 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 17 20:44:34.872433 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870431 2577 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 17 20:44:34.872433 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870433 2577 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 17 20:44:34.872433 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870436 2577 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 17 20:44:34.872433 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.870438 2577 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 17 20:44:34.873110 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873097 2577 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 17 20:44:34.873110 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873109 2577 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 17 20:44:34.873110 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873112 2577 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 17 20:44:34.873201 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873115 2577 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 17 20:44:34.873201 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873119 2577 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 17 20:44:34.873201 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873121 2577 feature_gate.go:328] unrecognized feature gate: Example2 Apr 17 20:44:34.873201 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873124 2577 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 17 20:44:34.873201 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873128 2577 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 17 20:44:34.873201 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873133 2577 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 17 20:44:34.873201 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873135 2577 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 17 20:44:34.873201 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873138 2577 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 17 20:44:34.873201 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873141 2577 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 17 20:44:34.873201 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873143 2577 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 17 20:44:34.873201 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873146 2577 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 17 20:44:34.873201 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873149 2577 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 17 20:44:34.873201 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873151 2577 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 17 20:44:34.873201 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873154 2577 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 17 20:44:34.873201 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873156 2577 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 17 20:44:34.873201 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873159 2577 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 17 20:44:34.873201 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873161 2577 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 17 20:44:34.873201 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873164 2577 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 17 20:44:34.873201 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873166 2577 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 17 20:44:34.873658 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873168 2577 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 17 20:44:34.873658 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873171 2577 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 17 20:44:34.873658 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873173 2577 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 17 20:44:34.873658 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873176 2577 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 17 20:44:34.873658 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873179 2577 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 17 20:44:34.873658 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873181 2577 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 17 20:44:34.873658 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873183 2577 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 17 20:44:34.873658 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873186 2577 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 17 20:44:34.873658 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873188 2577 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 17 20:44:34.873658 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873190 2577 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 17 20:44:34.873658 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873193 2577 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 17 20:44:34.873658 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873195 2577 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 17 20:44:34.873658 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873197 2577 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 17 20:44:34.873658 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873200 2577 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 17 20:44:34.873658 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873204 2577 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 17 20:44:34.873658 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873206 2577 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 17 20:44:34.873658 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873209 2577 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 17 20:44:34.873658 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873212 2577 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 17 20:44:34.873658 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873215 2577 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 17 20:44:34.873658 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873217 2577 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 17 20:44:34.874168 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873220 2577 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 17 20:44:34.874168 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873222 2577 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 17 20:44:34.874168 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873225 2577 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 17 20:44:34.874168 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873227 2577 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 17 20:44:34.874168 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873230 2577 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 17 20:44:34.874168 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873233 2577 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 17 20:44:34.874168 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873236 2577 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 17 20:44:34.874168 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873238 2577 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 17 20:44:34.874168 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873241 2577 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 17 20:44:34.874168 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873243 2577 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 17 20:44:34.874168 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873246 2577 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 17 20:44:34.874168 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873248 2577 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 17 20:44:34.874168 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873251 2577 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 17 20:44:34.874168 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873253 2577 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 17 20:44:34.874168 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873255 2577 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 17 20:44:34.874168 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873258 2577 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 17 20:44:34.874168 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873261 2577 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 17 20:44:34.874168 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873263 2577 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 17 20:44:34.874168 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873266 2577 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 17 20:44:34.874168 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873268 2577 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 17 20:44:34.874653 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873270 2577 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 17 20:44:34.874653 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873273 2577 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 17 20:44:34.874653 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873275 2577 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 17 20:44:34.874653 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873278 2577 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 17 20:44:34.874653 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873280 2577 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 17 20:44:34.874653 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873283 2577 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 17 20:44:34.874653 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873286 2577 feature_gate.go:328] unrecognized feature gate: Example Apr 17 20:44:34.874653 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873289 2577 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 17 20:44:34.874653 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873291 2577 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 17 20:44:34.874653 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873294 2577 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 17 20:44:34.874653 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873296 2577 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 17 20:44:34.874653 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873299 2577 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 17 20:44:34.874653 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873301 2577 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 17 20:44:34.874653 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873303 2577 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 17 20:44:34.874653 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873306 2577 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 17 20:44:34.874653 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873308 2577 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 17 20:44:34.874653 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873312 2577 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 17 20:44:34.874653 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873314 2577 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 17 20:44:34.874653 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873317 2577 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 17 20:44:34.874653 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873320 2577 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 17 20:44:34.875238 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873322 2577 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 17 20:44:34.875238 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873324 2577 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 17 20:44:34.875238 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873328 2577 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 17 20:44:34.875238 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.873332 2577 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 17 20:44:34.875238 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873404 2577 flags.go:64] FLAG: --address="0.0.0.0" Apr 17 20:44:34.875238 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873411 2577 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Apr 17 20:44:34.875238 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873418 2577 flags.go:64] FLAG: --anonymous-auth="true" Apr 17 20:44:34.875238 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873422 2577 flags.go:64] FLAG: --application-metrics-count-limit="100" Apr 17 20:44:34.875238 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873427 2577 flags.go:64] FLAG: --authentication-token-webhook="false" Apr 17 20:44:34.875238 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873431 2577 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Apr 17 20:44:34.875238 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873435 2577 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Apr 17 20:44:34.875238 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873439 2577 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Apr 17 20:44:34.875238 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873443 2577 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Apr 17 20:44:34.875238 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873446 2577 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Apr 17 20:44:34.875238 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873449 2577 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Apr 17 20:44:34.875238 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873452 2577 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Apr 17 20:44:34.875238 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873455 2577 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Apr 17 20:44:34.875238 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873458 2577 flags.go:64] FLAG: --cgroup-root="" Apr 17 20:44:34.875238 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873461 2577 flags.go:64] FLAG: --cgroups-per-qos="true" Apr 17 20:44:34.875238 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873465 2577 flags.go:64] FLAG: --client-ca-file="" Apr 17 20:44:34.875238 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873467 2577 flags.go:64] FLAG: --cloud-config="" Apr 17 20:44:34.875238 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873470 2577 flags.go:64] FLAG: --cloud-provider="external" Apr 17 20:44:34.875238 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873473 2577 flags.go:64] FLAG: --cluster-dns="[]" Apr 17 20:44:34.875839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873477 2577 flags.go:64] FLAG: --cluster-domain="" Apr 17 20:44:34.875839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873480 2577 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Apr 17 20:44:34.875839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873483 2577 flags.go:64] FLAG: --config-dir="" Apr 17 20:44:34.875839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873486 2577 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Apr 17 20:44:34.875839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873489 2577 flags.go:64] FLAG: --container-log-max-files="5" Apr 17 20:44:34.875839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873493 2577 flags.go:64] FLAG: --container-log-max-size="10Mi" Apr 17 20:44:34.875839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873497 2577 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Apr 17 20:44:34.875839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873500 2577 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Apr 17 20:44:34.875839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873503 2577 flags.go:64] FLAG: --containerd-namespace="k8s.io" Apr 17 20:44:34.875839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873506 2577 flags.go:64] FLAG: --contention-profiling="false" Apr 17 20:44:34.875839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873508 2577 flags.go:64] FLAG: --cpu-cfs-quota="true" Apr 17 20:44:34.875839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873511 2577 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Apr 17 20:44:34.875839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873514 2577 flags.go:64] FLAG: --cpu-manager-policy="none" Apr 17 20:44:34.875839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873517 2577 flags.go:64] FLAG: --cpu-manager-policy-options="" Apr 17 20:44:34.875839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873521 2577 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Apr 17 20:44:34.875839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873524 2577 flags.go:64] FLAG: --enable-controller-attach-detach="true" Apr 17 20:44:34.875839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873527 2577 flags.go:64] FLAG: --enable-debugging-handlers="true" Apr 17 20:44:34.875839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873529 2577 flags.go:64] FLAG: --enable-load-reader="false" Apr 17 20:44:34.875839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873532 2577 flags.go:64] FLAG: --enable-server="true" Apr 17 20:44:34.875839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873535 2577 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Apr 17 20:44:34.875839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873540 2577 flags.go:64] FLAG: --event-burst="100" Apr 17 20:44:34.875839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873543 2577 flags.go:64] FLAG: --event-qps="50" Apr 17 20:44:34.875839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873546 2577 flags.go:64] FLAG: --event-storage-age-limit="default=0" Apr 17 20:44:34.875839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873549 2577 flags.go:64] FLAG: --event-storage-event-limit="default=0" Apr 17 20:44:34.875839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873552 2577 flags.go:64] FLAG: --eviction-hard="" Apr 17 20:44:34.876434 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873555 2577 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Apr 17 20:44:34.876434 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873558 2577 flags.go:64] FLAG: --eviction-minimum-reclaim="" Apr 17 20:44:34.876434 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873562 2577 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Apr 17 20:44:34.876434 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873565 2577 flags.go:64] FLAG: --eviction-soft="" Apr 17 20:44:34.876434 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873567 2577 flags.go:64] FLAG: --eviction-soft-grace-period="" Apr 17 20:44:34.876434 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873570 2577 flags.go:64] FLAG: --exit-on-lock-contention="false" Apr 17 20:44:34.876434 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873573 2577 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Apr 17 20:44:34.876434 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873576 2577 flags.go:64] FLAG: --experimental-mounter-path="" Apr 17 20:44:34.876434 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873579 2577 flags.go:64] FLAG: --fail-cgroupv1="false" Apr 17 20:44:34.876434 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873581 2577 flags.go:64] FLAG: --fail-swap-on="true" Apr 17 20:44:34.876434 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873584 2577 flags.go:64] FLAG: --feature-gates="" Apr 17 20:44:34.876434 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873588 2577 flags.go:64] FLAG: --file-check-frequency="20s" Apr 17 20:44:34.876434 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873590 2577 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Apr 17 20:44:34.876434 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873593 2577 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Apr 17 20:44:34.876434 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873597 2577 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Apr 17 20:44:34.876434 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873600 2577 flags.go:64] FLAG: --healthz-port="10248" Apr 17 20:44:34.876434 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873603 2577 flags.go:64] FLAG: --help="false" Apr 17 20:44:34.876434 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873605 2577 flags.go:64] FLAG: --hostname-override="ip-10-0-137-110.ec2.internal" Apr 17 20:44:34.876434 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873609 2577 flags.go:64] FLAG: --housekeeping-interval="10s" Apr 17 20:44:34.876434 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873611 2577 flags.go:64] FLAG: --http-check-frequency="20s" Apr 17 20:44:34.876434 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873614 2577 flags.go:64] FLAG: --image-credential-provider-bin-dir="/usr/libexec/kubelet-image-credential-provider-plugins" Apr 17 20:44:34.876434 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873617 2577 flags.go:64] FLAG: --image-credential-provider-config="/etc/kubernetes/credential-providers/ecr-credential-provider.yaml" Apr 17 20:44:34.876434 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873621 2577 flags.go:64] FLAG: --image-gc-high-threshold="85" Apr 17 20:44:34.877033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873624 2577 flags.go:64] FLAG: --image-gc-low-threshold="80" Apr 17 20:44:34.877033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873626 2577 flags.go:64] FLAG: --image-service-endpoint="" Apr 17 20:44:34.877033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873629 2577 flags.go:64] FLAG: --kernel-memcg-notification="false" Apr 17 20:44:34.877033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873632 2577 flags.go:64] FLAG: --kube-api-burst="100" Apr 17 20:44:34.877033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873635 2577 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Apr 17 20:44:34.877033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873638 2577 flags.go:64] FLAG: --kube-api-qps="50" Apr 17 20:44:34.877033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873641 2577 flags.go:64] FLAG: --kube-reserved="" Apr 17 20:44:34.877033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873644 2577 flags.go:64] FLAG: --kube-reserved-cgroup="" Apr 17 20:44:34.877033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873646 2577 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Apr 17 20:44:34.877033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873649 2577 flags.go:64] FLAG: --kubelet-cgroups="" Apr 17 20:44:34.877033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873652 2577 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Apr 17 20:44:34.877033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873655 2577 flags.go:64] FLAG: --lock-file="" Apr 17 20:44:34.877033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873658 2577 flags.go:64] FLAG: --log-cadvisor-usage="false" Apr 17 20:44:34.877033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873661 2577 flags.go:64] FLAG: --log-flush-frequency="5s" Apr 17 20:44:34.877033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873664 2577 flags.go:64] FLAG: --log-json-info-buffer-size="0" Apr 17 20:44:34.877033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873669 2577 flags.go:64] FLAG: --log-json-split-stream="false" Apr 17 20:44:34.877033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873672 2577 flags.go:64] FLAG: --log-text-info-buffer-size="0" Apr 17 20:44:34.877033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873675 2577 flags.go:64] FLAG: --log-text-split-stream="false" Apr 17 20:44:34.877033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873678 2577 flags.go:64] FLAG: --logging-format="text" Apr 17 20:44:34.877033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873680 2577 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Apr 17 20:44:34.877033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873683 2577 flags.go:64] FLAG: --make-iptables-util-chains="true" Apr 17 20:44:34.877033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873686 2577 flags.go:64] FLAG: --manifest-url="" Apr 17 20:44:34.877033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873689 2577 flags.go:64] FLAG: --manifest-url-header="" Apr 17 20:44:34.877033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873693 2577 flags.go:64] FLAG: --max-housekeeping-interval="15s" Apr 17 20:44:34.877033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873696 2577 flags.go:64] FLAG: --max-open-files="1000000" Apr 17 20:44:34.877674 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873700 2577 flags.go:64] FLAG: --max-pods="110" Apr 17 20:44:34.877674 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873703 2577 flags.go:64] FLAG: --maximum-dead-containers="-1" Apr 17 20:44:34.877674 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873706 2577 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Apr 17 20:44:34.877674 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873708 2577 flags.go:64] FLAG: --memory-manager-policy="None" Apr 17 20:44:34.877674 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873711 2577 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Apr 17 20:44:34.877674 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873714 2577 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Apr 17 20:44:34.877674 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873716 2577 flags.go:64] FLAG: --node-ip="0.0.0.0" Apr 17 20:44:34.877674 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873719 2577 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/worker=,node.openshift.io/os_id=rhel" Apr 17 20:44:34.877674 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873727 2577 flags.go:64] FLAG: --node-status-max-images="50" Apr 17 20:44:34.877674 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873729 2577 flags.go:64] FLAG: --node-status-update-frequency="10s" Apr 17 20:44:34.877674 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873732 2577 flags.go:64] FLAG: --oom-score-adj="-999" Apr 17 20:44:34.877674 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873735 2577 flags.go:64] FLAG: --pod-cidr="" Apr 17 20:44:34.877674 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873738 2577 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c8cfe89231412ff3ee8cb6207fa0be33cad0f08e88c9c0f1e9f7e8c6f14d6715" Apr 17 20:44:34.877674 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873744 2577 flags.go:64] FLAG: --pod-manifest-path="" Apr 17 20:44:34.877674 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873747 2577 flags.go:64] FLAG: --pod-max-pids="-1" Apr 17 20:44:34.877674 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873750 2577 flags.go:64] FLAG: --pods-per-core="0" Apr 17 20:44:34.877674 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873752 2577 flags.go:64] FLAG: --port="10250" Apr 17 20:44:34.877674 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873755 2577 flags.go:64] FLAG: --protect-kernel-defaults="false" Apr 17 20:44:34.877674 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873758 2577 flags.go:64] FLAG: --provider-id="aws:///us-east-1a/i-041351118a9763f9d" Apr 17 20:44:34.877674 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873761 2577 flags.go:64] FLAG: --qos-reserved="" Apr 17 20:44:34.877674 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873764 2577 flags.go:64] FLAG: --read-only-port="10255" Apr 17 20:44:34.877674 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873767 2577 flags.go:64] FLAG: --register-node="true" Apr 17 20:44:34.877674 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873770 2577 flags.go:64] FLAG: --register-schedulable="true" Apr 17 20:44:34.877674 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873773 2577 flags.go:64] FLAG: --register-with-taints="" Apr 17 20:44:34.878290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873776 2577 flags.go:64] FLAG: --registry-burst="10" Apr 17 20:44:34.878290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873779 2577 flags.go:64] FLAG: --registry-qps="5" Apr 17 20:44:34.878290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873781 2577 flags.go:64] FLAG: --reserved-cpus="" Apr 17 20:44:34.878290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873784 2577 flags.go:64] FLAG: --reserved-memory="" Apr 17 20:44:34.878290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873788 2577 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Apr 17 20:44:34.878290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873790 2577 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Apr 17 20:44:34.878290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873793 2577 flags.go:64] FLAG: --rotate-certificates="false" Apr 17 20:44:34.878290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873796 2577 flags.go:64] FLAG: --rotate-server-certificates="false" Apr 17 20:44:34.878290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873814 2577 flags.go:64] FLAG: --runonce="false" Apr 17 20:44:34.878290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873817 2577 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Apr 17 20:44:34.878290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873820 2577 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Apr 17 20:44:34.878290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873823 2577 flags.go:64] FLAG: --seccomp-default="false" Apr 17 20:44:34.878290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873826 2577 flags.go:64] FLAG: --serialize-image-pulls="true" Apr 17 20:44:34.878290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873829 2577 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Apr 17 20:44:34.878290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873832 2577 flags.go:64] FLAG: --storage-driver-db="cadvisor" Apr 17 20:44:34.878290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873835 2577 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Apr 17 20:44:34.878290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873838 2577 flags.go:64] FLAG: --storage-driver-password="root" Apr 17 20:44:34.878290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873841 2577 flags.go:64] FLAG: --storage-driver-secure="false" Apr 17 20:44:34.878290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873844 2577 flags.go:64] FLAG: --storage-driver-table="stats" Apr 17 20:44:34.878290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873847 2577 flags.go:64] FLAG: --storage-driver-user="root" Apr 17 20:44:34.878290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873850 2577 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Apr 17 20:44:34.878290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873853 2577 flags.go:64] FLAG: --sync-frequency="1m0s" Apr 17 20:44:34.878290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873855 2577 flags.go:64] FLAG: --system-cgroups="" Apr 17 20:44:34.878290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873858 2577 flags.go:64] FLAG: --system-reserved="cpu=500m,ephemeral-storage=1Gi,memory=1Gi" Apr 17 20:44:34.878290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873864 2577 flags.go:64] FLAG: --system-reserved-cgroup="" Apr 17 20:44:34.878931 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873867 2577 flags.go:64] FLAG: --tls-cert-file="" Apr 17 20:44:34.878931 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873869 2577 flags.go:64] FLAG: --tls-cipher-suites="[]" Apr 17 20:44:34.878931 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873873 2577 flags.go:64] FLAG: --tls-min-version="" Apr 17 20:44:34.878931 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873876 2577 flags.go:64] FLAG: --tls-private-key-file="" Apr 17 20:44:34.878931 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873878 2577 flags.go:64] FLAG: --topology-manager-policy="none" Apr 17 20:44:34.878931 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873881 2577 flags.go:64] FLAG: --topology-manager-policy-options="" Apr 17 20:44:34.878931 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873884 2577 flags.go:64] FLAG: --topology-manager-scope="container" Apr 17 20:44:34.878931 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873887 2577 flags.go:64] FLAG: --v="2" Apr 17 20:44:34.878931 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873890 2577 flags.go:64] FLAG: --version="false" Apr 17 20:44:34.878931 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873894 2577 flags.go:64] FLAG: --vmodule="" Apr 17 20:44:34.878931 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873898 2577 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Apr 17 20:44:34.878931 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.873901 2577 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Apr 17 20:44:34.878931 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874009 2577 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 17 20:44:34.878931 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874014 2577 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 17 20:44:34.878931 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874017 2577 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 17 20:44:34.878931 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874020 2577 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 17 20:44:34.878931 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874023 2577 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 17 20:44:34.878931 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874026 2577 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 17 20:44:34.878931 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874029 2577 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 17 20:44:34.878931 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874031 2577 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 17 20:44:34.878931 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874034 2577 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 17 20:44:34.878931 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874036 2577 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 17 20:44:34.878931 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874039 2577 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 17 20:44:34.879473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874042 2577 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 17 20:44:34.879473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874045 2577 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 17 20:44:34.879473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874049 2577 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 17 20:44:34.879473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874052 2577 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 17 20:44:34.879473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874055 2577 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 17 20:44:34.879473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874058 2577 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 17 20:44:34.879473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874060 2577 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 17 20:44:34.879473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874063 2577 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 17 20:44:34.879473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874065 2577 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 17 20:44:34.879473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874068 2577 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 17 20:44:34.879473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874070 2577 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 17 20:44:34.879473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874073 2577 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 17 20:44:34.879473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874075 2577 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 17 20:44:34.879473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874077 2577 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 17 20:44:34.879473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874080 2577 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 17 20:44:34.879473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874084 2577 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 17 20:44:34.879473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874087 2577 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 17 20:44:34.879473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874090 2577 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 17 20:44:34.879473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874093 2577 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 17 20:44:34.879473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874096 2577 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 17 20:44:34.880233 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874099 2577 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 17 20:44:34.880233 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874101 2577 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 17 20:44:34.880233 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874103 2577 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 17 20:44:34.880233 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874106 2577 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 17 20:44:34.880233 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874108 2577 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 17 20:44:34.880233 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874111 2577 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 17 20:44:34.880233 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874114 2577 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 17 20:44:34.880233 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874118 2577 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 17 20:44:34.880233 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874120 2577 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 17 20:44:34.880233 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874123 2577 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 17 20:44:34.880233 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874126 2577 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 17 20:44:34.880233 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874128 2577 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 17 20:44:34.880233 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874130 2577 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 17 20:44:34.880233 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874133 2577 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 17 20:44:34.880233 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874136 2577 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 17 20:44:34.880233 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874139 2577 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 17 20:44:34.880233 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874141 2577 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 17 20:44:34.880233 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874143 2577 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 17 20:44:34.880233 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874146 2577 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 17 20:44:34.880233 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874148 2577 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 17 20:44:34.881112 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874151 2577 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 17 20:44:34.881112 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874153 2577 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 17 20:44:34.881112 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874156 2577 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 17 20:44:34.881112 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874158 2577 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 17 20:44:34.881112 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874161 2577 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 17 20:44:34.881112 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874163 2577 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 17 20:44:34.881112 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874166 2577 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 17 20:44:34.881112 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874168 2577 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 17 20:44:34.881112 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874170 2577 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 17 20:44:34.881112 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874173 2577 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 17 20:44:34.881112 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874176 2577 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 17 20:44:34.881112 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874178 2577 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 17 20:44:34.881112 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874181 2577 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 17 20:44:34.881112 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874183 2577 feature_gate.go:328] unrecognized feature gate: Example2 Apr 17 20:44:34.881112 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874186 2577 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 17 20:44:34.881112 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874188 2577 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 17 20:44:34.881112 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874190 2577 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 17 20:44:34.881112 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874193 2577 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 17 20:44:34.881112 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874195 2577 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 17 20:44:34.881112 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874200 2577 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 17 20:44:34.881877 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874203 2577 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 17 20:44:34.881877 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874206 2577 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 17 20:44:34.881877 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874208 2577 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 17 20:44:34.881877 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874210 2577 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 17 20:44:34.881877 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874213 2577 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 17 20:44:34.881877 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874215 2577 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 17 20:44:34.881877 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874222 2577 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 17 20:44:34.881877 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874226 2577 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 17 20:44:34.881877 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874230 2577 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 17 20:44:34.881877 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874232 2577 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 17 20:44:34.881877 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874235 2577 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 17 20:44:34.881877 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874238 2577 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 17 20:44:34.881877 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874240 2577 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 17 20:44:34.881877 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874254 2577 feature_gate.go:328] unrecognized feature gate: Example Apr 17 20:44:34.881877 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.874257 2577 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 17 20:44:34.882259 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.874857 2577 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 17 20:44:34.882482 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.882465 2577 server.go:530] "Kubelet version" kubeletVersion="v1.33.9" Apr 17 20:44:34.882539 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.882487 2577 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 17 20:44:34.882587 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882563 2577 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 17 20:44:34.882587 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882571 2577 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 17 20:44:34.882587 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882576 2577 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 17 20:44:34.882587 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882581 2577 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 17 20:44:34.882587 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882585 2577 feature_gate.go:328] unrecognized feature gate: Example2 Apr 17 20:44:34.882587 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882590 2577 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 17 20:44:34.882886 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882594 2577 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 17 20:44:34.882886 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882598 2577 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 17 20:44:34.882886 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882603 2577 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 17 20:44:34.882886 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882608 2577 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 17 20:44:34.882886 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882612 2577 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 17 20:44:34.882886 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882616 2577 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 17 20:44:34.882886 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882620 2577 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 17 20:44:34.882886 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882625 2577 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 17 20:44:34.882886 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882630 2577 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 17 20:44:34.882886 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882634 2577 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 17 20:44:34.882886 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882637 2577 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 17 20:44:34.882886 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882641 2577 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 17 20:44:34.882886 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882646 2577 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 17 20:44:34.882886 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882649 2577 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 17 20:44:34.882886 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882654 2577 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 17 20:44:34.882886 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882658 2577 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 17 20:44:34.882886 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882662 2577 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 17 20:44:34.882886 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882666 2577 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 17 20:44:34.882886 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882670 2577 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 17 20:44:34.882886 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882674 2577 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 17 20:44:34.883750 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882679 2577 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 17 20:44:34.883750 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882683 2577 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 17 20:44:34.883750 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882686 2577 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 17 20:44:34.883750 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882691 2577 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 17 20:44:34.883750 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882695 2577 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 17 20:44:34.883750 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882699 2577 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 17 20:44:34.883750 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882703 2577 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 17 20:44:34.883750 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882708 2577 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 17 20:44:34.883750 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882711 2577 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 17 20:44:34.883750 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882715 2577 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 17 20:44:34.883750 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882719 2577 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 17 20:44:34.883750 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882723 2577 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 17 20:44:34.883750 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882727 2577 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 17 20:44:34.883750 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882732 2577 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 17 20:44:34.883750 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882736 2577 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 17 20:44:34.883750 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882740 2577 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 17 20:44:34.883750 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882744 2577 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 17 20:44:34.883750 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882751 2577 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 17 20:44:34.883750 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882758 2577 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 17 20:44:34.884430 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882762 2577 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 17 20:44:34.884430 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882768 2577 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 17 20:44:34.884430 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882772 2577 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 17 20:44:34.884430 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882777 2577 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 17 20:44:34.884430 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882781 2577 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 17 20:44:34.884430 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882785 2577 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 17 20:44:34.884430 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882789 2577 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 17 20:44:34.884430 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882793 2577 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 17 20:44:34.884430 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882797 2577 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 17 20:44:34.884430 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882820 2577 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 17 20:44:34.884430 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882827 2577 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 17 20:44:34.884430 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882833 2577 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 17 20:44:34.884430 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882837 2577 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 17 20:44:34.884430 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882841 2577 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 17 20:44:34.884430 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882846 2577 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 17 20:44:34.884430 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882849 2577 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 17 20:44:34.884430 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882853 2577 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 17 20:44:34.884430 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882857 2577 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 17 20:44:34.884430 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882862 2577 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 17 20:44:34.884430 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882866 2577 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 17 20:44:34.885076 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882870 2577 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 17 20:44:34.885076 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882874 2577 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 17 20:44:34.885076 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882878 2577 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 17 20:44:34.885076 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882883 2577 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 17 20:44:34.885076 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882887 2577 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 17 20:44:34.885076 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882890 2577 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 17 20:44:34.885076 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882895 2577 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 17 20:44:34.885076 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882899 2577 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 17 20:44:34.885076 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882904 2577 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 17 20:44:34.885076 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882908 2577 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 17 20:44:34.885076 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882911 2577 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 17 20:44:34.885076 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882915 2577 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 17 20:44:34.885076 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882919 2577 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 17 20:44:34.885076 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882923 2577 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 17 20:44:34.885076 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882928 2577 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 17 20:44:34.885076 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882932 2577 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 17 20:44:34.885076 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882937 2577 feature_gate.go:328] unrecognized feature gate: Example Apr 17 20:44:34.885076 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882941 2577 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 17 20:44:34.885076 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882945 2577 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 17 20:44:34.885076 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882948 2577 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 17 20:44:34.885664 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.882952 2577 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 17 20:44:34.885664 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.882960 2577 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 17 20:44:34.885664 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883111 2577 feature_gate.go:328] unrecognized feature gate: OVNObservability Apr 17 20:44:34.885664 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883119 2577 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Apr 17 20:44:34.885664 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883123 2577 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Apr 17 20:44:34.885664 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883128 2577 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Apr 17 20:44:34.885664 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883132 2577 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Apr 17 20:44:34.885664 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883136 2577 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Apr 17 20:44:34.885664 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883140 2577 feature_gate.go:328] unrecognized feature gate: GatewayAPI Apr 17 20:44:34.885664 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883146 2577 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Apr 17 20:44:34.885664 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883152 2577 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Apr 17 20:44:34.885664 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883157 2577 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Apr 17 20:44:34.885664 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883161 2577 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Apr 17 20:44:34.885664 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883166 2577 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Apr 17 20:44:34.885664 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883170 2577 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Apr 17 20:44:34.886068 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883175 2577 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Apr 17 20:44:34.886068 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883180 2577 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Apr 17 20:44:34.886068 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883185 2577 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Apr 17 20:44:34.886068 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883189 2577 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Apr 17 20:44:34.886068 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883194 2577 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Apr 17 20:44:34.886068 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883198 2577 feature_gate.go:328] unrecognized feature gate: NewOLM Apr 17 20:44:34.886068 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883201 2577 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Apr 17 20:44:34.886068 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883205 2577 feature_gate.go:328] unrecognized feature gate: Example Apr 17 20:44:34.886068 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883209 2577 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Apr 17 20:44:34.886068 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883213 2577 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Apr 17 20:44:34.886068 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883217 2577 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Apr 17 20:44:34.886068 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883221 2577 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Apr 17 20:44:34.886068 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883226 2577 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Apr 17 20:44:34.886068 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883230 2577 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Apr 17 20:44:34.886068 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883234 2577 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Apr 17 20:44:34.886068 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883238 2577 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Apr 17 20:44:34.886068 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883242 2577 feature_gate.go:328] unrecognized feature gate: SignatureStores Apr 17 20:44:34.886068 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883246 2577 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Apr 17 20:44:34.886068 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883251 2577 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Apr 17 20:44:34.886660 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883256 2577 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Apr 17 20:44:34.886660 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883261 2577 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Apr 17 20:44:34.886660 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883265 2577 feature_gate.go:328] unrecognized feature gate: InsightsConfig Apr 17 20:44:34.886660 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883268 2577 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Apr 17 20:44:34.886660 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883272 2577 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Apr 17 20:44:34.886660 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883276 2577 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Apr 17 20:44:34.886660 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883281 2577 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Apr 17 20:44:34.886660 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883285 2577 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Apr 17 20:44:34.886660 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883289 2577 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Apr 17 20:44:34.886660 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883293 2577 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Apr 17 20:44:34.886660 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883296 2577 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Apr 17 20:44:34.886660 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883300 2577 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Apr 17 20:44:34.886660 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883304 2577 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Apr 17 20:44:34.886660 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883308 2577 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Apr 17 20:44:34.886660 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883312 2577 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Apr 17 20:44:34.886660 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883316 2577 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Apr 17 20:44:34.886660 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883320 2577 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Apr 17 20:44:34.886660 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883325 2577 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Apr 17 20:44:34.886660 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883329 2577 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Apr 17 20:44:34.886660 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883333 2577 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Apr 17 20:44:34.887473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883336 2577 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Apr 17 20:44:34.887473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883340 2577 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Apr 17 20:44:34.887473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883345 2577 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Apr 17 20:44:34.887473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883348 2577 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Apr 17 20:44:34.887473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883352 2577 feature_gate.go:328] unrecognized feature gate: PinnedImages Apr 17 20:44:34.887473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883356 2577 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Apr 17 20:44:34.887473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883361 2577 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Apr 17 20:44:34.887473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883365 2577 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Apr 17 20:44:34.887473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883369 2577 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Apr 17 20:44:34.887473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883373 2577 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Apr 17 20:44:34.887473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883377 2577 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Apr 17 20:44:34.887473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883381 2577 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Apr 17 20:44:34.887473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883384 2577 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Apr 17 20:44:34.887473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883388 2577 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Apr 17 20:44:34.887473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883393 2577 feature_gate.go:328] unrecognized feature gate: Example2 Apr 17 20:44:34.887473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883397 2577 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Apr 17 20:44:34.887473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883401 2577 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Apr 17 20:44:34.887473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883405 2577 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Apr 17 20:44:34.887473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883409 2577 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Apr 17 20:44:34.887473 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883413 2577 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Apr 17 20:44:34.888060 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883417 2577 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Apr 17 20:44:34.888060 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883421 2577 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Apr 17 20:44:34.888060 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883425 2577 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Apr 17 20:44:34.888060 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883429 2577 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Apr 17 20:44:34.888060 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883433 2577 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Apr 17 20:44:34.888060 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883437 2577 feature_gate.go:328] unrecognized feature gate: DualReplica Apr 17 20:44:34.888060 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883440 2577 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Apr 17 20:44:34.888060 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883444 2577 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Apr 17 20:44:34.888060 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883449 2577 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Apr 17 20:44:34.888060 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883453 2577 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Apr 17 20:44:34.888060 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883457 2577 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Apr 17 20:44:34.888060 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883461 2577 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Apr 17 20:44:34.888060 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883465 2577 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Apr 17 20:44:34.888060 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:34.883469 2577 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Apr 17 20:44:34.888060 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.883476 2577 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Apr 17 20:44:34.888584 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.884204 2577 server.go:962] "Client rotation is on, will bootstrap in background" Apr 17 20:44:34.888584 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.888250 2577 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Apr 17 20:44:34.889407 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.889395 2577 server.go:1019] "Starting client certificate rotation" Apr 17 20:44:34.889512 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.889496 2577 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kube-apiserver-client-kubelet" Apr 17 20:44:34.889551 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.889537 2577 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kube-apiserver-client-kubelet" Apr 17 20:44:34.918795 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.918773 2577 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Apr 17 20:44:34.923787 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.923756 2577 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Apr 17 20:44:34.937323 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.937300 2577 log.go:25] "Validated CRI v1 runtime API" Apr 17 20:44:34.942887 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.942874 2577 log.go:25] "Validated CRI v1 image API" Apr 17 20:44:34.947839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.947821 2577 server.go:1452] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Apr 17 20:44:34.948815 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.948782 2577 reflector.go:430] "Caches populated" logger="kubernetes.io/kube-apiserver-client-kubelet" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Apr 17 20:44:34.952102 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.952079 2577 fs.go:135] Filesystem UUIDs: map[06331d64-56a5-4385-beb4-12687bd460a4:/dev/nvme0n1p3 40ad2d99-8d6d-4f41-a2e1-ed1f35c52f91:/dev/nvme0n1p4 7B77-95E7:/dev/nvme0n1p2] Apr 17 20:44:34.952163 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.952103 2577 fs.go:136] Filesystem partitions: map[/dev/nvme0n1p3:{mountpoint:/boot major:259 minor:3 fsType:ext4 blockSize:0} /dev/nvme0n1p4:{mountpoint:/var major:259 minor:4 fsType:xfs blockSize:0} /dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:31 fsType:tmpfs blockSize:0} composefs_0-33:{mountpoint:/ major:0 minor:33 fsType:overlay blockSize:0}] Apr 17 20:44:34.957938 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.957821 2577 manager.go:217] Machine: {Timestamp:2026-04-17 20:44:34.955559973 +0000 UTC m=+0.429338251 CPUVendorID:GenuineIntel NumCores:8 NumPhysicalCores:4 NumSockets:1 CpuFrequency:3099434 MemoryCapacity:32812171264 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:ec2a938da738d2c0ea647f678f42cd93 SystemUUID:ec2a938d-a738-d2c0-ea64-7f678f42cd93 BootID:d3be9ebf-06fa-44ba-a3d4-a5dafe794c9f Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16406085632 Type:vfs Inodes:4005392 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6562435072 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/nvme0n1p4 DeviceMajor:259 DeviceMinor:4 Capacity:128243970048 Type:vfs Inodes:62651840 HasInodes:true} {Device:composefs_0-33 DeviceMajor:0 DeviceMinor:33 Capacity:6103040 Type:vfs Inodes:18446744073709551615 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:31 Capacity:16406085632 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/nvme0n1p3 DeviceMajor:259 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[259:0:{Name:nvme0n1 Major:259 Minor:0 Size:128849018880 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:02:1f:54:15:21:ab Speed:0 Mtu:9001} {Name:ens5 MacAddress:02:1f:54:15:21:ab Speed:0 Mtu:9001} {Name:ovs-system MacAddress:f2:cc:e1:31:f4:bf Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:32812171264 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0 4] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:1 Threads:[1 5] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:2 Threads:[2 6] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:3 Threads:[3 7] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:}] Caches:[{Id:0 Size:34603008 Type:Unified Level:3}] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Apr 17 20:44:34.957938 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.957933 2577 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Apr 17 20:44:34.958051 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.958003 2577 manager.go:233] Version: {KernelVersion:5.14.0-570.107.1.el9_6.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 9.6.20260414-0 (Plow) DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Apr 17 20:44:34.959124 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.959101 2577 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 17 20:44:34.959282 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.959128 2577 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-10-0-137-110.ec2.internal","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"500m","ephemeral-storage":"1Gi","memory":"1Gi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 17 20:44:34.959364 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.959297 2577 topology_manager.go:138] "Creating topology manager with none policy" Apr 17 20:44:34.959364 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.959309 2577 container_manager_linux.go:306] "Creating device plugin manager" Apr 17 20:44:34.959364 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.959327 2577 manager.go:141] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Apr 17 20:44:34.959364 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.959354 2577 server.go:72] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Apr 17 20:44:34.960188 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.960175 2577 state_mem.go:36] "Initialized new in-memory state store" Apr 17 20:44:34.960355 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.960343 2577 server.go:1267] "Using root directory" path="/var/lib/kubelet" Apr 17 20:44:34.963324 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.963313 2577 kubelet.go:491] "Attempting to sync node with API server" Apr 17 20:44:34.963391 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.963331 2577 kubelet.go:386] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 17 20:44:34.963391 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.963347 2577 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Apr 17 20:44:34.963391 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.963360 2577 kubelet.go:397] "Adding apiserver pod source" Apr 17 20:44:34.963391 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.963387 2577 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 17 20:44:34.964510 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.964497 2577 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Apr 17 20:44:34.964613 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.964520 2577 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Apr 17 20:44:34.964732 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.964715 2577 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-xbbx8" Apr 17 20:44:34.967626 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.967612 2577 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="cri-o" version="1.33.10-2.rhaos4.20.gita4d0894.el9" apiVersion="v1" Apr 17 20:44:34.969263 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.969250 2577 kubelet.go:953] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Apr 17 20:44:34.970595 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.970583 2577 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Apr 17 20:44:34.970655 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.970601 2577 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Apr 17 20:44:34.970655 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.970612 2577 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Apr 17 20:44:34.970655 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.970621 2577 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Apr 17 20:44:34.970655 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.970627 2577 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Apr 17 20:44:34.970655 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.970634 2577 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/secret" Apr 17 20:44:34.970655 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.970639 2577 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Apr 17 20:44:34.970655 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.970645 2577 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Apr 17 20:44:34.970655 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.970652 2577 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/fc" Apr 17 20:44:34.970655 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.970657 2577 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Apr 17 20:44:34.970942 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.970683 2577 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/projected" Apr 17 20:44:34.970942 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.970694 2577 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Apr 17 20:44:34.971666 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.971640 2577 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-xbbx8" Apr 17 20:44:34.972599 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.972586 2577 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/csi" Apr 17 20:44:34.972662 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.972605 2577 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/image" Apr 17 20:44:34.975914 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:34.975894 2577 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Apr 17 20:44:34.976005 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.975991 2577 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "ip-10-0-137-110.ec2.internal" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Apr 17 20:44:34.976082 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.976070 2577 watchdog_linux.go:99] "Systemd watchdog is not enabled" Apr 17 20:44:34.976121 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.976101 2577 server.go:1295] "Started kubelet" Apr 17 20:44:34.976231 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:34.976210 2577 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: nodes \"ip-10-0-137-110.ec2.internal\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Apr 17 20:44:34.976273 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.976213 2577 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 17 20:44:34.976304 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.976272 2577 server_v1.go:47] "podresources" method="list" useActivePods=true Apr 17 20:44:34.976340 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.976269 2577 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Apr 17 20:44:34.978319 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.978283 2577 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 17 20:44:34.978326 ip-10-0-137-110 systemd[1]: Started Kubernetes Kubelet. Apr 17 20:44:34.981169 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.981143 2577 server.go:317] "Adding debug handlers to kubelet server" Apr 17 20:44:34.985543 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.985526 2577 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 17 20:44:34.985620 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.985559 2577 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kubelet-serving" Apr 17 20:44:34.986331 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.986257 2577 volume_manager.go:295] "The desired_state_of_world populator starts" Apr 17 20:44:34.986597 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.986577 2577 volume_manager.go:297] "Starting Kubelet Volume Manager" Apr 17 20:44:34.986696 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.986324 2577 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Apr 17 20:44:34.986746 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:34.986510 2577 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-137-110.ec2.internal\" not found" Apr 17 20:44:34.986787 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.986776 2577 reconstruct.go:97] "Volume reconstruction finished" Apr 17 20:44:34.986857 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.986789 2577 reconciler.go:26] "Reconciler: start to sync state" Apr 17 20:44:34.988844 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.988826 2577 reflector.go:430] "Caches populated" type="*v1.CSIDriver" reflector="k8s.io/client-go/informers/factory.go:160" Apr 17 20:44:34.988953 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.988826 2577 factory.go:221] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Apr 17 20:44:34.989013 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.988957 2577 factory.go:55] Registering systemd factory Apr 17 20:44:34.989013 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.988968 2577 factory.go:223] Registration of the systemd container factory successfully Apr 17 20:44:34.989292 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.989281 2577 factory.go:153] Registering CRI-O factory Apr 17 20:44:34.989292 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.989294 2577 factory.go:223] Registration of the crio container factory successfully Apr 17 20:44:34.989402 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.989334 2577 factory.go:103] Registering Raw factory Apr 17 20:44:34.989402 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.989348 2577 manager.go:1196] Started watching for new ooms in manager Apr 17 20:44:34.990135 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.990081 2577 manager.go:319] Starting recovery of all containers Apr 17 20:44:34.991235 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:34.991212 2577 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ip-10-0-137-110.ec2.internal\" not found" node="ip-10-0-137-110.ec2.internal" Apr 17 20:44:34.995311 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:34.995284 2577 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Apr 17 20:44:35.000638 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.000621 2577 manager.go:324] Recovery completed Apr 17 20:44:35.002360 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:35.002338 2577 watcher.go:152] Failed to watch directory "/sys/fs/cgroup/system.slice/systemd-update-utmp-runlevel.service": inotify_add_watch /sys/fs/cgroup/system.slice/systemd-update-utmp-runlevel.service: no such file or directory Apr 17 20:44:35.005057 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.005044 2577 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 17 20:44:35.007516 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.007416 2577 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-110.ec2.internal" event="NodeHasSufficientMemory" Apr 17 20:44:35.007585 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.007529 2577 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-110.ec2.internal" event="NodeHasNoDiskPressure" Apr 17 20:44:35.007585 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.007554 2577 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-110.ec2.internal" event="NodeHasSufficientPID" Apr 17 20:44:35.008018 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.008004 2577 cpu_manager.go:222] "Starting CPU manager" policy="none" Apr 17 20:44:35.008068 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.008018 2577 cpu_manager.go:223] "Reconciling" reconcilePeriod="10s" Apr 17 20:44:35.008068 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.008033 2577 state_mem.go:36] "Initialized new in-memory state store" Apr 17 20:44:35.010298 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.010285 2577 policy_none.go:49] "None policy: Start" Apr 17 20:44:35.010340 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.010306 2577 memory_manager.go:186] "Starting memorymanager" policy="None" Apr 17 20:44:35.010340 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.010321 2577 state_mem.go:35] "Initializing new in-memory state store" Apr 17 20:44:35.055353 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.055337 2577 manager.go:341] "Starting Device Plugin manager" Apr 17 20:44:35.063559 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:35.055370 2577 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Apr 17 20:44:35.063559 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.055384 2577 server.go:85] "Starting device plugin registration server" Apr 17 20:44:35.063559 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.055578 2577 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 17 20:44:35.063559 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.055588 2577 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 17 20:44:35.063559 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.055669 2577 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Apr 17 20:44:35.063559 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.055733 2577 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Apr 17 20:44:35.063559 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.055742 2577 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 17 20:44:35.063559 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:35.056117 2577 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="non-existent label \"crio-containers\"" Apr 17 20:44:35.063559 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:35.056146 2577 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ip-10-0-137-110.ec2.internal\" not found" Apr 17 20:44:35.115463 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.115411 2577 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Apr 17 20:44:35.115463 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.115438 2577 status_manager.go:230] "Starting to sync pod status with apiserver" Apr 17 20:44:35.115463 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.115453 2577 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Apr 17 20:44:35.115463 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.115462 2577 kubelet.go:2451] "Starting kubelet main sync loop" Apr 17 20:44:35.115662 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:35.115497 2577 kubelet.go:2475] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Apr 17 20:44:35.119470 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.119452 2577 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Apr 17 20:44:35.156328 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.156311 2577 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 17 20:44:35.157117 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.157104 2577 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-110.ec2.internal" event="NodeHasSufficientMemory" Apr 17 20:44:35.157205 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.157135 2577 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-110.ec2.internal" event="NodeHasNoDiskPressure" Apr 17 20:44:35.157205 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.157151 2577 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-110.ec2.internal" event="NodeHasSufficientPID" Apr 17 20:44:35.157205 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.157177 2577 kubelet_node_status.go:78] "Attempting to register node" node="ip-10-0-137-110.ec2.internal" Apr 17 20:44:35.167184 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.167158 2577 kubelet_node_status.go:81] "Successfully registered node" node="ip-10-0-137-110.ec2.internal" Apr 17 20:44:35.167273 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:35.167185 2577 kubelet_node_status.go:597] "Error updating node status, will retry" err="error getting node \"ip-10-0-137-110.ec2.internal\": node \"ip-10-0-137-110.ec2.internal\" not found" Apr 17 20:44:35.184688 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:35.184667 2577 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-137-110.ec2.internal\" not found" Apr 17 20:44:35.216020 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.215975 2577 kubelet.go:2537] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-110.ec2.internal","kube-system/kube-apiserver-proxy-ip-10-0-137-110.ec2.internal"] Apr 17 20:44:35.216073 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.216063 2577 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 17 20:44:35.216824 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.216794 2577 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-110.ec2.internal" event="NodeHasSufficientMemory" Apr 17 20:44:35.216889 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.216833 2577 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-110.ec2.internal" event="NodeHasNoDiskPressure" Apr 17 20:44:35.216889 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.216843 2577 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-110.ec2.internal" event="NodeHasSufficientPID" Apr 17 20:44:35.218251 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.218240 2577 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 17 20:44:35.218394 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.218380 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-110.ec2.internal" Apr 17 20:44:35.218447 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.218407 2577 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 17 20:44:35.218895 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.218879 2577 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-110.ec2.internal" event="NodeHasSufficientMemory" Apr 17 20:44:35.218895 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.218888 2577 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-110.ec2.internal" event="NodeHasSufficientMemory" Apr 17 20:44:35.219001 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.218909 2577 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-110.ec2.internal" event="NodeHasNoDiskPressure" Apr 17 20:44:35.219001 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.218925 2577 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-110.ec2.internal" event="NodeHasSufficientPID" Apr 17 20:44:35.219001 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.218910 2577 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-110.ec2.internal" event="NodeHasNoDiskPressure" Apr 17 20:44:35.219087 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.218999 2577 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-110.ec2.internal" event="NodeHasSufficientPID" Apr 17 20:44:35.220250 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.220237 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/kube-apiserver-proxy-ip-10-0-137-110.ec2.internal" Apr 17 20:44:35.220297 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.220260 2577 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Apr 17 20:44:35.220938 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.220922 2577 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-110.ec2.internal" event="NodeHasSufficientMemory" Apr 17 20:44:35.221008 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.220953 2577 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-110.ec2.internal" event="NodeHasNoDiskPressure" Apr 17 20:44:35.221008 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.220967 2577 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-110.ec2.internal" event="NodeHasSufficientPID" Apr 17 20:44:35.244446 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:35.244431 2577 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-10-0-137-110.ec2.internal\" not found" node="ip-10-0-137-110.ec2.internal" Apr 17 20:44:35.248838 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:35.248823 2577 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ip-10-0-137-110.ec2.internal\" not found" node="ip-10-0-137-110.ec2.internal" Apr 17 20:44:35.285508 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:35.285488 2577 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-137-110.ec2.internal\" not found" Apr 17 20:44:35.288821 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.288791 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/a8586432e8767e95c199e223475265b6-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-137-110.ec2.internal\" (UID: \"a8586432e8767e95c199e223475265b6\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-110.ec2.internal" Apr 17 20:44:35.288886 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.288830 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/a8586432e8767e95c199e223475265b6-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-137-110.ec2.internal\" (UID: \"a8586432e8767e95c199e223475265b6\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-110.ec2.internal" Apr 17 20:44:35.288886 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.288847 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/3114a7bb68dc8d0f555ef7f4a1089352-config\") pod \"kube-apiserver-proxy-ip-10-0-137-110.ec2.internal\" (UID: \"3114a7bb68dc8d0f555ef7f4a1089352\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-137-110.ec2.internal" Apr 17 20:44:35.386322 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:35.386276 2577 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-137-110.ec2.internal\" not found" Apr 17 20:44:35.389595 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.389581 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/a8586432e8767e95c199e223475265b6-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-137-110.ec2.internal\" (UID: \"a8586432e8767e95c199e223475265b6\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-110.ec2.internal" Apr 17 20:44:35.389665 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.389607 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/a8586432e8767e95c199e223475265b6-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-137-110.ec2.internal\" (UID: \"a8586432e8767e95c199e223475265b6\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-110.ec2.internal" Apr 17 20:44:35.389722 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.389661 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/3114a7bb68dc8d0f555ef7f4a1089352-config\") pod \"kube-apiserver-proxy-ip-10-0-137-110.ec2.internal\" (UID: \"3114a7bb68dc8d0f555ef7f4a1089352\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-137-110.ec2.internal" Apr 17 20:44:35.389722 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.389688 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/a8586432e8767e95c199e223475265b6-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-137-110.ec2.internal\" (UID: \"a8586432e8767e95c199e223475265b6\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-110.ec2.internal" Apr 17 20:44:35.389824 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.389734 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/a8586432e8767e95c199e223475265b6-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-137-110.ec2.internal\" (UID: \"a8586432e8767e95c199e223475265b6\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-110.ec2.internal" Apr 17 20:44:35.389824 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.389752 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/host-path/3114a7bb68dc8d0f555ef7f4a1089352-config\") pod \"kube-apiserver-proxy-ip-10-0-137-110.ec2.internal\" (UID: \"3114a7bb68dc8d0f555ef7f4a1089352\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-137-110.ec2.internal" Apr 17 20:44:35.487047 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:35.487024 2577 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-137-110.ec2.internal\" not found" Apr 17 20:44:35.546271 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.546256 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-110.ec2.internal" Apr 17 20:44:35.550533 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.550517 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/kube-apiserver-proxy-ip-10-0-137-110.ec2.internal" Apr 17 20:44:35.588087 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:35.588069 2577 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-137-110.ec2.internal\" not found" Apr 17 20:44:35.688588 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:35.688535 2577 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-137-110.ec2.internal\" not found" Apr 17 20:44:35.789082 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:35.789061 2577 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-137-110.ec2.internal\" not found" Apr 17 20:44:35.888567 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.888549 2577 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Apr 17 20:44:35.888999 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.888668 2577 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" err="very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received" Apr 17 20:44:35.888999 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.888722 2577 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" err="very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received" Apr 17 20:44:35.889624 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:35.889604 2577 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-137-110.ec2.internal\" not found" Apr 17 20:44:35.976477 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.976433 2577 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kube-apiserver-client-kubelet" expiration="2028-04-16 20:39:34 +0000 UTC" deadline="2027-12-23 21:13:15.300586293 +0000 UTC" Apr 17 20:44:35.976477 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.976478 2577 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kube-apiserver-client-kubelet" sleep="14760h28m39.324112475s" Apr 17 20:44:35.986536 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:35.986518 2577 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kubelet-serving" Apr 17 20:44:35.990219 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:35.990192 2577 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-137-110.ec2.internal\" not found" Apr 17 20:44:36.002033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.002012 2577 reflector.go:430] "Caches populated" logger="kubernetes.io/kubelet-serving" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Apr 17 20:44:36.020000 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.019978 2577 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kubelet-serving" csr="csr-b7l56" Apr 17 20:44:36.027121 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.027101 2577 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kubelet-serving" csr="csr-b7l56" Apr 17 20:44:36.046161 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:36.046136 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8586432e8767e95c199e223475265b6.slice/crio-5acd9ada18a4230dcc715048f2d6a795f23e559c456afff3254da35bd1984db2 WatchSource:0}: Error finding container 5acd9ada18a4230dcc715048f2d6a795f23e559c456afff3254da35bd1984db2: Status 404 returned error can't find the container with id 5acd9ada18a4230dcc715048f2d6a795f23e559c456afff3254da35bd1984db2 Apr 17 20:44:36.046697 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:36.046677 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3114a7bb68dc8d0f555ef7f4a1089352.slice/crio-ff5911c340104c4e24078ca60d3b73804c2f2ef703e31ae67dc812d3b0071f76 WatchSource:0}: Error finding container ff5911c340104c4e24078ca60d3b73804c2f2ef703e31ae67dc812d3b0071f76: Status 404 returned error can't find the container with id ff5911c340104c4e24078ca60d3b73804c2f2ef703e31ae67dc812d3b0071f76 Apr 17 20:44:36.050683 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.050669 2577 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 17 20:44:36.090856 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:36.090836 2577 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-137-110.ec2.internal\" not found" Apr 17 20:44:36.117835 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.117778 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-110.ec2.internal" event={"ID":"a8586432e8767e95c199e223475265b6","Type":"ContainerStarted","Data":"5acd9ada18a4230dcc715048f2d6a795f23e559c456afff3254da35bd1984db2"} Apr 17 20:44:36.118668 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.118650 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-137-110.ec2.internal" event={"ID":"3114a7bb68dc8d0f555ef7f4a1089352","Type":"ContainerStarted","Data":"ff5911c340104c4e24078ca60d3b73804c2f2ef703e31ae67dc812d3b0071f76"} Apr 17 20:44:36.191050 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:36.191032 2577 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-137-110.ec2.internal\" not found" Apr 17 20:44:36.291612 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:36.291570 2577 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-137-110.ec2.internal\" not found" Apr 17 20:44:36.362726 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.362706 2577 reflector.go:430] "Caches populated" type="*v1.Node" reflector="k8s.io/client-go/informers/factory.go:160" Apr 17 20:44:36.386057 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.386032 2577 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-137-110.ec2.internal" Apr 17 20:44:36.395339 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.395322 2577 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 17 20:44:36.397323 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.397311 2577 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-110.ec2.internal" Apr 17 20:44:36.410525 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.410509 2577 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Apr 17 20:44:36.493944 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.493926 2577 reflector.go:430] "Caches populated" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" Apr 17 20:44:36.965017 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.964993 2577 apiserver.go:52] "Watching apiserver" Apr 17 20:44:36.970915 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.970887 2577 reflector.go:430] "Caches populated" type="*v1.Pod" reflector="pkg/kubelet/config/apiserver.go:66" Apr 17 20:44:36.971294 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.971272 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kube-system/kube-apiserver-proxy-ip-10-0-137-110.ec2.internal","openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np","openshift-cluster-node-tuning-operator/tuned-4898w","openshift-image-registry/node-ca-v8fbw","openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-110.ec2.internal","openshift-multus/multus-2lsfw","openshift-multus/multus-additional-cni-plugins-zds6c","openshift-dns/node-resolver-cwp9d","openshift-multus/network-metrics-daemon-7ddt5","openshift-network-diagnostics/network-check-target-cm2lg","openshift-network-operator/iptables-alerter-mkzlr","openshift-ovn-kubernetes/ovnkube-node-dk97h","kube-system/konnectivity-agent-cxc79"] Apr 17 20:44:36.972899 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.972877 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-zds6c" Apr 17 20:44:36.974101 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.974081 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" Apr 17 20:44:36.974723 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.974696 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"whereabouts-flatfile-config\"" Apr 17 20:44:36.974834 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.974782 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"openshift-service-ca.crt\"" Apr 17 20:44:36.975321 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.975279 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"default-cni-sysctl-allowlist\"" Apr 17 20:44:36.975728 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.975709 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-csi-drivers\"/\"aws-ebs-csi-driver-node-metrics-serving-cert\"" Apr 17 20:44:36.975839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.975764 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-8dbbt\"" Apr 17 20:44:36.975907 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.975891 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"kube-root-ca.crt\"" Apr 17 20:44:36.977070 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.976353 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"cni-copy-resources\"" Apr 17 20:44:36.977070 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.976536 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-csi-drivers\"/\"kube-root-ca.crt\"" Apr 17 20:44:36.977070 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.976677 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-csi-drivers\"/\"aws-ebs-csi-driver-node-sa-dockercfg-7zmnk\"" Apr 17 20:44:36.977070 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.976970 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-csi-drivers\"/\"openshift-service-ca.crt\"" Apr 17 20:44:36.977958 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.977648 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:36.977958 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.977751 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-v8fbw" Apr 17 20:44:36.979327 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.979309 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"openshift-service-ca.crt\"" Apr 17 20:44:36.979673 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.979656 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"node-ca-dockercfg-s5rtq\"" Apr 17 20:44:36.980301 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.980286 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-2lsfw" Apr 17 20:44:36.980492 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.980474 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"image-registry-certificates\"" Apr 17 20:44:36.980587 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.980572 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"openshift-service-ca.crt\"" Apr 17 20:44:36.980679 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.980662 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"kube-root-ca.crt\"" Apr 17 20:44:36.980819 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.980770 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"tuned-dockercfg-kbs4v\"" Apr 17 20:44:36.980918 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.980770 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"kube-root-ca.crt\"" Apr 17 20:44:36.981896 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.981556 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-cwp9d" Apr 17 20:44:36.981896 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.981676 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:44:36.981896 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:36.981788 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7ddt5" podUID="1a6721ad-2d96-4d01-a9db-b8654256e62f" Apr 17 20:44:36.982454 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.982428 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"multus-daemon-config\"" Apr 17 20:44:36.982568 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.982540 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"default-dockercfg-4bp6l\"" Apr 17 20:44:36.982845 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.982830 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:44:36.982943 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:36.982917 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cm2lg" podUID="ac379788-302d-4181-8970-56f47ae312c1" Apr 17 20:44:36.983010 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.982968 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"openshift-service-ca.crt\"" Apr 17 20:44:36.983974 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.983952 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-mkzlr" Apr 17 20:44:36.984111 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.984090 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"kube-root-ca.crt\"" Apr 17 20:44:36.984375 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.984359 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"node-resolver-dockercfg-694gh\"" Apr 17 20:44:36.985434 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.985416 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:36.986014 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.985996 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"kube-root-ca.crt\"" Apr 17 20:44:36.986102 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.986076 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"openshift-service-ca.crt\"" Apr 17 20:44:36.986151 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.986130 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-dockercfg-xdkzp\"" Apr 17 20:44:36.986334 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.986317 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-script\"" Apr 17 20:44:36.986570 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.986552 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/konnectivity-agent-cxc79" Apr 17 20:44:36.987198 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.987180 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-node-dockercfg-rgfn6\"" Apr 17 20:44:36.987289 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.987213 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-config\"" Apr 17 20:44:36.987449 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.987416 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"openshift-service-ca.crt\"" Apr 17 20:44:36.987539 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.987466 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"env-overrides\"" Apr 17 20:44:36.987886 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.987865 2577 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Apr 17 20:44:36.987978 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.987956 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"kube-root-ca.crt\"" Apr 17 20:44:36.988036 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.988002 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-node-metrics-cert\"" Apr 17 20:44:36.988347 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.988330 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-script-lib\"" Apr 17 20:44:36.988430 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.988332 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kube-system\"/\"konnectivity-ca-bundle\"" Apr 17 20:44:36.988430 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.988408 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"konnectivity-agent\"" Apr 17 20:44:36.988537 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.988489 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"default-dockercfg-spbq6\"" Apr 17 20:44:36.997839 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.997796 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/06332f89-1179-4cde-9a2b-4335714933a1-cni-binary-copy\") pod \"multus-additional-cni-plugins-zds6c\" (UID: \"06332f89-1179-4cde-9a2b-4335714933a1\") " pod="openshift-multus/multus-additional-cni-plugins-zds6c" Apr 17 20:44:36.997952 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.997854 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/83de7e89-23a8-4a38-833c-4fa3b9c105ca-registration-dir\") pod \"aws-ebs-csi-driver-node-fj8np\" (UID: \"83de7e89-23a8-4a38-833c-4fa3b9c105ca\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" Apr 17 20:44:36.997952 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.997903 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-var-lib-kubelet\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:36.997952 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.997922 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/48141b3e-9ad0-4dca-a6d1-055ed87fba90-etc-tuned\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:36.997952 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.997937 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-multus-conf-dir\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:36.998156 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.997955 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-host-run-multus-certs\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:36.998156 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.997977 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/06332f89-1179-4cde-9a2b-4335714933a1-os-release\") pod \"multus-additional-cni-plugins-zds6c\" (UID: \"06332f89-1179-4cde-9a2b-4335714933a1\") " pod="openshift-multus/multus-additional-cni-plugins-zds6c" Apr 17 20:44:36.998156 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.997998 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-etc-openvswitch\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:36.998156 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998020 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-run-openvswitch\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:36.998156 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998094 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1a6721ad-2d96-4d01-a9db-b8654256e62f-metrics-certs\") pod \"network-metrics-daemon-7ddt5\" (UID: \"1a6721ad-2d96-4d01-a9db-b8654256e62f\") " pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:44:36.998386 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998158 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wncm7\" (UniqueName: \"kubernetes.io/projected/1a6721ad-2d96-4d01-a9db-b8654256e62f-kube-api-access-wncm7\") pod \"network-metrics-daemon-7ddt5\" (UID: \"1a6721ad-2d96-4d01-a9db-b8654256e62f\") " pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:44:36.998386 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998188 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-ovnkube-script-lib\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:36.998386 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998217 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-etc-sysctl-d\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:36.998386 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998245 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-multus-cni-dir\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:36.998386 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998270 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4b68ff3e-8a24-45e1-a70f-83f0539a9623-multus-daemon-config\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:36.998386 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998296 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxmfs\" (UniqueName: \"kubernetes.io/projected/06332f89-1179-4cde-9a2b-4335714933a1-kube-api-access-gxmfs\") pod \"multus-additional-cni-plugins-zds6c\" (UID: \"06332f89-1179-4cde-9a2b-4335714933a1\") " pod="openshift-multus/multus-additional-cni-plugins-zds6c" Apr 17 20:44:36.998386 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998323 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-host-run-ovn-kubernetes\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:36.998386 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998347 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-host-cni-bin\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:36.998386 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998368 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-run\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:36.998819 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998391 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/06332f89-1179-4cde-9a2b-4335714933a1-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zds6c\" (UID: \"06332f89-1179-4cde-9a2b-4335714933a1\") " pod="openshift-multus/multus-additional-cni-plugins-zds6c" Apr 17 20:44:36.998819 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998414 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-systemd-units\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:36.998819 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998436 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:36.998819 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998499 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/906a412c-f8af-4800-9e64-23e1b3be650e-konnectivity-ca\") pod \"konnectivity-agent-cxc79\" (UID: \"906a412c-f8af-4800-9e64-23e1b3be650e\") " pod="kube-system/konnectivity-agent-cxc79" Apr 17 20:44:36.998819 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998546 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-etc-modprobe-d\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:36.998819 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998563 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-host\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:36.998819 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998577 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4b68ff3e-8a24-45e1-a70f-83f0539a9623-cni-binary-copy\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:36.998819 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998596 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-host-run-netns\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:36.998819 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998615 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-env-overrides\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:36.998819 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998629 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5fa81115-694f-4f2f-8797-8962d102b7e7-host\") pod \"node-ca-v8fbw\" (UID: \"5fa81115-694f-4f2f-8797-8962d102b7e7\") " pod="openshift-image-registry/node-ca-v8fbw" Apr 17 20:44:36.998819 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998656 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4d52\" (UniqueName: \"kubernetes.io/projected/5fa81115-694f-4f2f-8797-8962d102b7e7-kube-api-access-m4d52\") pod \"node-ca-v8fbw\" (UID: \"5fa81115-694f-4f2f-8797-8962d102b7e7\") " pod="openshift-image-registry/node-ca-v8fbw" Apr 17 20:44:36.998819 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998707 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-cnibin\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:36.998819 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998730 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-os-release\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:36.998819 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998750 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-node-log\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:36.998819 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998765 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-ovnkube-config\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:36.998819 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998779 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/83de7e89-23a8-4a38-833c-4fa3b9c105ca-etc-selinux\") pod \"aws-ebs-csi-driver-node-fj8np\" (UID: \"83de7e89-23a8-4a38-833c-4fa3b9c105ca\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" Apr 17 20:44:36.999300 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998826 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szhzf\" (UniqueName: \"kubernetes.io/projected/83de7e89-23a8-4a38-833c-4fa3b9c105ca-kube-api-access-szhzf\") pod \"aws-ebs-csi-driver-node-fj8np\" (UID: \"83de7e89-23a8-4a38-833c-4fa3b9c105ca\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" Apr 17 20:44:36.999300 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998845 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-etc-systemd\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:36.999300 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998858 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-host-kubelet\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:36.999300 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998872 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cctxf\" (UniqueName: \"kubernetes.io/projected/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-kube-api-access-cctxf\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:36.999300 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998888 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-host-var-lib-cni-bin\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:36.999300 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998917 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-host-var-lib-kubelet\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:36.999300 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998935 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-etc-kubernetes\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:36.999300 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998949 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c9f9af16-3b9f-4ebb-8025-9b9162c7eaa1-hosts-file\") pod \"node-resolver-cwp9d\" (UID: \"c9f9af16-3b9f-4ebb-8025-9b9162c7eaa1\") " pod="openshift-dns/node-resolver-cwp9d" Apr 17 20:44:36.999300 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998963 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79v97\" (UniqueName: \"kubernetes.io/projected/ac379788-302d-4181-8970-56f47ae312c1-kube-api-access-79v97\") pod \"network-check-target-cm2lg\" (UID: \"ac379788-302d-4181-8970-56f47ae312c1\") " pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:44:36.999300 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.998994 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/83de7e89-23a8-4a38-833c-4fa3b9c105ca-kubelet-dir\") pod \"aws-ebs-csi-driver-node-fj8np\" (UID: \"83de7e89-23a8-4a38-833c-4fa3b9c105ca\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" Apr 17 20:44:36.999300 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999013 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24w7l\" (UniqueName: \"kubernetes.io/projected/c9f9af16-3b9f-4ebb-8025-9b9162c7eaa1-kube-api-access-24w7l\") pod \"node-resolver-cwp9d\" (UID: \"c9f9af16-3b9f-4ebb-8025-9b9162c7eaa1\") " pod="openshift-dns/node-resolver-cwp9d" Apr 17 20:44:36.999300 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999033 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/06332f89-1179-4cde-9a2b-4335714933a1-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-zds6c\" (UID: \"06332f89-1179-4cde-9a2b-4335714933a1\") " pod="openshift-multus/multus-additional-cni-plugins-zds6c" Apr 17 20:44:36.999300 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999056 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-run-ovn\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:36.999300 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999076 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/83de7e89-23a8-4a38-833c-4fa3b9c105ca-sys-fs\") pod \"aws-ebs-csi-driver-node-fj8np\" (UID: \"83de7e89-23a8-4a38-833c-4fa3b9c105ca\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" Apr 17 20:44:36.999300 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999089 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-etc-sysctl-conf\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:36.999300 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999105 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5fa81115-694f-4f2f-8797-8962d102b7e7-serviceca\") pod \"node-ca-v8fbw\" (UID: \"5fa81115-694f-4f2f-8797-8962d102b7e7\") " pod="openshift-image-registry/node-ca-v8fbw" Apr 17 20:44:36.999923 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999143 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxhbh\" (UniqueName: \"kubernetes.io/projected/4b68ff3e-8a24-45e1-a70f-83f0539a9623-kube-api-access-lxhbh\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:36.999923 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999180 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/06332f89-1179-4cde-9a2b-4335714933a1-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zds6c\" (UID: \"06332f89-1179-4cde-9a2b-4335714933a1\") " pod="openshift-multus/multus-additional-cni-plugins-zds6c" Apr 17 20:44:36.999923 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999212 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-host-cni-netd\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:36.999923 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999235 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-sys\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:36.999923 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999258 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-host-run-k8s-cni-cncf-io\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:36.999923 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999283 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-host-var-lib-cni-multus\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:36.999923 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999307 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/c9f9af16-3b9f-4ebb-8025-9b9162c7eaa1-tmp-dir\") pod \"node-resolver-cwp9d\" (UID: \"c9f9af16-3b9f-4ebb-8025-9b9162c7eaa1\") " pod="openshift-dns/node-resolver-cwp9d" Apr 17 20:44:36.999923 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999330 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-log-socket\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:36.999923 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999352 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-ovn-node-metrics-cert\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:36.999923 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999373 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-etc-kubernetes\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:36.999923 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999412 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/06332f89-1179-4cde-9a2b-4335714933a1-system-cni-dir\") pod \"multus-additional-cni-plugins-zds6c\" (UID: \"06332f89-1179-4cde-9a2b-4335714933a1\") " pod="openshift-multus/multus-additional-cni-plugins-zds6c" Apr 17 20:44:36.999923 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999439 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-host-slash\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:36.999923 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999473 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-run-systemd\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:36.999923 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999508 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-var-lib-openvswitch\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:36.999923 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999536 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qhhh\" (UniqueName: \"kubernetes.io/projected/2e78f338-82b0-4062-8438-2b45fae0f4cf-kube-api-access-2qhhh\") pod \"iptables-alerter-mkzlr\" (UID: \"2e78f338-82b0-4062-8438-2b45fae0f4cf\") " pod="openshift-network-operator/iptables-alerter-mkzlr" Apr 17 20:44:36.999923 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999570 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktj7v\" (UniqueName: \"kubernetes.io/projected/48141b3e-9ad0-4dca-a6d1-055ed87fba90-kube-api-access-ktj7v\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.000648 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999600 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-hostroot\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.000648 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999647 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2e78f338-82b0-4062-8438-2b45fae0f4cf-host-slash\") pod \"iptables-alerter-mkzlr\" (UID: \"2e78f338-82b0-4062-8438-2b45fae0f4cf\") " pod="openshift-network-operator/iptables-alerter-mkzlr" Apr 17 20:44:37.000648 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999676 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/06332f89-1179-4cde-9a2b-4335714933a1-cnibin\") pod \"multus-additional-cni-plugins-zds6c\" (UID: \"06332f89-1179-4cde-9a2b-4335714933a1\") " pod="openshift-multus/multus-additional-cni-plugins-zds6c" Apr 17 20:44:37.000648 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999699 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-lib-modules\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.000648 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999728 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/48141b3e-9ad0-4dca-a6d1-055ed87fba90-tmp\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.000648 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999750 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/2e78f338-82b0-4062-8438-2b45fae0f4cf-iptables-alerter-script\") pod \"iptables-alerter-mkzlr\" (UID: \"2e78f338-82b0-4062-8438-2b45fae0f4cf\") " pod="openshift-network-operator/iptables-alerter-mkzlr" Apr 17 20:44:37.000648 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999770 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/906a412c-f8af-4800-9e64-23e1b3be650e-agent-certs\") pod \"konnectivity-agent-cxc79\" (UID: \"906a412c-f8af-4800-9e64-23e1b3be650e\") " pod="kube-system/konnectivity-agent-cxc79" Apr 17 20:44:37.000648 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999788 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/83de7e89-23a8-4a38-833c-4fa3b9c105ca-socket-dir\") pod \"aws-ebs-csi-driver-node-fj8np\" (UID: \"83de7e89-23a8-4a38-833c-4fa3b9c105ca\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" Apr 17 20:44:37.000648 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999826 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/83de7e89-23a8-4a38-833c-4fa3b9c105ca-device-dir\") pod \"aws-ebs-csi-driver-node-fj8np\" (UID: \"83de7e89-23a8-4a38-833c-4fa3b9c105ca\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" Apr 17 20:44:37.000648 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999843 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-etc-sysconfig\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.000648 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999856 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-system-cni-dir\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.000648 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999869 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-multus-socket-dir-parent\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.000648 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:36.999890 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-host-run-netns\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.028560 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.028536 2577 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kubelet-serving" expiration="2028-04-16 20:39:36 +0000 UTC" deadline="2027-09-11 13:27:02.141397535 +0000 UTC" Apr 17 20:44:37.028560 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.028558 2577 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kubelet-serving" sleep="12280h42m25.112841604s" Apr 17 20:44:37.054568 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.054548 2577 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Apr 17 20:44:37.100761 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.100742 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-2qhhh\" (UniqueName: \"kubernetes.io/projected/2e78f338-82b0-4062-8438-2b45fae0f4cf-kube-api-access-2qhhh\") pod \"iptables-alerter-mkzlr\" (UID: \"2e78f338-82b0-4062-8438-2b45fae0f4cf\") " pod="openshift-network-operator/iptables-alerter-mkzlr" Apr 17 20:44:37.100874 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.100776 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-ktj7v\" (UniqueName: \"kubernetes.io/projected/48141b3e-9ad0-4dca-a6d1-055ed87fba90-kube-api-access-ktj7v\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.100874 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.100824 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-hostroot\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.100874 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.100849 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2e78f338-82b0-4062-8438-2b45fae0f4cf-host-slash\") pod \"iptables-alerter-mkzlr\" (UID: \"2e78f338-82b0-4062-8438-2b45fae0f4cf\") " pod="openshift-network-operator/iptables-alerter-mkzlr" Apr 17 20:44:37.101035 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.100906 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-hostroot\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.101035 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.100923 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2e78f338-82b0-4062-8438-2b45fae0f4cf-host-slash\") pod \"iptables-alerter-mkzlr\" (UID: \"2e78f338-82b0-4062-8438-2b45fae0f4cf\") " pod="openshift-network-operator/iptables-alerter-mkzlr" Apr 17 20:44:37.101035 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.100953 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/06332f89-1179-4cde-9a2b-4335714933a1-cnibin\") pod \"multus-additional-cni-plugins-zds6c\" (UID: \"06332f89-1179-4cde-9a2b-4335714933a1\") " pod="openshift-multus/multus-additional-cni-plugins-zds6c" Apr 17 20:44:37.101035 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.100972 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-lib-modules\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.101035 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.100986 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/48141b3e-9ad0-4dca-a6d1-055ed87fba90-tmp\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.101035 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101008 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/2e78f338-82b0-4062-8438-2b45fae0f4cf-iptables-alerter-script\") pod \"iptables-alerter-mkzlr\" (UID: \"2e78f338-82b0-4062-8438-2b45fae0f4cf\") " pod="openshift-network-operator/iptables-alerter-mkzlr" Apr 17 20:44:37.101035 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101016 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/06332f89-1179-4cde-9a2b-4335714933a1-cnibin\") pod \"multus-additional-cni-plugins-zds6c\" (UID: \"06332f89-1179-4cde-9a2b-4335714933a1\") " pod="openshift-multus/multus-additional-cni-plugins-zds6c" Apr 17 20:44:37.101035 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101033 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/906a412c-f8af-4800-9e64-23e1b3be650e-agent-certs\") pod \"konnectivity-agent-cxc79\" (UID: \"906a412c-f8af-4800-9e64-23e1b3be650e\") " pod="kube-system/konnectivity-agent-cxc79" Apr 17 20:44:37.101385 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101057 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/83de7e89-23a8-4a38-833c-4fa3b9c105ca-socket-dir\") pod \"aws-ebs-csi-driver-node-fj8np\" (UID: \"83de7e89-23a8-4a38-833c-4fa3b9c105ca\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" Apr 17 20:44:37.101385 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101079 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/83de7e89-23a8-4a38-833c-4fa3b9c105ca-device-dir\") pod \"aws-ebs-csi-driver-node-fj8np\" (UID: \"83de7e89-23a8-4a38-833c-4fa3b9c105ca\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" Apr 17 20:44:37.101385 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101106 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-etc-sysconfig\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.101385 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101127 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-lib-modules\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.101385 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101200 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-system-cni-dir\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.101385 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101130 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-system-cni-dir\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.101385 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101234 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-multus-socket-dir-parent\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.101385 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101250 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-host-run-netns\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.101385 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101241 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/83de7e89-23a8-4a38-833c-4fa3b9c105ca-socket-dir\") pod \"aws-ebs-csi-driver-node-fj8np\" (UID: \"83de7e89-23a8-4a38-833c-4fa3b9c105ca\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" Apr 17 20:44:37.101385 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101274 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/06332f89-1179-4cde-9a2b-4335714933a1-cni-binary-copy\") pod \"multus-additional-cni-plugins-zds6c\" (UID: \"06332f89-1179-4cde-9a2b-4335714933a1\") " pod="openshift-multus/multus-additional-cni-plugins-zds6c" Apr 17 20:44:37.101385 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101296 2577 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Apr 17 20:44:37.101385 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101316 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/83de7e89-23a8-4a38-833c-4fa3b9c105ca-device-dir\") pod \"aws-ebs-csi-driver-node-fj8np\" (UID: \"83de7e89-23a8-4a38-833c-4fa3b9c105ca\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" Apr 17 20:44:37.101385 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101347 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/83de7e89-23a8-4a38-833c-4fa3b9c105ca-registration-dir\") pod \"aws-ebs-csi-driver-node-fj8np\" (UID: \"83de7e89-23a8-4a38-833c-4fa3b9c105ca\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" Apr 17 20:44:37.101385 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101362 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-etc-sysconfig\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.105995 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101395 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-multus-socket-dir-parent\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.105995 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101397 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-host-run-netns\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.105995 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101300 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/83de7e89-23a8-4a38-833c-4fa3b9c105ca-registration-dir\") pod \"aws-ebs-csi-driver-node-fj8np\" (UID: \"83de7e89-23a8-4a38-833c-4fa3b9c105ca\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" Apr 17 20:44:37.105995 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101430 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-var-lib-kubelet\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.105995 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101444 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/48141b3e-9ad0-4dca-a6d1-055ed87fba90-etc-tuned\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.105995 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101458 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-multus-conf-dir\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.105995 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101502 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-host-run-multus-certs\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.105995 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101527 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/06332f89-1179-4cde-9a2b-4335714933a1-os-release\") pod \"multus-additional-cni-plugins-zds6c\" (UID: \"06332f89-1179-4cde-9a2b-4335714933a1\") " pod="openshift-multus/multus-additional-cni-plugins-zds6c" Apr 17 20:44:37.105995 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101551 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-etc-openvswitch\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.105995 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101574 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-run-openvswitch\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.105995 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101620 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1a6721ad-2d96-4d01-a9db-b8654256e62f-metrics-certs\") pod \"network-metrics-daemon-7ddt5\" (UID: \"1a6721ad-2d96-4d01-a9db-b8654256e62f\") " pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:44:37.105995 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101646 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-wncm7\" (UniqueName: \"kubernetes.io/projected/1a6721ad-2d96-4d01-a9db-b8654256e62f-kube-api-access-wncm7\") pod \"network-metrics-daemon-7ddt5\" (UID: \"1a6721ad-2d96-4d01-a9db-b8654256e62f\") " pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:44:37.105995 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101714 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-ovnkube-script-lib\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.105995 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101744 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-etc-sysctl-d\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.105995 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101751 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/2e78f338-82b0-4062-8438-2b45fae0f4cf-iptables-alerter-script\") pod \"iptables-alerter-mkzlr\" (UID: \"2e78f338-82b0-4062-8438-2b45fae0f4cf\") " pod="openshift-network-operator/iptables-alerter-mkzlr" Apr 17 20:44:37.105995 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101768 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-multus-cni-dir\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.105995 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101783 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4b68ff3e-8a24-45e1-a70f-83f0539a9623-multus-daemon-config\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.106609 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101826 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gxmfs\" (UniqueName: \"kubernetes.io/projected/06332f89-1179-4cde-9a2b-4335714933a1-kube-api-access-gxmfs\") pod \"multus-additional-cni-plugins-zds6c\" (UID: \"06332f89-1179-4cde-9a2b-4335714933a1\") " pod="openshift-multus/multus-additional-cni-plugins-zds6c" Apr 17 20:44:37.106609 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101852 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-host-run-ovn-kubernetes\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.106609 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101891 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/06332f89-1179-4cde-9a2b-4335714933a1-cni-binary-copy\") pod \"multus-additional-cni-plugins-zds6c\" (UID: \"06332f89-1179-4cde-9a2b-4335714933a1\") " pod="openshift-multus/multus-additional-cni-plugins-zds6c" Apr 17 20:44:37.106609 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101901 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-host-cni-bin\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.106609 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101927 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-run\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.106609 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101893 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-var-lib-kubelet\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.106609 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101954 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/06332f89-1179-4cde-9a2b-4335714933a1-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zds6c\" (UID: \"06332f89-1179-4cde-9a2b-4335714933a1\") " pod="openshift-multus/multus-additional-cni-plugins-zds6c" Apr 17 20:44:37.106609 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101978 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-systemd-units\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.106609 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101976 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-multus-conf-dir\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.106609 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102009 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.106609 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102036 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/906a412c-f8af-4800-9e64-23e1b3be650e-konnectivity-ca\") pod \"konnectivity-agent-cxc79\" (UID: \"906a412c-f8af-4800-9e64-23e1b3be650e\") " pod="kube-system/konnectivity-agent-cxc79" Apr 17 20:44:37.106609 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102062 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-etc-modprobe-d\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.106609 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102084 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-host\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.106609 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102109 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4b68ff3e-8a24-45e1-a70f-83f0539a9623-cni-binary-copy\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.106609 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:37.102129 2577 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 20:44:37.106609 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102137 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-host-run-netns\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.106609 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102179 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-env-overrides\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.107259 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:37.102198 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1a6721ad-2d96-4d01-a9db-b8654256e62f-metrics-certs podName:1a6721ad-2d96-4d01-a9db-b8654256e62f nodeName:}" failed. No retries permitted until 2026-04-17 20:44:37.602171487 +0000 UTC m=+3.075949786 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1a6721ad-2d96-4d01-a9db-b8654256e62f-metrics-certs") pod "network-metrics-daemon-7ddt5" (UID: "1a6721ad-2d96-4d01-a9db-b8654256e62f") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 20:44:37.107259 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102226 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5fa81115-694f-4f2f-8797-8962d102b7e7-host\") pod \"node-ca-v8fbw\" (UID: \"5fa81115-694f-4f2f-8797-8962d102b7e7\") " pod="openshift-image-registry/node-ca-v8fbw" Apr 17 20:44:37.107259 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102250 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-m4d52\" (UniqueName: \"kubernetes.io/projected/5fa81115-694f-4f2f-8797-8962d102b7e7-kube-api-access-m4d52\") pod \"node-ca-v8fbw\" (UID: \"5fa81115-694f-4f2f-8797-8962d102b7e7\") " pod="openshift-image-registry/node-ca-v8fbw" Apr 17 20:44:37.107259 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102270 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-cnibin\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.107259 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102292 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-os-release\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.107259 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102312 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-node-log\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.107259 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102333 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-ovnkube-config\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.107259 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102357 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/83de7e89-23a8-4a38-833c-4fa3b9c105ca-etc-selinux\") pod \"aws-ebs-csi-driver-node-fj8np\" (UID: \"83de7e89-23a8-4a38-833c-4fa3b9c105ca\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" Apr 17 20:44:37.107259 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102379 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-szhzf\" (UniqueName: \"kubernetes.io/projected/83de7e89-23a8-4a38-833c-4fa3b9c105ca-kube-api-access-szhzf\") pod \"aws-ebs-csi-driver-node-fj8np\" (UID: \"83de7e89-23a8-4a38-833c-4fa3b9c105ca\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" Apr 17 20:44:37.107259 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102399 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-etc-systemd\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.107259 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102419 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-host-kubelet\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.107259 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102443 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-cctxf\" (UniqueName: \"kubernetes.io/projected/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-kube-api-access-cctxf\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.107259 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102460 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-host-run-multus-certs\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.107259 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102463 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-host-var-lib-cni-bin\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.107259 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102499 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-host-var-lib-cni-bin\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.107259 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102507 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-host-var-lib-kubelet\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.107259 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102535 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-etc-kubernetes\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.107840 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102547 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5fa81115-694f-4f2f-8797-8962d102b7e7-host\") pod \"node-ca-v8fbw\" (UID: \"5fa81115-694f-4f2f-8797-8962d102b7e7\") " pod="openshift-image-registry/node-ca-v8fbw" Apr 17 20:44:37.107840 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102561 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c9f9af16-3b9f-4ebb-8025-9b9162c7eaa1-hosts-file\") pod \"node-resolver-cwp9d\" (UID: \"c9f9af16-3b9f-4ebb-8025-9b9162c7eaa1\") " pod="openshift-dns/node-resolver-cwp9d" Apr 17 20:44:37.107840 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102589 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-79v97\" (UniqueName: \"kubernetes.io/projected/ac379788-302d-4181-8970-56f47ae312c1-kube-api-access-79v97\") pod \"network-check-target-cm2lg\" (UID: \"ac379788-302d-4181-8970-56f47ae312c1\") " pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:44:37.107840 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102572 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-env-overrides\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.107840 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102619 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/83de7e89-23a8-4a38-833c-4fa3b9c105ca-kubelet-dir\") pod \"aws-ebs-csi-driver-node-fj8np\" (UID: \"83de7e89-23a8-4a38-833c-4fa3b9c105ca\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" Apr 17 20:44:37.107840 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102647 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-24w7l\" (UniqueName: \"kubernetes.io/projected/c9f9af16-3b9f-4ebb-8025-9b9162c7eaa1-kube-api-access-24w7l\") pod \"node-resolver-cwp9d\" (UID: \"c9f9af16-3b9f-4ebb-8025-9b9162c7eaa1\") " pod="openshift-dns/node-resolver-cwp9d" Apr 17 20:44:37.107840 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102678 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/06332f89-1179-4cde-9a2b-4335714933a1-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-zds6c\" (UID: \"06332f89-1179-4cde-9a2b-4335714933a1\") " pod="openshift-multus/multus-additional-cni-plugins-zds6c" Apr 17 20:44:37.107840 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102703 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-run-ovn\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.107840 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102737 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/83de7e89-23a8-4a38-833c-4fa3b9c105ca-sys-fs\") pod \"aws-ebs-csi-driver-node-fj8np\" (UID: \"83de7e89-23a8-4a38-833c-4fa3b9c105ca\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" Apr 17 20:44:37.107840 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102740 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-etc-sysctl-d\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.107840 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102776 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-etc-sysctl-conf\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.107840 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102821 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/06332f89-1179-4cde-9a2b-4335714933a1-os-release\") pod \"multus-additional-cni-plugins-zds6c\" (UID: \"06332f89-1179-4cde-9a2b-4335714933a1\") " pod="openshift-multus/multus-additional-cni-plugins-zds6c" Apr 17 20:44:37.107840 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102844 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5fa81115-694f-4f2f-8797-8962d102b7e7-serviceca\") pod \"node-ca-v8fbw\" (UID: \"5fa81115-694f-4f2f-8797-8962d102b7e7\") " pod="openshift-image-registry/node-ca-v8fbw" Apr 17 20:44:37.107840 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102862 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-host-var-lib-kubelet\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.107840 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102894 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-lxhbh\" (UniqueName: \"kubernetes.io/projected/4b68ff3e-8a24-45e1-a70f-83f0539a9623-kube-api-access-lxhbh\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.107840 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102938 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/06332f89-1179-4cde-9a2b-4335714933a1-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zds6c\" (UID: \"06332f89-1179-4cde-9a2b-4335714933a1\") " pod="openshift-multus/multus-additional-cni-plugins-zds6c" Apr 17 20:44:37.107840 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102953 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-cnibin\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.108531 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102967 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-host-cni-netd\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.108531 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102994 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-sys\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.108531 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.103033 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-host-run-k8s-cni-cncf-io\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.108531 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.103059 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-host-var-lib-cni-multus\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.108531 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.103085 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/c9f9af16-3b9f-4ebb-8025-9b9162c7eaa1-tmp-dir\") pod \"node-resolver-cwp9d\" (UID: \"c9f9af16-3b9f-4ebb-8025-9b9162c7eaa1\") " pod="openshift-dns/node-resolver-cwp9d" Apr 17 20:44:37.108531 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.103112 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-log-socket\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.108531 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.103122 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-os-release\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.108531 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.103138 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-ovn-node-metrics-cert\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.108531 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.103169 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-etc-kubernetes\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.108531 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.103196 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/06332f89-1179-4cde-9a2b-4335714933a1-system-cni-dir\") pod \"multus-additional-cni-plugins-zds6c\" (UID: \"06332f89-1179-4cde-9a2b-4335714933a1\") " pod="openshift-multus/multus-additional-cni-plugins-zds6c" Apr 17 20:44:37.108531 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.103220 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-host-slash\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.108531 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.103245 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-run-systemd\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.108531 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.103250 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/83de7e89-23a8-4a38-833c-4fa3b9c105ca-etc-selinux\") pod \"aws-ebs-csi-driver-node-fj8np\" (UID: \"83de7e89-23a8-4a38-833c-4fa3b9c105ca\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" Apr 17 20:44:37.108531 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.103270 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-var-lib-openvswitch\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.108531 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.103483 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-etc-systemd\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.108531 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.103497 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-multus-cni-dir\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.108531 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.103530 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-host-kubelet\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.108531 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102418 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-ovnkube-script-lib\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.109379 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.103650 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/06332f89-1179-4cde-9a2b-4335714933a1-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-zds6c\" (UID: \"06332f89-1179-4cde-9a2b-4335714933a1\") " pod="openshift-multus/multus-additional-cni-plugins-zds6c" Apr 17 20:44:37.109379 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.103943 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c9f9af16-3b9f-4ebb-8025-9b9162c7eaa1-hosts-file\") pod \"node-resolver-cwp9d\" (UID: \"c9f9af16-3b9f-4ebb-8025-9b9162c7eaa1\") " pod="openshift-dns/node-resolver-cwp9d" Apr 17 20:44:37.109379 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.104162 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-ovnkube-config\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.109379 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.104223 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-run-ovn\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.109379 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.104271 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5fa81115-694f-4f2f-8797-8962d102b7e7-serviceca\") pod \"node-ca-v8fbw\" (UID: \"5fa81115-694f-4f2f-8797-8962d102b7e7\") " pod="openshift-image-registry/node-ca-v8fbw" Apr 17 20:44:37.109379 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.104513 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.109379 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.104568 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-host-run-ovn-kubernetes\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.109379 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101851 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-etc-openvswitch\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.109379 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.104616 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-host-cni-bin\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.109379 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.101896 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-run-openvswitch\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.109379 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.104671 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-run\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.109379 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.104730 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4b68ff3e-8a24-45e1-a70f-83f0539a9623-multus-daemon-config\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.109379 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.102908 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-etc-kubernetes\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.109379 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.104967 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/906a412c-f8af-4800-9e64-23e1b3be650e-konnectivity-ca\") pod \"konnectivity-agent-cxc79\" (UID: \"906a412c-f8af-4800-9e64-23e1b3be650e\") " pod="kube-system/konnectivity-agent-cxc79" Apr 17 20:44:37.109379 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.105007 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/83de7e89-23a8-4a38-833c-4fa3b9c105ca-kubelet-dir\") pod \"aws-ebs-csi-driver-node-fj8np\" (UID: \"83de7e89-23a8-4a38-833c-4fa3b9c105ca\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" Apr 17 20:44:37.109379 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.105066 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-etc-modprobe-d\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.109379 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.105122 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-host\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.109379 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.105152 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/06332f89-1179-4cde-9a2b-4335714933a1-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zds6c\" (UID: \"06332f89-1179-4cde-9a2b-4335714933a1\") " pod="openshift-multus/multus-additional-cni-plugins-zds6c" Apr 17 20:44:37.110016 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.105201 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-host-cni-netd\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.110016 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.105205 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/48141b3e-9ad0-4dca-a6d1-055ed87fba90-tmp\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.110016 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.105212 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/06332f89-1179-4cde-9a2b-4335714933a1-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zds6c\" (UID: \"06332f89-1179-4cde-9a2b-4335714933a1\") " pod="openshift-multus/multus-additional-cni-plugins-zds6c" Apr 17 20:44:37.110016 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.105251 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-sys\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.110016 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.105259 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-host-var-lib-cni-multus\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.110016 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.105286 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4b68ff3e-8a24-45e1-a70f-83f0539a9623-host-run-k8s-cni-cncf-io\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.110016 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.105316 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-systemd-units\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.110016 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.103170 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-node-log\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.110016 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.105361 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-log-socket\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.110016 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.105513 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-etc-kubernetes\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.110016 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.105532 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/c9f9af16-3b9f-4ebb-8025-9b9162c7eaa1-tmp-dir\") pod \"node-resolver-cwp9d\" (UID: \"c9f9af16-3b9f-4ebb-8025-9b9162c7eaa1\") " pod="openshift-dns/node-resolver-cwp9d" Apr 17 20:44:37.110016 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.105565 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4b68ff3e-8a24-45e1-a70f-83f0539a9623-cni-binary-copy\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.110016 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.105581 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-host-slash\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.110016 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.105607 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-run-systemd\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.110016 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.105626 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/06332f89-1179-4cde-9a2b-4335714933a1-system-cni-dir\") pod \"multus-additional-cni-plugins-zds6c\" (UID: \"06332f89-1179-4cde-9a2b-4335714933a1\") " pod="openshift-multus/multus-additional-cni-plugins-zds6c" Apr 17 20:44:37.110016 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.105616 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-host-run-netns\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.110016 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.105654 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-var-lib-openvswitch\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.110016 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.105677 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/83de7e89-23a8-4a38-833c-4fa3b9c105ca-sys-fs\") pod \"aws-ebs-csi-driver-node-fj8np\" (UID: \"83de7e89-23a8-4a38-833c-4fa3b9c105ca\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" Apr 17 20:44:37.110620 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.105731 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/48141b3e-9ad0-4dca-a6d1-055ed87fba90-etc-sysctl-conf\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.110620 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.106204 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/48141b3e-9ad0-4dca-a6d1-055ed87fba90-etc-tuned\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.110620 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.107396 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/906a412c-f8af-4800-9e64-23e1b3be650e-agent-certs\") pod \"konnectivity-agent-cxc79\" (UID: \"906a412c-f8af-4800-9e64-23e1b3be650e\") " pod="kube-system/konnectivity-agent-cxc79" Apr 17 20:44:37.110620 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.107516 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-ovn-node-metrics-cert\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.110849 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.110831 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qhhh\" (UniqueName: \"kubernetes.io/projected/2e78f338-82b0-4062-8438-2b45fae0f4cf-kube-api-access-2qhhh\") pod \"iptables-alerter-mkzlr\" (UID: \"2e78f338-82b0-4062-8438-2b45fae0f4cf\") " pod="openshift-network-operator/iptables-alerter-mkzlr" Apr 17 20:44:37.110912 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.110859 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-wncm7\" (UniqueName: \"kubernetes.io/projected/1a6721ad-2d96-4d01-a9db-b8654256e62f-kube-api-access-wncm7\") pod \"network-metrics-daemon-7ddt5\" (UID: \"1a6721ad-2d96-4d01-a9db-b8654256e62f\") " pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:44:37.110912 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.110838 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktj7v\" (UniqueName: \"kubernetes.io/projected/48141b3e-9ad0-4dca-a6d1-055ed87fba90-kube-api-access-ktj7v\") pod \"tuned-4898w\" (UID: \"48141b3e-9ad0-4dca-a6d1-055ed87fba90\") " pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.111377 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:37.111100 2577 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 17 20:44:37.111377 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:37.111123 2577 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 17 20:44:37.111377 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:37.111137 2577 projected.go:194] Error preparing data for projected volume kube-api-access-79v97 for pod openshift-network-diagnostics/network-check-target-cm2lg: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 20:44:37.111377 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:37.111194 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ac379788-302d-4181-8970-56f47ae312c1-kube-api-access-79v97 podName:ac379788-302d-4181-8970-56f47ae312c1 nodeName:}" failed. No retries permitted until 2026-04-17 20:44:37.611178012 +0000 UTC m=+3.084956280 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-79v97" (UniqueName: "kubernetes.io/projected/ac379788-302d-4181-8970-56f47ae312c1-kube-api-access-79v97") pod "network-check-target-cm2lg" (UID: "ac379788-302d-4181-8970-56f47ae312c1") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 20:44:37.113877 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.113855 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-24w7l\" (UniqueName: \"kubernetes.io/projected/c9f9af16-3b9f-4ebb-8025-9b9162c7eaa1-kube-api-access-24w7l\") pod \"node-resolver-cwp9d\" (UID: \"c9f9af16-3b9f-4ebb-8025-9b9162c7eaa1\") " pod="openshift-dns/node-resolver-cwp9d" Apr 17 20:44:37.114007 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.113881 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4d52\" (UniqueName: \"kubernetes.io/projected/5fa81115-694f-4f2f-8797-8962d102b7e7-kube-api-access-m4d52\") pod \"node-ca-v8fbw\" (UID: \"5fa81115-694f-4f2f-8797-8962d102b7e7\") " pod="openshift-image-registry/node-ca-v8fbw" Apr 17 20:44:37.114007 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.113934 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-szhzf\" (UniqueName: \"kubernetes.io/projected/83de7e89-23a8-4a38-833c-4fa3b9c105ca-kube-api-access-szhzf\") pod \"aws-ebs-csi-driver-node-fj8np\" (UID: \"83de7e89-23a8-4a38-833c-4fa3b9c105ca\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" Apr 17 20:44:37.114454 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.114431 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-cctxf\" (UniqueName: \"kubernetes.io/projected/37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0-kube-api-access-cctxf\") pod \"ovnkube-node-dk97h\" (UID: \"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.117563 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.117537 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxmfs\" (UniqueName: \"kubernetes.io/projected/06332f89-1179-4cde-9a2b-4335714933a1-kube-api-access-gxmfs\") pod \"multus-additional-cni-plugins-zds6c\" (UID: \"06332f89-1179-4cde-9a2b-4335714933a1\") " pod="openshift-multus/multus-additional-cni-plugins-zds6c" Apr 17 20:44:37.117652 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.117604 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxhbh\" (UniqueName: \"kubernetes.io/projected/4b68ff3e-8a24-45e1-a70f-83f0539a9623-kube-api-access-lxhbh\") pod \"multus-2lsfw\" (UID: \"4b68ff3e-8a24-45e1-a70f-83f0539a9623\") " pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.286850 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.286769 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-zds6c" Apr 17 20:44:37.293756 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.293728 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" Apr 17 20:44:37.301394 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.301366 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-4898w" Apr 17 20:44:37.307288 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.307062 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-v8fbw" Apr 17 20:44:37.314004 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.313981 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-2lsfw" Apr 17 20:44:37.321530 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.321512 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-cwp9d" Apr 17 20:44:37.330100 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.330083 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-mkzlr" Apr 17 20:44:37.337643 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.337623 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:44:37.343252 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.343233 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/konnectivity-agent-cxc79" Apr 17 20:44:37.352066 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.352040 2577 reflector.go:430] "Caches populated" type="*v1.CSIDriver" reflector="k8s.io/client-go/informers/factory.go:160" Apr 17 20:44:37.606488 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.606417 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1a6721ad-2d96-4d01-a9db-b8654256e62f-metrics-certs\") pod \"network-metrics-daemon-7ddt5\" (UID: \"1a6721ad-2d96-4d01-a9db-b8654256e62f\") " pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:44:37.606634 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:37.606543 2577 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 20:44:37.606634 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:37.606620 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1a6721ad-2d96-4d01-a9db-b8654256e62f-metrics-certs podName:1a6721ad-2d96-4d01-a9db-b8654256e62f nodeName:}" failed. No retries permitted until 2026-04-17 20:44:38.606600644 +0000 UTC m=+4.080378922 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1a6721ad-2d96-4d01-a9db-b8654256e62f-metrics-certs") pod "network-metrics-daemon-7ddt5" (UID: "1a6721ad-2d96-4d01-a9db-b8654256e62f") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 20:44:37.707375 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:37.707338 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-79v97\" (UniqueName: \"kubernetes.io/projected/ac379788-302d-4181-8970-56f47ae312c1-kube-api-access-79v97\") pod \"network-check-target-cm2lg\" (UID: \"ac379788-302d-4181-8970-56f47ae312c1\") " pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:44:37.707555 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:37.707494 2577 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 17 20:44:37.707555 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:37.707510 2577 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 17 20:44:37.707555 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:37.707519 2577 projected.go:194] Error preparing data for projected volume kube-api-access-79v97 for pod openshift-network-diagnostics/network-check-target-cm2lg: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 20:44:37.707716 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:37.707579 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ac379788-302d-4181-8970-56f47ae312c1-kube-api-access-79v97 podName:ac379788-302d-4181-8970-56f47ae312c1 nodeName:}" failed. No retries permitted until 2026-04-17 20:44:38.707558327 +0000 UTC m=+4.181336594 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-79v97" (UniqueName: "kubernetes.io/projected/ac379788-302d-4181-8970-56f47ae312c1-kube-api-access-79v97") pod "network-check-target-cm2lg" (UID: "ac379788-302d-4181-8970-56f47ae312c1") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 20:44:37.750432 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:37.750332 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fa81115_694f_4f2f_8797_8962d102b7e7.slice/crio-d268dd6e258aba152260026a04de7e6f14c588541c6d30e3cb4ecafb315b3a44 WatchSource:0}: Error finding container d268dd6e258aba152260026a04de7e6f14c588541c6d30e3cb4ecafb315b3a44: Status 404 returned error can't find the container with id d268dd6e258aba152260026a04de7e6f14c588541c6d30e3cb4ecafb315b3a44 Apr 17 20:44:37.751431 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:37.751411 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod83de7e89_23a8_4a38_833c_4fa3b9c105ca.slice/crio-1cd2644d5a58a3798554f1138a628ad5fa0c5c613d5af65daa0e1a7d74b6df06 WatchSource:0}: Error finding container 1cd2644d5a58a3798554f1138a628ad5fa0c5c613d5af65daa0e1a7d74b6df06: Status 404 returned error can't find the container with id 1cd2644d5a58a3798554f1138a628ad5fa0c5c613d5af65daa0e1a7d74b6df06 Apr 17 20:44:37.752140 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:37.752118 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b68ff3e_8a24_45e1_a70f_83f0539a9623.slice/crio-f2d5b5d3cd5efa43ce799bf96edbbd3c7bfafb7eae348042ab39e1f862b15958 WatchSource:0}: Error finding container f2d5b5d3cd5efa43ce799bf96edbbd3c7bfafb7eae348042ab39e1f862b15958: Status 404 returned error can't find the container with id f2d5b5d3cd5efa43ce799bf96edbbd3c7bfafb7eae348042ab39e1f862b15958 Apr 17 20:44:37.755859 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:37.755839 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc9f9af16_3b9f_4ebb_8025_9b9162c7eaa1.slice/crio-1d50308fa22983ba37a37ddcae7f1537f78367b3f9d38c5846f4bcafe5b652f7 WatchSource:0}: Error finding container 1d50308fa22983ba37a37ddcae7f1537f78367b3f9d38c5846f4bcafe5b652f7: Status 404 returned error can't find the container with id 1d50308fa22983ba37a37ddcae7f1537f78367b3f9d38c5846f4bcafe5b652f7 Apr 17 20:44:37.756947 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:37.756925 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod906a412c_f8af_4800_9e64_23e1b3be650e.slice/crio-4edff9c89957f3dd9176f931eaf64852430adb844abb5e83652549335dc685d3 WatchSource:0}: Error finding container 4edff9c89957f3dd9176f931eaf64852430adb844abb5e83652549335dc685d3: Status 404 returned error can't find the container with id 4edff9c89957f3dd9176f931eaf64852430adb844abb5e83652549335dc685d3 Apr 17 20:44:37.758265 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:37.758230 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48141b3e_9ad0_4dca_a6d1_055ed87fba90.slice/crio-cc8963c1e13abffb29a687ff8892f1fdaad6938cc14a20a7acf493923cdb13ee WatchSource:0}: Error finding container cc8963c1e13abffb29a687ff8892f1fdaad6938cc14a20a7acf493923cdb13ee: Status 404 returned error can't find the container with id cc8963c1e13abffb29a687ff8892f1fdaad6938cc14a20a7acf493923cdb13ee Apr 17 20:44:37.760689 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:44:37.760668 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37eb787b_5bf1_4d2b_b68e_5b7e766fe8e0.slice/crio-bd39e625894ca69a3360f7093fc07ba7c5e10fd9208ace41eaf6b17309b85a4f WatchSource:0}: Error finding container bd39e625894ca69a3360f7093fc07ba7c5e10fd9208ace41eaf6b17309b85a4f: Status 404 returned error can't find the container with id bd39e625894ca69a3360f7093fc07ba7c5e10fd9208ace41eaf6b17309b85a4f Apr 17 20:44:38.029752 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:38.029514 2577 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kubelet-serving" expiration="2028-04-16 20:39:36 +0000 UTC" deadline="2027-10-20 22:56:53.862358201 +0000 UTC" Apr 17 20:44:38.029752 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:38.029700 2577 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kubelet-serving" sleep="13226h12m15.832662604s" Apr 17 20:44:38.116172 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:38.116148 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:44:38.117602 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:38.116607 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7ddt5" podUID="1a6721ad-2d96-4d01-a9db-b8654256e62f" Apr 17 20:44:38.122313 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:38.122274 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" event={"ID":"83de7e89-23a8-4a38-833c-4fa3b9c105ca","Type":"ContainerStarted","Data":"1cd2644d5a58a3798554f1138a628ad5fa0c5c613d5af65daa0e1a7d74b6df06"} Apr 17 20:44:38.123218 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:38.123196 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-v8fbw" event={"ID":"5fa81115-694f-4f2f-8797-8962d102b7e7","Type":"ContainerStarted","Data":"d268dd6e258aba152260026a04de7e6f14c588541c6d30e3cb4ecafb315b3a44"} Apr 17 20:44:38.124586 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:38.124560 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-137-110.ec2.internal" event={"ID":"3114a7bb68dc8d0f555ef7f4a1089352","Type":"ContainerStarted","Data":"5a06e8669873507720992e0b6e6ad89934054433d94fb1ede8165a8a484da751"} Apr 17 20:44:38.125535 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:38.125514 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zds6c" event={"ID":"06332f89-1179-4cde-9a2b-4335714933a1","Type":"ContainerStarted","Data":"90c64ec152e81bc5f05de8dd4134ead3874002e5f72660792b1a20174669ed20"} Apr 17 20:44:38.126456 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:38.126432 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" event={"ID":"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0","Type":"ContainerStarted","Data":"bd39e625894ca69a3360f7093fc07ba7c5e10fd9208ace41eaf6b17309b85a4f"} Apr 17 20:44:38.127488 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:38.127463 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2lsfw" event={"ID":"4b68ff3e-8a24-45e1-a70f-83f0539a9623","Type":"ContainerStarted","Data":"f2d5b5d3cd5efa43ce799bf96edbbd3c7bfafb7eae348042ab39e1f862b15958"} Apr 17 20:44:38.128364 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:38.128341 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-mkzlr" event={"ID":"2e78f338-82b0-4062-8438-2b45fae0f4cf","Type":"ContainerStarted","Data":"0f630d67e8f03b707b40a580bd919403f9bc2e066e795465b559f4cddf162cc0"} Apr 17 20:44:38.129365 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:38.129349 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-4898w" event={"ID":"48141b3e-9ad0-4dca-a6d1-055ed87fba90","Type":"ContainerStarted","Data":"cc8963c1e13abffb29a687ff8892f1fdaad6938cc14a20a7acf493923cdb13ee"} Apr 17 20:44:38.130255 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:38.130225 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/konnectivity-agent-cxc79" event={"ID":"906a412c-f8af-4800-9e64-23e1b3be650e","Type":"ContainerStarted","Data":"4edff9c89957f3dd9176f931eaf64852430adb844abb5e83652549335dc685d3"} Apr 17 20:44:38.131120 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:38.131103 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-cwp9d" event={"ID":"c9f9af16-3b9f-4ebb-8025-9b9162c7eaa1","Type":"ContainerStarted","Data":"1d50308fa22983ba37a37ddcae7f1537f78367b3f9d38c5846f4bcafe5b652f7"} Apr 17 20:44:38.135572 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:38.135540 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-proxy-ip-10-0-137-110.ec2.internal" podStartSLOduration=2.135530459 podStartE2EDuration="2.135530459s" podCreationTimestamp="2026-04-17 20:44:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 20:44:38.135471302 +0000 UTC m=+3.609249589" watchObservedRunningTime="2026-04-17 20:44:38.135530459 +0000 UTC m=+3.609308737" Apr 17 20:44:38.616400 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:38.616362 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1a6721ad-2d96-4d01-a9db-b8654256e62f-metrics-certs\") pod \"network-metrics-daemon-7ddt5\" (UID: \"1a6721ad-2d96-4d01-a9db-b8654256e62f\") " pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:44:38.616541 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:38.616516 2577 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 20:44:38.616643 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:38.616609 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1a6721ad-2d96-4d01-a9db-b8654256e62f-metrics-certs podName:1a6721ad-2d96-4d01-a9db-b8654256e62f nodeName:}" failed. No retries permitted until 2026-04-17 20:44:40.616561326 +0000 UTC m=+6.090339596 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1a6721ad-2d96-4d01-a9db-b8654256e62f-metrics-certs") pod "network-metrics-daemon-7ddt5" (UID: "1a6721ad-2d96-4d01-a9db-b8654256e62f") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 20:44:38.717602 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:38.717519 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-79v97\" (UniqueName: \"kubernetes.io/projected/ac379788-302d-4181-8970-56f47ae312c1-kube-api-access-79v97\") pod \"network-check-target-cm2lg\" (UID: \"ac379788-302d-4181-8970-56f47ae312c1\") " pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:44:38.717747 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:38.717674 2577 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 17 20:44:38.717747 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:38.717694 2577 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 17 20:44:38.717747 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:38.717707 2577 projected.go:194] Error preparing data for projected volume kube-api-access-79v97 for pod openshift-network-diagnostics/network-check-target-cm2lg: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 20:44:38.717937 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:38.717766 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ac379788-302d-4181-8970-56f47ae312c1-kube-api-access-79v97 podName:ac379788-302d-4181-8970-56f47ae312c1 nodeName:}" failed. No retries permitted until 2026-04-17 20:44:40.717746418 +0000 UTC m=+6.191524686 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-79v97" (UniqueName: "kubernetes.io/projected/ac379788-302d-4181-8970-56f47ae312c1-kube-api-access-79v97") pod "network-check-target-cm2lg" (UID: "ac379788-302d-4181-8970-56f47ae312c1") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 20:44:39.116515 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:39.116018 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:44:39.116515 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:39.116143 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cm2lg" podUID="ac379788-302d-4181-8970-56f47ae312c1" Apr 17 20:44:39.146421 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:39.146379 2577 generic.go:358] "Generic (PLEG): container finished" podID="a8586432e8767e95c199e223475265b6" containerID="f5a623c9e3af594888b9503182d43ed3c649783862e949ab3973fc69d9a60177" exitCode=0 Apr 17 20:44:39.147268 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:39.147241 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-110.ec2.internal" event={"ID":"a8586432e8767e95c199e223475265b6","Type":"ContainerDied","Data":"f5a623c9e3af594888b9503182d43ed3c649783862e949ab3973fc69d9a60177"} Apr 17 20:44:40.116435 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:40.116403 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:44:40.116611 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:40.116548 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7ddt5" podUID="1a6721ad-2d96-4d01-a9db-b8654256e62f" Apr 17 20:44:40.171723 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:40.171687 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-110.ec2.internal" event={"ID":"a8586432e8767e95c199e223475265b6","Type":"ContainerStarted","Data":"688b251d64d19f92ed69addb9cf84f09b735d5a3a967582f61cee3ed9e687ed2"} Apr 17 20:44:40.633584 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:40.633550 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1a6721ad-2d96-4d01-a9db-b8654256e62f-metrics-certs\") pod \"network-metrics-daemon-7ddt5\" (UID: \"1a6721ad-2d96-4d01-a9db-b8654256e62f\") " pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:44:40.633753 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:40.633722 2577 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 20:44:40.633833 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:40.633785 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1a6721ad-2d96-4d01-a9db-b8654256e62f-metrics-certs podName:1a6721ad-2d96-4d01-a9db-b8654256e62f nodeName:}" failed. No retries permitted until 2026-04-17 20:44:44.633765372 +0000 UTC m=+10.107543651 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1a6721ad-2d96-4d01-a9db-b8654256e62f-metrics-certs") pod "network-metrics-daemon-7ddt5" (UID: "1a6721ad-2d96-4d01-a9db-b8654256e62f") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 20:44:40.734725 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:40.734685 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-79v97\" (UniqueName: \"kubernetes.io/projected/ac379788-302d-4181-8970-56f47ae312c1-kube-api-access-79v97\") pod \"network-check-target-cm2lg\" (UID: \"ac379788-302d-4181-8970-56f47ae312c1\") " pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:44:40.734919 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:40.734865 2577 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 17 20:44:40.734919 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:40.734885 2577 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 17 20:44:40.734919 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:40.734897 2577 projected.go:194] Error preparing data for projected volume kube-api-access-79v97 for pod openshift-network-diagnostics/network-check-target-cm2lg: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 20:44:40.735067 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:40.734953 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ac379788-302d-4181-8970-56f47ae312c1-kube-api-access-79v97 podName:ac379788-302d-4181-8970-56f47ae312c1 nodeName:}" failed. No retries permitted until 2026-04-17 20:44:44.734935481 +0000 UTC m=+10.208713750 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-79v97" (UniqueName: "kubernetes.io/projected/ac379788-302d-4181-8970-56f47ae312c1-kube-api-access-79v97") pod "network-check-target-cm2lg" (UID: "ac379788-302d-4181-8970-56f47ae312c1") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 20:44:41.116688 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:41.116184 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:44:41.116688 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:41.116309 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cm2lg" podUID="ac379788-302d-4181-8970-56f47ae312c1" Apr 17 20:44:42.116664 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:42.116632 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:44:42.116857 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:42.116779 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7ddt5" podUID="1a6721ad-2d96-4d01-a9db-b8654256e62f" Apr 17 20:44:43.119819 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:43.119773 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:44:43.120191 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:43.119909 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cm2lg" podUID="ac379788-302d-4181-8970-56f47ae312c1" Apr 17 20:44:44.116400 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:44.116364 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:44:44.116573 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:44.116490 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7ddt5" podUID="1a6721ad-2d96-4d01-a9db-b8654256e62f" Apr 17 20:44:44.669504 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:44.669464 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1a6721ad-2d96-4d01-a9db-b8654256e62f-metrics-certs\") pod \"network-metrics-daemon-7ddt5\" (UID: \"1a6721ad-2d96-4d01-a9db-b8654256e62f\") " pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:44:44.670013 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:44.669775 2577 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 20:44:44.670013 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:44.669864 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1a6721ad-2d96-4d01-a9db-b8654256e62f-metrics-certs podName:1a6721ad-2d96-4d01-a9db-b8654256e62f nodeName:}" failed. No retries permitted until 2026-04-17 20:44:52.669841789 +0000 UTC m=+18.143620057 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1a6721ad-2d96-4d01-a9db-b8654256e62f-metrics-certs") pod "network-metrics-daemon-7ddt5" (UID: "1a6721ad-2d96-4d01-a9db-b8654256e62f") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 20:44:44.770046 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:44.770004 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-79v97\" (UniqueName: \"kubernetes.io/projected/ac379788-302d-4181-8970-56f47ae312c1-kube-api-access-79v97\") pod \"network-check-target-cm2lg\" (UID: \"ac379788-302d-4181-8970-56f47ae312c1\") " pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:44:44.770213 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:44.770177 2577 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 17 20:44:44.770213 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:44.770196 2577 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 17 20:44:44.770213 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:44.770210 2577 projected.go:194] Error preparing data for projected volume kube-api-access-79v97 for pod openshift-network-diagnostics/network-check-target-cm2lg: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 20:44:44.770378 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:44.770264 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ac379788-302d-4181-8970-56f47ae312c1-kube-api-access-79v97 podName:ac379788-302d-4181-8970-56f47ae312c1 nodeName:}" failed. No retries permitted until 2026-04-17 20:44:52.770246658 +0000 UTC m=+18.244024929 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-79v97" (UniqueName: "kubernetes.io/projected/ac379788-302d-4181-8970-56f47ae312c1-kube-api-access-79v97") pod "network-check-target-cm2lg" (UID: "ac379788-302d-4181-8970-56f47ae312c1") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 20:44:45.117381 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:45.117292 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:44:45.120754 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:45.120714 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cm2lg" podUID="ac379788-302d-4181-8970-56f47ae312c1" Apr 17 20:44:46.116294 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:46.116266 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:44:46.116742 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:46.116392 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7ddt5" podUID="1a6721ad-2d96-4d01-a9db-b8654256e62f" Apr 17 20:44:47.118987 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:47.118956 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:44:47.119365 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:47.119068 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cm2lg" podUID="ac379788-302d-4181-8970-56f47ae312c1" Apr 17 20:44:48.116447 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:48.116419 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:44:48.116596 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:48.116543 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7ddt5" podUID="1a6721ad-2d96-4d01-a9db-b8654256e62f" Apr 17 20:44:49.116732 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:49.116694 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:44:49.117216 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:49.116842 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cm2lg" podUID="ac379788-302d-4181-8970-56f47ae312c1" Apr 17 20:44:50.116610 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:50.116575 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:44:50.116812 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:50.116685 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7ddt5" podUID="1a6721ad-2d96-4d01-a9db-b8654256e62f" Apr 17 20:44:51.116194 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:51.116156 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:44:51.116367 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:51.116281 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cm2lg" podUID="ac379788-302d-4181-8970-56f47ae312c1" Apr 17 20:44:52.116053 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:52.116020 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:44:52.116443 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:52.116133 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7ddt5" podUID="1a6721ad-2d96-4d01-a9db-b8654256e62f" Apr 17 20:44:52.725039 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:52.725000 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1a6721ad-2d96-4d01-a9db-b8654256e62f-metrics-certs\") pod \"network-metrics-daemon-7ddt5\" (UID: \"1a6721ad-2d96-4d01-a9db-b8654256e62f\") " pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:44:52.725201 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:52.725118 2577 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 20:44:52.725201 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:52.725181 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1a6721ad-2d96-4d01-a9db-b8654256e62f-metrics-certs podName:1a6721ad-2d96-4d01-a9db-b8654256e62f nodeName:}" failed. No retries permitted until 2026-04-17 20:45:08.725164219 +0000 UTC m=+34.198942484 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1a6721ad-2d96-4d01-a9db-b8654256e62f-metrics-certs") pod "network-metrics-daemon-7ddt5" (UID: "1a6721ad-2d96-4d01-a9db-b8654256e62f") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 20:44:52.825737 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:52.825696 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-79v97\" (UniqueName: \"kubernetes.io/projected/ac379788-302d-4181-8970-56f47ae312c1-kube-api-access-79v97\") pod \"network-check-target-cm2lg\" (UID: \"ac379788-302d-4181-8970-56f47ae312c1\") " pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:44:52.825904 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:52.825868 2577 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 17 20:44:52.825904 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:52.825892 2577 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 17 20:44:52.825904 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:52.825904 2577 projected.go:194] Error preparing data for projected volume kube-api-access-79v97 for pod openshift-network-diagnostics/network-check-target-cm2lg: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 20:44:52.826063 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:52.825961 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ac379788-302d-4181-8970-56f47ae312c1-kube-api-access-79v97 podName:ac379788-302d-4181-8970-56f47ae312c1 nodeName:}" failed. No retries permitted until 2026-04-17 20:45:08.825941558 +0000 UTC m=+34.299719825 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-79v97" (UniqueName: "kubernetes.io/projected/ac379788-302d-4181-8970-56f47ae312c1-kube-api-access-79v97") pod "network-check-target-cm2lg" (UID: "ac379788-302d-4181-8970-56f47ae312c1") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 20:44:53.115871 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:53.115709 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:44:53.115871 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:53.115855 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cm2lg" podUID="ac379788-302d-4181-8970-56f47ae312c1" Apr 17 20:44:54.116163 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:54.116135 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:44:54.116616 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:54.116252 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7ddt5" podUID="1a6721ad-2d96-4d01-a9db-b8654256e62f" Apr 17 20:44:55.116537 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:55.116510 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:44:55.116919 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:55.116612 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cm2lg" podUID="ac379788-302d-4181-8970-56f47ae312c1" Apr 17 20:44:56.116743 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:56.116386 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:44:56.117551 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:56.116825 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7ddt5" podUID="1a6721ad-2d96-4d01-a9db-b8654256e62f" Apr 17 20:44:56.198401 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:56.198338 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2lsfw" event={"ID":"4b68ff3e-8a24-45e1-a70f-83f0539a9623","Type":"ContainerStarted","Data":"74f4851de6df816c9dee6c7609820fd3632a842e415554bda82c8b19f1304ad7"} Apr 17 20:44:56.200686 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:56.200649 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-4898w" event={"ID":"48141b3e-9ad0-4dca-a6d1-055ed87fba90","Type":"ContainerStarted","Data":"790d74dbfd9c2c133b40e2616ddf1f961320e2a775da6599842855ff9d808eb4"} Apr 17 20:44:56.202827 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:56.202754 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/konnectivity-agent-cxc79" event={"ID":"906a412c-f8af-4800-9e64-23e1b3be650e","Type":"ContainerStarted","Data":"ea06c51d4c232cd14fb1bfb3cd80be38ed0044944225ea6c9c8c157a96f944b4"} Apr 17 20:44:56.204394 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:56.204374 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-cwp9d" event={"ID":"c9f9af16-3b9f-4ebb-8025-9b9162c7eaa1","Type":"ContainerStarted","Data":"5f7308bc033a2d55bec4bb708ebc29d80a33e13063582653e65eb52965beba95"} Apr 17 20:44:56.205784 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:56.205757 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" event={"ID":"83de7e89-23a8-4a38-833c-4fa3b9c105ca","Type":"ContainerStarted","Data":"97c3c3e5f03ccd39a06d29229d1eaeff470ac1285e89b6b4e7319655b50cb1e9"} Apr 17 20:44:56.207261 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:56.207183 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-v8fbw" event={"ID":"5fa81115-694f-4f2f-8797-8962d102b7e7","Type":"ContainerStarted","Data":"3ecbe78118e9ea7dbf4786877a5388337af2f012fa27d311d6898447d2947555"} Apr 17 20:44:56.208542 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:56.208517 2577 generic.go:358] "Generic (PLEG): container finished" podID="06332f89-1179-4cde-9a2b-4335714933a1" containerID="40f224c032018264e48176abeae6f6e8b6ce98922aadee9c61e32c07b0112eae" exitCode=0 Apr 17 20:44:56.208637 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:56.208591 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zds6c" event={"ID":"06332f89-1179-4cde-9a2b-4335714933a1","Type":"ContainerDied","Data":"40f224c032018264e48176abeae6f6e8b6ce98922aadee9c61e32c07b0112eae"} Apr 17 20:44:56.211460 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:56.211444 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dk97h_37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0/ovn-acl-logging/0.log" Apr 17 20:44:56.211817 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:56.211774 2577 generic.go:358] "Generic (PLEG): container finished" podID="37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0" containerID="c7d6d3180d41358f509d39fed3b9c295237de1d43bbb7aeff23b36fe17b27f13" exitCode=1 Apr 17 20:44:56.211890 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:56.211819 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" event={"ID":"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0","Type":"ContainerStarted","Data":"bd0c25b49eee3a6252cb8d5d6040dc34c43ae2bd7b4d6a245dd76544ebc9d3b3"} Apr 17 20:44:56.211890 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:56.211848 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" event={"ID":"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0","Type":"ContainerStarted","Data":"360286d9c5b0fba2b355740534939af94aab6cd02bc76a38cebcf24fd475910b"} Apr 17 20:44:56.211890 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:56.211863 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" event={"ID":"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0","Type":"ContainerStarted","Data":"e4ed9dc261f054d7e72e4a5a9c1e700451dfceb3b5b98a4b0e0668384a133925"} Apr 17 20:44:56.211890 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:56.211876 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" event={"ID":"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0","Type":"ContainerStarted","Data":"6692dbea95004496f0686360d66eb6e3afc6ddd6ccf3faf84a2c02991f5e8683"} Apr 17 20:44:56.211890 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:56.211887 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" event={"ID":"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0","Type":"ContainerDied","Data":"c7d6d3180d41358f509d39fed3b9c295237de1d43bbb7aeff23b36fe17b27f13"} Apr 17 20:44:56.212072 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:56.211902 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" event={"ID":"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0","Type":"ContainerStarted","Data":"5f959f24ad1b832a342b59fbe534c10c1f4d1af4d0e8798fb083ab633c618845"} Apr 17 20:44:56.215134 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:56.215101 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-137-110.ec2.internal" podStartSLOduration=20.215089968 podStartE2EDuration="20.215089968s" podCreationTimestamp="2026-04-17 20:44:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 20:44:40.187073912 +0000 UTC m=+5.660852199" watchObservedRunningTime="2026-04-17 20:44:56.215089968 +0000 UTC m=+21.688868255" Apr 17 20:44:56.215222 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:56.215175 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-2lsfw" podStartSLOduration=3.846214304 podStartE2EDuration="21.215165919s" podCreationTimestamp="2026-04-17 20:44:35 +0000 UTC" firstStartedPulling="2026-04-17 20:44:37.754179523 +0000 UTC m=+3.227957789" lastFinishedPulling="2026-04-17 20:44:55.123131138 +0000 UTC m=+20.596909404" observedRunningTime="2026-04-17 20:44:56.214594108 +0000 UTC m=+21.688372397" watchObservedRunningTime="2026-04-17 20:44:56.215165919 +0000 UTC m=+21.688944209" Apr 17 20:44:56.236881 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:56.236847 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-v8fbw" podStartSLOduration=3.9059169540000003 podStartE2EDuration="21.236838061s" podCreationTimestamp="2026-04-17 20:44:35 +0000 UTC" firstStartedPulling="2026-04-17 20:44:37.752963459 +0000 UTC m=+3.226741724" lastFinishedPulling="2026-04-17 20:44:55.083884559 +0000 UTC m=+20.557662831" observedRunningTime="2026-04-17 20:44:56.225533294 +0000 UTC m=+21.699311583" watchObservedRunningTime="2026-04-17 20:44:56.236838061 +0000 UTC m=+21.710616389" Apr 17 20:44:56.237093 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:56.237064 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-cwp9d" podStartSLOduration=3.910684333 podStartE2EDuration="21.237057887s" podCreationTimestamp="2026-04-17 20:44:35 +0000 UTC" firstStartedPulling="2026-04-17 20:44:37.757537092 +0000 UTC m=+3.231315371" lastFinishedPulling="2026-04-17 20:44:55.083910657 +0000 UTC m=+20.557688925" observedRunningTime="2026-04-17 20:44:56.236694924 +0000 UTC m=+21.710473212" watchObservedRunningTime="2026-04-17 20:44:56.237057887 +0000 UTC m=+21.710836173" Apr 17 20:44:56.272747 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:56.272712 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/konnectivity-agent-cxc79" podStartSLOduration=3.947739531 podStartE2EDuration="21.272702805s" podCreationTimestamp="2026-04-17 20:44:35 +0000 UTC" firstStartedPulling="2026-04-17 20:44:37.75891934 +0000 UTC m=+3.232697609" lastFinishedPulling="2026-04-17 20:44:55.08388261 +0000 UTC m=+20.557660883" observedRunningTime="2026-04-17 20:44:56.249428737 +0000 UTC m=+21.723207026" watchObservedRunningTime="2026-04-17 20:44:56.272702805 +0000 UTC m=+21.746481091" Apr 17 20:44:56.285321 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:56.285288 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-node-tuning-operator/tuned-4898w" podStartSLOduration=3.962941224 podStartE2EDuration="21.285276449s" podCreationTimestamp="2026-04-17 20:44:35 +0000 UTC" firstStartedPulling="2026-04-17 20:44:37.761643609 +0000 UTC m=+3.235421881" lastFinishedPulling="2026-04-17 20:44:55.083978826 +0000 UTC m=+20.557757106" observedRunningTime="2026-04-17 20:44:56.285182523 +0000 UTC m=+21.758960810" watchObservedRunningTime="2026-04-17 20:44:56.285276449 +0000 UTC m=+21.759054738" Apr 17 20:44:56.455789 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:56.455761 2577 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="kube-system/konnectivity-agent-cxc79" Apr 17 20:44:56.456673 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:56.456658 2577 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="kube-system/konnectivity-agent-cxc79" Apr 17 20:44:56.765659 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:56.765637 2577 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/ebs.csi.aws.com-reg.sock" Apr 17 20:44:57.067690 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:57.067532 2577 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/ebs.csi.aws.com-reg.sock","Timestamp":"2026-04-17T20:44:56.765654837Z","UUID":"581040cc-7ec4-46e2-95ec-cdacb3264935","Handler":null,"Name":"","Endpoint":""} Apr 17 20:44:57.070362 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:57.070344 2577 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: ebs.csi.aws.com endpoint: /var/lib/kubelet/plugins/ebs.csi.aws.com/csi.sock versions: 1.0.0 Apr 17 20:44:57.070362 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:57.070366 2577 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: ebs.csi.aws.com at endpoint: /var/lib/kubelet/plugins/ebs.csi.aws.com/csi.sock Apr 17 20:44:57.116171 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:57.116146 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:44:57.116321 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:57.116262 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cm2lg" podUID="ac379788-302d-4181-8970-56f47ae312c1" Apr 17 20:44:57.216442 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:57.216413 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-mkzlr" event={"ID":"2e78f338-82b0-4062-8438-2b45fae0f4cf","Type":"ContainerStarted","Data":"2bf6ef0667a95eac4221310531cf05fc511849b2848a2e212e1a3911d37c5923"} Apr 17 20:44:57.218902 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:57.218874 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" event={"ID":"83de7e89-23a8-4a38-833c-4fa3b9c105ca","Type":"ContainerStarted","Data":"f59637f7221324eff1b6ca5dc94ba60b9de46d0be5f9c2472640b80aa3a60a9d"} Apr 17 20:44:57.219463 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:57.219444 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kube-system/konnectivity-agent-cxc79" Apr 17 20:44:57.219966 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:57.219948 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kube-system/konnectivity-agent-cxc79" Apr 17 20:44:57.229426 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:57.229070 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-operator/iptables-alerter-mkzlr" podStartSLOduration=4.932918532 podStartE2EDuration="22.229054642s" podCreationTimestamp="2026-04-17 20:44:35 +0000 UTC" firstStartedPulling="2026-04-17 20:44:37.762904903 +0000 UTC m=+3.236683168" lastFinishedPulling="2026-04-17 20:44:55.059041013 +0000 UTC m=+20.532819278" observedRunningTime="2026-04-17 20:44:57.228602963 +0000 UTC m=+22.702381250" watchObservedRunningTime="2026-04-17 20:44:57.229054642 +0000 UTC m=+22.702832932" Apr 17 20:44:58.115682 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:58.115655 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:44:58.115841 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:58.115771 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7ddt5" podUID="1a6721ad-2d96-4d01-a9db-b8654256e62f" Apr 17 20:44:58.223845 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:58.223826 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dk97h_37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0/ovn-acl-logging/0.log" Apr 17 20:44:58.224270 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:58.224256 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" event={"ID":"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0","Type":"ContainerStarted","Data":"8d5c1ee74c4276fbbe66d5cc4ea50fa105c2e2967caf8e4de0046ba22483f5ac"} Apr 17 20:44:58.226301 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:58.226258 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" event={"ID":"83de7e89-23a8-4a38-833c-4fa3b9c105ca","Type":"ContainerStarted","Data":"efa202a022b9b6ac104b32dac13936d426274e517a0f8b942b6861a22b5ceb3c"} Apr 17 20:44:58.241313 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:58.241271 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-fj8np" podStartSLOduration=3.261276115 podStartE2EDuration="23.241259259s" podCreationTimestamp="2026-04-17 20:44:35 +0000 UTC" firstStartedPulling="2026-04-17 20:44:37.753633041 +0000 UTC m=+3.227411306" lastFinishedPulling="2026-04-17 20:44:57.733616185 +0000 UTC m=+23.207394450" observedRunningTime="2026-04-17 20:44:58.240657912 +0000 UTC m=+23.714436198" watchObservedRunningTime="2026-04-17 20:44:58.241259259 +0000 UTC m=+23.715037546" Apr 17 20:44:59.116097 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:44:59.116066 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:44:59.116289 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:44:59.116195 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cm2lg" podUID="ac379788-302d-4181-8970-56f47ae312c1" Apr 17 20:45:00.116847 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:00.116653 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:45:00.117214 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:00.116937 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7ddt5" podUID="1a6721ad-2d96-4d01-a9db-b8654256e62f" Apr 17 20:45:01.116567 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:01.116541 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:45:01.116678 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:01.116648 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cm2lg" podUID="ac379788-302d-4181-8970-56f47ae312c1" Apr 17 20:45:01.236229 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:01.236205 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dk97h_37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0/ovn-acl-logging/0.log" Apr 17 20:45:01.237132 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:01.237104 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" event={"ID":"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0","Type":"ContainerStarted","Data":"88f1768a0bfe3a0855e9945155e941ca70f261f1b64656f3adb5c19f53221591"} Apr 17 20:45:01.237661 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:01.237639 2577 scope.go:117] "RemoveContainer" containerID="c7d6d3180d41358f509d39fed3b9c295237de1d43bbb7aeff23b36fe17b27f13" Apr 17 20:45:01.238161 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:01.238109 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:45:01.238214 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:01.238179 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:45:01.258414 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:01.258359 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:45:01.259471 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:01.259438 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:45:02.116264 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:02.116093 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:45:02.116359 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:02.116341 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7ddt5" podUID="1a6721ad-2d96-4d01-a9db-b8654256e62f" Apr 17 20:45:02.242953 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:02.242928 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dk97h_37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0/ovn-acl-logging/0.log" Apr 17 20:45:02.243357 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:02.243267 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" event={"ID":"37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0","Type":"ContainerStarted","Data":"0536567f6b8748b6ae9e8211050ba882df39af87c39bd9bbb7153e639498d557"} Apr 17 20:45:02.243357 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:02.243337 2577 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 17 20:45:02.244672 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:02.244647 2577 generic.go:358] "Generic (PLEG): container finished" podID="06332f89-1179-4cde-9a2b-4335714933a1" containerID="6d4b945d8bda9e8c4069550c26fa9720af8db9c336530e3f35bf298c2a837067" exitCode=0 Apr 17 20:45:02.244769 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:02.244681 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zds6c" event={"ID":"06332f89-1179-4cde-9a2b-4335714933a1","Type":"ContainerDied","Data":"6d4b945d8bda9e8c4069550c26fa9720af8db9c336530e3f35bf298c2a837067"} Apr 17 20:45:02.267429 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:02.267384 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" podStartSLOduration=9.869835917 podStartE2EDuration="27.2673748s" podCreationTimestamp="2026-04-17 20:44:35 +0000 UTC" firstStartedPulling="2026-04-17 20:44:37.763432255 +0000 UTC m=+3.237210526" lastFinishedPulling="2026-04-17 20:44:55.160971143 +0000 UTC m=+20.634749409" observedRunningTime="2026-04-17 20:45:02.266037854 +0000 UTC m=+27.739816141" watchObservedRunningTime="2026-04-17 20:45:02.2673748 +0000 UTC m=+27.741153117" Apr 17 20:45:03.088769 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:03.088733 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-7ddt5"] Apr 17 20:45:03.088900 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:03.088887 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:45:03.089010 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:03.088984 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7ddt5" podUID="1a6721ad-2d96-4d01-a9db-b8654256e62f" Apr 17 20:45:03.091127 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:03.091105 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-target-cm2lg"] Apr 17 20:45:03.091208 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:03.091192 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:45:03.091271 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:03.091255 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cm2lg" podUID="ac379788-302d-4181-8970-56f47ae312c1" Apr 17 20:45:03.248244 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:03.248214 2577 generic.go:358] "Generic (PLEG): container finished" podID="06332f89-1179-4cde-9a2b-4335714933a1" containerID="b0733de7f7844acf130ef5ca0a680990f6e22a0a6223771b92eafee5416bdf3a" exitCode=0 Apr 17 20:45:03.248579 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:03.248340 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zds6c" event={"ID":"06332f89-1179-4cde-9a2b-4335714933a1","Type":"ContainerDied","Data":"b0733de7f7844acf130ef5ca0a680990f6e22a0a6223771b92eafee5416bdf3a"} Apr 17 20:45:03.248579 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:03.248478 2577 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 17 20:45:04.251976 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:04.251946 2577 generic.go:358] "Generic (PLEG): container finished" podID="06332f89-1179-4cde-9a2b-4335714933a1" containerID="d86bc03f933e61c434e046ad50a3ae01e1982dac39eed06ec068dee77d54512e" exitCode=0 Apr 17 20:45:04.252228 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:04.252023 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zds6c" event={"ID":"06332f89-1179-4cde-9a2b-4335714933a1","Type":"ContainerDied","Data":"d86bc03f933e61c434e046ad50a3ae01e1982dac39eed06ec068dee77d54512e"} Apr 17 20:45:05.008491 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:05.008449 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:45:05.119431 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:05.119405 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:45:05.119595 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:05.119431 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:45:05.119595 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:05.119502 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cm2lg" podUID="ac379788-302d-4181-8970-56f47ae312c1" Apr 17 20:45:05.119704 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:05.119584 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7ddt5" podUID="1a6721ad-2d96-4d01-a9db-b8654256e62f" Apr 17 20:45:07.116554 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:07.116307 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:45:07.117015 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:07.116357 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:45:07.119924 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:07.116683 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-7ddt5" podUID="1a6721ad-2d96-4d01-a9db-b8654256e62f" Apr 17 20:45:07.119924 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:07.117245 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-cm2lg" podUID="ac379788-302d-4181-8970-56f47ae312c1" Apr 17 20:45:08.370194 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.370166 2577 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-137-110.ec2.internal" event="NodeReady" Apr 17 20:45:08.370695 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.370286 2577 kubelet_node_status.go:550] "Fast updating node status as it just became ready" Apr 17 20:45:08.413378 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.413344 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-bc6wf"] Apr 17 20:45:08.417890 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.417860 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-q59g7"] Apr 17 20:45:08.418050 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.418031 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-bc6wf" Apr 17 20:45:08.419752 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.419675 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-dockercfg-8pvd4\"" Apr 17 20:45:08.419910 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.419852 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"dns-default\"" Apr 17 20:45:08.419973 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.419962 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-default-metrics-tls\"" Apr 17 20:45:08.421222 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.421196 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-q59g7" Apr 17 20:45:08.422751 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.422731 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"openshift-service-ca.crt\"" Apr 17 20:45:08.422967 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.422952 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"canary-serving-cert\"" Apr 17 20:45:08.423250 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.423228 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"default-dockercfg-pn5wd\"" Apr 17 20:45:08.423932 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.423912 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"kube-root-ca.crt\"" Apr 17 20:45:08.424841 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.424787 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-bc6wf"] Apr 17 20:45:08.425444 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.425424 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-q59g7"] Apr 17 20:45:08.548083 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.548049 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/821ba3d4-448f-45b9-bcec-701682592a0f-config-volume\") pod \"dns-default-bc6wf\" (UID: \"821ba3d4-448f-45b9-bcec-701682592a0f\") " pod="openshift-dns/dns-default-bc6wf" Apr 17 20:45:08.548083 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.548083 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/821ba3d4-448f-45b9-bcec-701682592a0f-metrics-tls\") pod \"dns-default-bc6wf\" (UID: \"821ba3d4-448f-45b9-bcec-701682592a0f\") " pod="openshift-dns/dns-default-bc6wf" Apr 17 20:45:08.548287 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.548113 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v77dc\" (UniqueName: \"kubernetes.io/projected/79b05e51-04fa-4473-b612-c762aa0e3856-kube-api-access-v77dc\") pod \"ingress-canary-q59g7\" (UID: \"79b05e51-04fa-4473-b612-c762aa0e3856\") " pod="openshift-ingress-canary/ingress-canary-q59g7" Apr 17 20:45:08.548287 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.548192 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/821ba3d4-448f-45b9-bcec-701682592a0f-tmp-dir\") pod \"dns-default-bc6wf\" (UID: \"821ba3d4-448f-45b9-bcec-701682592a0f\") " pod="openshift-dns/dns-default-bc6wf" Apr 17 20:45:08.548287 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.548278 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sfgn\" (UniqueName: \"kubernetes.io/projected/821ba3d4-448f-45b9-bcec-701682592a0f-kube-api-access-6sfgn\") pod \"dns-default-bc6wf\" (UID: \"821ba3d4-448f-45b9-bcec-701682592a0f\") " pod="openshift-dns/dns-default-bc6wf" Apr 17 20:45:08.548424 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.548306 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79b05e51-04fa-4473-b612-c762aa0e3856-cert\") pod \"ingress-canary-q59g7\" (UID: \"79b05e51-04fa-4473-b612-c762aa0e3856\") " pod="openshift-ingress-canary/ingress-canary-q59g7" Apr 17 20:45:08.648923 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.648849 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6sfgn\" (UniqueName: \"kubernetes.io/projected/821ba3d4-448f-45b9-bcec-701682592a0f-kube-api-access-6sfgn\") pod \"dns-default-bc6wf\" (UID: \"821ba3d4-448f-45b9-bcec-701682592a0f\") " pod="openshift-dns/dns-default-bc6wf" Apr 17 20:45:08.648923 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.648891 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79b05e51-04fa-4473-b612-c762aa0e3856-cert\") pod \"ingress-canary-q59g7\" (UID: \"79b05e51-04fa-4473-b612-c762aa0e3856\") " pod="openshift-ingress-canary/ingress-canary-q59g7" Apr 17 20:45:08.649136 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.648925 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/821ba3d4-448f-45b9-bcec-701682592a0f-config-volume\") pod \"dns-default-bc6wf\" (UID: \"821ba3d4-448f-45b9-bcec-701682592a0f\") " pod="openshift-dns/dns-default-bc6wf" Apr 17 20:45:08.649136 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:08.649028 2577 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 17 20:45:08.649136 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.649044 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/821ba3d4-448f-45b9-bcec-701682592a0f-metrics-tls\") pod \"dns-default-bc6wf\" (UID: \"821ba3d4-448f-45b9-bcec-701682592a0f\") " pod="openshift-dns/dns-default-bc6wf" Apr 17 20:45:08.649136 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:08.649097 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79b05e51-04fa-4473-b612-c762aa0e3856-cert podName:79b05e51-04fa-4473-b612-c762aa0e3856 nodeName:}" failed. No retries permitted until 2026-04-17 20:45:09.149076098 +0000 UTC m=+34.622854363 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/79b05e51-04fa-4473-b612-c762aa0e3856-cert") pod "ingress-canary-q59g7" (UID: "79b05e51-04fa-4473-b612-c762aa0e3856") : secret "canary-serving-cert" not found Apr 17 20:45:08.649136 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.649117 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-v77dc\" (UniqueName: \"kubernetes.io/projected/79b05e51-04fa-4473-b612-c762aa0e3856-kube-api-access-v77dc\") pod \"ingress-canary-q59g7\" (UID: \"79b05e51-04fa-4473-b612-c762aa0e3856\") " pod="openshift-ingress-canary/ingress-canary-q59g7" Apr 17 20:45:08.649136 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:08.649128 2577 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 17 20:45:08.649443 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:08.649171 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/821ba3d4-448f-45b9-bcec-701682592a0f-metrics-tls podName:821ba3d4-448f-45b9-bcec-701682592a0f nodeName:}" failed. No retries permitted until 2026-04-17 20:45:09.149159691 +0000 UTC m=+34.622937956 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/821ba3d4-448f-45b9-bcec-701682592a0f-metrics-tls") pod "dns-default-bc6wf" (UID: "821ba3d4-448f-45b9-bcec-701682592a0f") : secret "dns-default-metrics-tls" not found Apr 17 20:45:08.649443 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.649198 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/821ba3d4-448f-45b9-bcec-701682592a0f-tmp-dir\") pod \"dns-default-bc6wf\" (UID: \"821ba3d4-448f-45b9-bcec-701682592a0f\") " pod="openshift-dns/dns-default-bc6wf" Apr 17 20:45:08.649537 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.649473 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/821ba3d4-448f-45b9-bcec-701682592a0f-config-volume\") pod \"dns-default-bc6wf\" (UID: \"821ba3d4-448f-45b9-bcec-701682592a0f\") " pod="openshift-dns/dns-default-bc6wf" Apr 17 20:45:08.649537 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.649506 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/821ba3d4-448f-45b9-bcec-701682592a0f-tmp-dir\") pod \"dns-default-bc6wf\" (UID: \"821ba3d4-448f-45b9-bcec-701682592a0f\") " pod="openshift-dns/dns-default-bc6wf" Apr 17 20:45:08.658921 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.658899 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sfgn\" (UniqueName: \"kubernetes.io/projected/821ba3d4-448f-45b9-bcec-701682592a0f-kube-api-access-6sfgn\") pod \"dns-default-bc6wf\" (UID: \"821ba3d4-448f-45b9-bcec-701682592a0f\") " pod="openshift-dns/dns-default-bc6wf" Apr 17 20:45:08.659047 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.658992 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-v77dc\" (UniqueName: \"kubernetes.io/projected/79b05e51-04fa-4473-b612-c762aa0e3856-kube-api-access-v77dc\") pod \"ingress-canary-q59g7\" (UID: \"79b05e51-04fa-4473-b612-c762aa0e3856\") " pod="openshift-ingress-canary/ingress-canary-q59g7" Apr 17 20:45:08.750298 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.750262 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1a6721ad-2d96-4d01-a9db-b8654256e62f-metrics-certs\") pod \"network-metrics-daemon-7ddt5\" (UID: \"1a6721ad-2d96-4d01-a9db-b8654256e62f\") " pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:45:08.750435 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:08.750413 2577 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 20:45:08.750485 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:08.750474 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1a6721ad-2d96-4d01-a9db-b8654256e62f-metrics-certs podName:1a6721ad-2d96-4d01-a9db-b8654256e62f nodeName:}" failed. No retries permitted until 2026-04-17 20:45:40.750456458 +0000 UTC m=+66.224234723 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1a6721ad-2d96-4d01-a9db-b8654256e62f-metrics-certs") pod "network-metrics-daemon-7ddt5" (UID: "1a6721ad-2d96-4d01-a9db-b8654256e62f") : object "openshift-multus"/"metrics-daemon-secret" not registered Apr 17 20:45:08.851179 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:08.851151 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-79v97\" (UniqueName: \"kubernetes.io/projected/ac379788-302d-4181-8970-56f47ae312c1-kube-api-access-79v97\") pod \"network-check-target-cm2lg\" (UID: \"ac379788-302d-4181-8970-56f47ae312c1\") " pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:45:08.851353 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:08.851327 2577 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Apr 17 20:45:08.851353 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:08.851349 2577 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Apr 17 20:45:08.851497 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:08.851359 2577 projected.go:194] Error preparing data for projected volume kube-api-access-79v97 for pod openshift-network-diagnostics/network-check-target-cm2lg: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 20:45:08.851497 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:08.851409 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ac379788-302d-4181-8970-56f47ae312c1-kube-api-access-79v97 podName:ac379788-302d-4181-8970-56f47ae312c1 nodeName:}" failed. No retries permitted until 2026-04-17 20:45:40.851395408 +0000 UTC m=+66.325173672 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-79v97" (UniqueName: "kubernetes.io/projected/ac379788-302d-4181-8970-56f47ae312c1-kube-api-access-79v97") pod "network-check-target-cm2lg" (UID: "ac379788-302d-4181-8970-56f47ae312c1") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Apr 17 20:45:09.119141 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:09.119109 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:45:09.119317 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:09.119118 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:45:09.122475 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:09.122286 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Apr 17 20:45:09.122475 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:09.122298 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Apr 17 20:45:09.122475 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:09.122308 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-jzn7k\"" Apr 17 20:45:09.122475 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:09.122345 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Apr 17 20:45:09.122475 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:09.122286 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-diagnostics\"/\"default-dockercfg-nb4cm\"" Apr 17 20:45:09.153749 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:09.153725 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79b05e51-04fa-4473-b612-c762aa0e3856-cert\") pod \"ingress-canary-q59g7\" (UID: \"79b05e51-04fa-4473-b612-c762aa0e3856\") " pod="openshift-ingress-canary/ingress-canary-q59g7" Apr 17 20:45:09.153856 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:09.153776 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/821ba3d4-448f-45b9-bcec-701682592a0f-metrics-tls\") pod \"dns-default-bc6wf\" (UID: \"821ba3d4-448f-45b9-bcec-701682592a0f\") " pod="openshift-dns/dns-default-bc6wf" Apr 17 20:45:09.153920 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:09.153902 2577 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 17 20:45:09.153920 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:09.153908 2577 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 17 20:45:09.153998 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:09.153966 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/821ba3d4-448f-45b9-bcec-701682592a0f-metrics-tls podName:821ba3d4-448f-45b9-bcec-701682592a0f nodeName:}" failed. No retries permitted until 2026-04-17 20:45:10.153951751 +0000 UTC m=+35.627730016 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/821ba3d4-448f-45b9-bcec-701682592a0f-metrics-tls") pod "dns-default-bc6wf" (UID: "821ba3d4-448f-45b9-bcec-701682592a0f") : secret "dns-default-metrics-tls" not found Apr 17 20:45:09.153998 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:09.153984 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79b05e51-04fa-4473-b612-c762aa0e3856-cert podName:79b05e51-04fa-4473-b612-c762aa0e3856 nodeName:}" failed. No retries permitted until 2026-04-17 20:45:10.153975563 +0000 UTC m=+35.627753834 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/79b05e51-04fa-4473-b612-c762aa0e3856-cert") pod "ingress-canary-q59g7" (UID: "79b05e51-04fa-4473-b612-c762aa0e3856") : secret "canary-serving-cert" not found Apr 17 20:45:10.162178 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:10.162144 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79b05e51-04fa-4473-b612-c762aa0e3856-cert\") pod \"ingress-canary-q59g7\" (UID: \"79b05e51-04fa-4473-b612-c762aa0e3856\") " pod="openshift-ingress-canary/ingress-canary-q59g7" Apr 17 20:45:10.162608 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:10.162185 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/821ba3d4-448f-45b9-bcec-701682592a0f-metrics-tls\") pod \"dns-default-bc6wf\" (UID: \"821ba3d4-448f-45b9-bcec-701682592a0f\") " pod="openshift-dns/dns-default-bc6wf" Apr 17 20:45:10.162608 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:10.162299 2577 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 17 20:45:10.162608 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:10.162299 2577 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 17 20:45:10.162608 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:10.162361 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/821ba3d4-448f-45b9-bcec-701682592a0f-metrics-tls podName:821ba3d4-448f-45b9-bcec-701682592a0f nodeName:}" failed. No retries permitted until 2026-04-17 20:45:12.162344608 +0000 UTC m=+37.636122882 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/821ba3d4-448f-45b9-bcec-701682592a0f-metrics-tls") pod "dns-default-bc6wf" (UID: "821ba3d4-448f-45b9-bcec-701682592a0f") : secret "dns-default-metrics-tls" not found Apr 17 20:45:10.162608 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:10.162378 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79b05e51-04fa-4473-b612-c762aa0e3856-cert podName:79b05e51-04fa-4473-b612-c762aa0e3856 nodeName:}" failed. No retries permitted until 2026-04-17 20:45:12.16236985 +0000 UTC m=+37.636148115 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/79b05e51-04fa-4473-b612-c762aa0e3856-cert") pod "ingress-canary-q59g7" (UID: "79b05e51-04fa-4473-b612-c762aa0e3856") : secret "canary-serving-cert" not found Apr 17 20:45:11.266953 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:11.266917 2577 generic.go:358] "Generic (PLEG): container finished" podID="06332f89-1179-4cde-9a2b-4335714933a1" containerID="a74be867d9c2b7a74a8a02c2f05d36e70b6344addb29ec7a08d2f709a4cb8954" exitCode=0 Apr 17 20:45:11.267475 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:11.266986 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zds6c" event={"ID":"06332f89-1179-4cde-9a2b-4335714933a1","Type":"ContainerDied","Data":"a74be867d9c2b7a74a8a02c2f05d36e70b6344addb29ec7a08d2f709a4cb8954"} Apr 17 20:45:12.175761 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:12.175733 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/821ba3d4-448f-45b9-bcec-701682592a0f-metrics-tls\") pod \"dns-default-bc6wf\" (UID: \"821ba3d4-448f-45b9-bcec-701682592a0f\") " pod="openshift-dns/dns-default-bc6wf" Apr 17 20:45:12.175929 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:12.175794 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79b05e51-04fa-4473-b612-c762aa0e3856-cert\") pod \"ingress-canary-q59g7\" (UID: \"79b05e51-04fa-4473-b612-c762aa0e3856\") " pod="openshift-ingress-canary/ingress-canary-q59g7" Apr 17 20:45:12.175929 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:12.175881 2577 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 17 20:45:12.175929 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:12.175888 2577 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 17 20:45:12.176020 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:12.175936 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/821ba3d4-448f-45b9-bcec-701682592a0f-metrics-tls podName:821ba3d4-448f-45b9-bcec-701682592a0f nodeName:}" failed. No retries permitted until 2026-04-17 20:45:16.175921785 +0000 UTC m=+41.649700049 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/821ba3d4-448f-45b9-bcec-701682592a0f-metrics-tls") pod "dns-default-bc6wf" (UID: "821ba3d4-448f-45b9-bcec-701682592a0f") : secret "dns-default-metrics-tls" not found Apr 17 20:45:12.176020 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:12.175949 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79b05e51-04fa-4473-b612-c762aa0e3856-cert podName:79b05e51-04fa-4473-b612-c762aa0e3856 nodeName:}" failed. No retries permitted until 2026-04-17 20:45:16.17594322 +0000 UTC m=+41.649721485 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/79b05e51-04fa-4473-b612-c762aa0e3856-cert") pod "ingress-canary-q59g7" (UID: "79b05e51-04fa-4473-b612-c762aa0e3856") : secret "canary-serving-cert" not found Apr 17 20:45:12.271401 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:12.271377 2577 generic.go:358] "Generic (PLEG): container finished" podID="06332f89-1179-4cde-9a2b-4335714933a1" containerID="05c64cffed10666ed864fc828d1c0e9d82cf47aef181eaad75366fecd205d353" exitCode=0 Apr 17 20:45:12.271887 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:12.271428 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zds6c" event={"ID":"06332f89-1179-4cde-9a2b-4335714933a1","Type":"ContainerDied","Data":"05c64cffed10666ed864fc828d1c0e9d82cf47aef181eaad75366fecd205d353"} Apr 17 20:45:13.278489 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:13.278306 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zds6c" event={"ID":"06332f89-1179-4cde-9a2b-4335714933a1","Type":"ContainerStarted","Data":"8e5f06e5d03d6a56580ff3fae26b76ded2543a4df9a2330edb3c77d31c889590"} Apr 17 20:45:13.297311 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:13.297264 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-zds6c" podStartSLOduration=5.583172799 podStartE2EDuration="38.297251664s" podCreationTimestamp="2026-04-17 20:44:35 +0000 UTC" firstStartedPulling="2026-04-17 20:44:37.764161467 +0000 UTC m=+3.237939733" lastFinishedPulling="2026-04-17 20:45:10.478240333 +0000 UTC m=+35.952018598" observedRunningTime="2026-04-17 20:45:13.29622767 +0000 UTC m=+38.770005959" watchObservedRunningTime="2026-04-17 20:45:13.297251664 +0000 UTC m=+38.771029951" Apr 17 20:45:16.202662 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:16.202619 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79b05e51-04fa-4473-b612-c762aa0e3856-cert\") pod \"ingress-canary-q59g7\" (UID: \"79b05e51-04fa-4473-b612-c762aa0e3856\") " pod="openshift-ingress-canary/ingress-canary-q59g7" Apr 17 20:45:16.202662 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:16.202667 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/821ba3d4-448f-45b9-bcec-701682592a0f-metrics-tls\") pod \"dns-default-bc6wf\" (UID: \"821ba3d4-448f-45b9-bcec-701682592a0f\") " pod="openshift-dns/dns-default-bc6wf" Apr 17 20:45:16.203118 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:16.202772 2577 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 17 20:45:16.203118 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:16.202792 2577 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 17 20:45:16.203118 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:16.202860 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79b05e51-04fa-4473-b612-c762aa0e3856-cert podName:79b05e51-04fa-4473-b612-c762aa0e3856 nodeName:}" failed. No retries permitted until 2026-04-17 20:45:24.202842939 +0000 UTC m=+49.676621208 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/79b05e51-04fa-4473-b612-c762aa0e3856-cert") pod "ingress-canary-q59g7" (UID: "79b05e51-04fa-4473-b612-c762aa0e3856") : secret "canary-serving-cert" not found Apr 17 20:45:16.203118 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:16.202874 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/821ba3d4-448f-45b9-bcec-701682592a0f-metrics-tls podName:821ba3d4-448f-45b9-bcec-701682592a0f nodeName:}" failed. No retries permitted until 2026-04-17 20:45:24.202867527 +0000 UTC m=+49.676645791 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/821ba3d4-448f-45b9-bcec-701682592a0f-metrics-tls") pod "dns-default-bc6wf" (UID: "821ba3d4-448f-45b9-bcec-701682592a0f") : secret "dns-default-metrics-tls" not found Apr 17 20:45:24.252699 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:24.252663 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79b05e51-04fa-4473-b612-c762aa0e3856-cert\") pod \"ingress-canary-q59g7\" (UID: \"79b05e51-04fa-4473-b612-c762aa0e3856\") " pod="openshift-ingress-canary/ingress-canary-q59g7" Apr 17 20:45:24.252699 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:24.252703 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/821ba3d4-448f-45b9-bcec-701682592a0f-metrics-tls\") pod \"dns-default-bc6wf\" (UID: \"821ba3d4-448f-45b9-bcec-701682592a0f\") " pod="openshift-dns/dns-default-bc6wf" Apr 17 20:45:24.253205 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:24.252828 2577 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 17 20:45:24.253205 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:24.252869 2577 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 17 20:45:24.253205 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:24.252883 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/821ba3d4-448f-45b9-bcec-701682592a0f-metrics-tls podName:821ba3d4-448f-45b9-bcec-701682592a0f nodeName:}" failed. No retries permitted until 2026-04-17 20:45:40.252869424 +0000 UTC m=+65.726647689 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/821ba3d4-448f-45b9-bcec-701682592a0f-metrics-tls") pod "dns-default-bc6wf" (UID: "821ba3d4-448f-45b9-bcec-701682592a0f") : secret "dns-default-metrics-tls" not found Apr 17 20:45:24.253205 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:24.252954 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79b05e51-04fa-4473-b612-c762aa0e3856-cert podName:79b05e51-04fa-4473-b612-c762aa0e3856 nodeName:}" failed. No retries permitted until 2026-04-17 20:45:40.252935397 +0000 UTC m=+65.726713679 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/79b05e51-04fa-4473-b612-c762aa0e3856-cert") pod "ingress-canary-q59g7" (UID: "79b05e51-04fa-4473-b612-c762aa0e3856") : secret "canary-serving-cert" not found Apr 17 20:45:35.020895 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:35.020861 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dk97h" Apr 17 20:45:40.260356 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:40.260322 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79b05e51-04fa-4473-b612-c762aa0e3856-cert\") pod \"ingress-canary-q59g7\" (UID: \"79b05e51-04fa-4473-b612-c762aa0e3856\") " pod="openshift-ingress-canary/ingress-canary-q59g7" Apr 17 20:45:40.260356 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:40.260359 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/821ba3d4-448f-45b9-bcec-701682592a0f-metrics-tls\") pod \"dns-default-bc6wf\" (UID: \"821ba3d4-448f-45b9-bcec-701682592a0f\") " pod="openshift-dns/dns-default-bc6wf" Apr 17 20:45:40.260885 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:40.260452 2577 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 17 20:45:40.260885 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:40.260457 2577 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 17 20:45:40.260885 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:40.260503 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/821ba3d4-448f-45b9-bcec-701682592a0f-metrics-tls podName:821ba3d4-448f-45b9-bcec-701682592a0f nodeName:}" failed. No retries permitted until 2026-04-17 20:46:12.260488665 +0000 UTC m=+97.734266930 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/821ba3d4-448f-45b9-bcec-701682592a0f-metrics-tls") pod "dns-default-bc6wf" (UID: "821ba3d4-448f-45b9-bcec-701682592a0f") : secret "dns-default-metrics-tls" not found Apr 17 20:45:40.260885 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:40.260515 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79b05e51-04fa-4473-b612-c762aa0e3856-cert podName:79b05e51-04fa-4473-b612-c762aa0e3856 nodeName:}" failed. No retries permitted until 2026-04-17 20:46:12.260509503 +0000 UTC m=+97.734287767 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/79b05e51-04fa-4473-b612-c762aa0e3856-cert") pod "ingress-canary-q59g7" (UID: "79b05e51-04fa-4473-b612-c762aa0e3856") : secret "canary-serving-cert" not found Apr 17 20:45:40.764179 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:40.764138 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1a6721ad-2d96-4d01-a9db-b8654256e62f-metrics-certs\") pod \"network-metrics-daemon-7ddt5\" (UID: \"1a6721ad-2d96-4d01-a9db-b8654256e62f\") " pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:45:40.766326 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:40.766305 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Apr 17 20:45:40.774834 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:40.774797 2577 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: secret "metrics-daemon-secret" not found Apr 17 20:45:40.774925 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:45:40.774876 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1a6721ad-2d96-4d01-a9db-b8654256e62f-metrics-certs podName:1a6721ad-2d96-4d01-a9db-b8654256e62f nodeName:}" failed. No retries permitted until 2026-04-17 20:46:44.774856034 +0000 UTC m=+130.248634299 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1a6721ad-2d96-4d01-a9db-b8654256e62f-metrics-certs") pod "network-metrics-daemon-7ddt5" (UID: "1a6721ad-2d96-4d01-a9db-b8654256e62f") : secret "metrics-daemon-secret" not found Apr 17 20:45:40.865117 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:40.865090 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-79v97\" (UniqueName: \"kubernetes.io/projected/ac379788-302d-4181-8970-56f47ae312c1-kube-api-access-79v97\") pod \"network-check-target-cm2lg\" (UID: \"ac379788-302d-4181-8970-56f47ae312c1\") " pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:45:40.867088 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:40.867069 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Apr 17 20:45:40.876903 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:40.876884 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Apr 17 20:45:40.889533 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:40.889505 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-79v97\" (UniqueName: \"kubernetes.io/projected/ac379788-302d-4181-8970-56f47ae312c1-kube-api-access-79v97\") pod \"network-check-target-cm2lg\" (UID: \"ac379788-302d-4181-8970-56f47ae312c1\") " pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:45:40.933106 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:40.933082 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-diagnostics\"/\"default-dockercfg-nb4cm\"" Apr 17 20:45:40.942533 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:40.942519 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:45:41.108997 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:41.108969 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-target-cm2lg"] Apr 17 20:45:41.112276 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:45:41.112247 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac379788_302d_4181_8970_56f47ae312c1.slice/crio-e33eafc59d3fa6a696f840e5aed74fe8cc90fae8ca9a5dbb815ebf4039929bbb WatchSource:0}: Error finding container e33eafc59d3fa6a696f840e5aed74fe8cc90fae8ca9a5dbb815ebf4039929bbb: Status 404 returned error can't find the container with id e33eafc59d3fa6a696f840e5aed74fe8cc90fae8ca9a5dbb815ebf4039929bbb Apr 17 20:45:41.329423 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:41.329346 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-cm2lg" event={"ID":"ac379788-302d-4181-8970-56f47ae312c1","Type":"ContainerStarted","Data":"e33eafc59d3fa6a696f840e5aed74fe8cc90fae8ca9a5dbb815ebf4039929bbb"} Apr 17 20:45:44.336000 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:44.335965 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-cm2lg" event={"ID":"ac379788-302d-4181-8970-56f47ae312c1","Type":"ContainerStarted","Data":"497bca8dd55b0b84a25b3b86ef1d9761c3b608602e04dfc51bf285785645194a"} Apr 17 20:45:44.336364 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:44.336087 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:45:44.349452 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:45:44.349363 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-diagnostics/network-check-target-cm2lg" podStartSLOduration=66.622830748 podStartE2EDuration="1m9.349347092s" podCreationTimestamp="2026-04-17 20:44:35 +0000 UTC" firstStartedPulling="2026-04-17 20:45:41.114048562 +0000 UTC m=+66.587826827" lastFinishedPulling="2026-04-17 20:45:43.840564903 +0000 UTC m=+69.314343171" observedRunningTime="2026-04-17 20:45:44.348876503 +0000 UTC m=+69.822654790" watchObservedRunningTime="2026-04-17 20:45:44.349347092 +0000 UTC m=+69.823125372" Apr 17 20:46:12.264072 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:12.263939 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79b05e51-04fa-4473-b612-c762aa0e3856-cert\") pod \"ingress-canary-q59g7\" (UID: \"79b05e51-04fa-4473-b612-c762aa0e3856\") " pod="openshift-ingress-canary/ingress-canary-q59g7" Apr 17 20:46:12.264072 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:12.263984 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/821ba3d4-448f-45b9-bcec-701682592a0f-metrics-tls\") pod \"dns-default-bc6wf\" (UID: \"821ba3d4-448f-45b9-bcec-701682592a0f\") " pod="openshift-dns/dns-default-bc6wf" Apr 17 20:46:12.264072 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:12.264046 2577 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Apr 17 20:46:12.264072 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:12.264068 2577 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Apr 17 20:46:12.264621 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:12.264124 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/821ba3d4-448f-45b9-bcec-701682592a0f-metrics-tls podName:821ba3d4-448f-45b9-bcec-701682592a0f nodeName:}" failed. No retries permitted until 2026-04-17 20:47:16.264110758 +0000 UTC m=+161.737889022 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/821ba3d4-448f-45b9-bcec-701682592a0f-metrics-tls") pod "dns-default-bc6wf" (UID: "821ba3d4-448f-45b9-bcec-701682592a0f") : secret "dns-default-metrics-tls" not found Apr 17 20:46:12.264621 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:12.264138 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/79b05e51-04fa-4473-b612-c762aa0e3856-cert podName:79b05e51-04fa-4473-b612-c762aa0e3856 nodeName:}" failed. No retries permitted until 2026-04-17 20:47:16.264131135 +0000 UTC m=+161.737909399 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/79b05e51-04fa-4473-b612-c762aa0e3856-cert") pod "ingress-canary-q59g7" (UID: "79b05e51-04fa-4473-b612-c762aa0e3856") : secret "canary-serving-cert" not found Apr 17 20:46:15.340064 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:15.340034 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-cm2lg" Apr 17 20:46:20.311212 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.311173 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-dhmj2"] Apr 17 20:46:20.315760 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.315735 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-dhmj2" Apr 17 20:46:20.316061 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.316030 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-rz8d2"] Apr 17 20:46:20.318389 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.318352 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-storage-operator\"/\"volume-data-source-validator-dockercfg-7v6xq\"" Apr 17 20:46:20.318850 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.318827 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-storage-operator\"/\"openshift-service-ca.crt\"" Apr 17 20:46:20.319010 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.318978 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-storage-operator\"/\"kube-root-ca.crt\"" Apr 17 20:46:20.320738 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.320707 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-rz8d2" Apr 17 20:46:20.323136 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.323111 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"samples-operator-tls\"" Apr 17 20:46:20.323237 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.323212 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"cluster-samples-operator-dockercfg-g4s6r\"" Apr 17 20:46:20.323298 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.323234 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-samples-operator\"/\"openshift-service-ca.crt\"" Apr 17 20:46:20.323349 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.323293 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-samples-operator\"/\"kube-root-ca.crt\"" Apr 17 20:46:20.326886 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.326863 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-insights/insights-operator-585dfdc468-5jzh4"] Apr 17 20:46:20.329651 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.329630 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-operator-585dfdc468-5jzh4" Apr 17 20:46:20.331932 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.331911 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"operator-dockercfg-7dgc7\"" Apr 17 20:46:20.332037 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.331913 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"openshift-service-ca.crt\"" Apr 17 20:46:20.332221 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.332205 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"openshift-insights-serving-cert\"" Apr 17 20:46:20.340172 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.340152 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"kube-root-ca.crt\"" Apr 17 20:46:20.340551 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.340530 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"service-ca-bundle\"" Apr 17 20:46:20.341451 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.341428 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-dhmj2"] Apr 17 20:46:20.342092 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.342066 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-rz8d2"] Apr 17 20:46:20.347027 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.347005 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-operator-585dfdc468-5jzh4"] Apr 17 20:46:20.349061 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.349022 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"trusted-ca-bundle\"" Apr 17 20:46:20.400861 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.400836 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-d6fc45fc5-rxj8l"] Apr 17 20:46:20.403675 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.403658 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-75587bd455-7cqn6"] Apr 17 20:46:20.403862 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.403844 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-rxj8l" Apr 17 20:46:20.405491 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.405468 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca-operator\"/\"openshift-service-ca.crt\"" Apr 17 20:46:20.405679 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.405659 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca-operator\"/\"service-ca-operator-config\"" Apr 17 20:46:20.405786 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.405768 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca-operator\"/\"kube-root-ca.crt\"" Apr 17 20:46:20.405890 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.405876 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca-operator\"/\"service-ca-operator-dockercfg-xqz96\"" Apr 17 20:46:20.405995 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.405981 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca-operator\"/\"serving-cert\"" Apr 17 20:46:20.406614 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.406598 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-9d4b6777b-9rmnh"] Apr 17 20:46:20.406741 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.406728 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-7cqn6" Apr 17 20:46:20.408334 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.408317 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"cluster-monitoring-operator-tls\"" Apr 17 20:46:20.408420 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.408384 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"openshift-service-ca.crt\"" Apr 17 20:46:20.408519 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.408504 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"kube-root-ca.crt\"" Apr 17 20:46:20.408635 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.408621 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"cluster-monitoring-operator-dockercfg-kx4g8\"" Apr 17 20:46:20.409313 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.409291 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-fs4ts"] Apr 17 20:46:20.409414 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.409398 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"telemetry-config\"" Apr 17 20:46:20.409514 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.409494 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" Apr 17 20:46:20.411429 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.411410 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console-operator\"/\"console-operator-dockercfg-htk46\"" Apr 17 20:46:20.411569 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.411548 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console-operator\"/\"serving-cert\"" Apr 17 20:46:20.411642 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.411490 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"console-operator-config\"" Apr 17 20:46:20.411734 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.411507 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"openshift-service-ca.crt\"" Apr 17 20:46:20.411734 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.411462 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"kube-root-ca.crt\"" Apr 17 20:46:20.412157 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.412142 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-fs4ts" Apr 17 20:46:20.412634 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.412568 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-d6fc45fc5-rxj8l"] Apr 17 20:46:20.413673 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.413653 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-75587bd455-7cqn6"] Apr 17 20:46:20.414285 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.414267 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"serving-cert\"" Apr 17 20:46:20.415536 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.415518 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"kube-root-ca.crt\"" Apr 17 20:46:20.416600 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.416576 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"config\"" Apr 17 20:46:20.416757 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.416736 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"kube-storage-version-migrator-operator-dockercfg-794fp\"" Apr 17 20:46:20.418741 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.418722 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-fs4ts"] Apr 17 20:46:20.419857 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.419835 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18ff00e6-c4f7-4852-b689-94ec992eb400-serving-cert\") pod \"insights-operator-585dfdc468-5jzh4\" (UID: \"18ff00e6-c4f7-4852-b689-94ec992eb400\") " pod="openshift-insights/insights-operator-585dfdc468-5jzh4" Apr 17 20:46:20.420135 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.420116 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5f4b3fb0-6561-4993-a933-390f93a354ea-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-rz8d2\" (UID: \"5f4b3fb0-6561-4993-a933-390f93a354ea\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-rz8d2" Apr 17 20:46:20.420231 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.420210 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"openshift-service-ca.crt\"" Apr 17 20:46:20.420338 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.420321 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pkmh\" (UniqueName: \"kubernetes.io/projected/2cfc1b29-dfac-4fb3-878c-0961d357f28e-kube-api-access-2pkmh\") pod \"volume-data-source-validator-7c6cbb6c87-dhmj2\" (UID: \"2cfc1b29-dfac-4fb3-878c-0961d357f28e\") " pod="openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-dhmj2" Apr 17 20:46:20.420490 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.420472 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/18ff00e6-c4f7-4852-b689-94ec992eb400-trusted-ca-bundle\") pod \"insights-operator-585dfdc468-5jzh4\" (UID: \"18ff00e6-c4f7-4852-b689-94ec992eb400\") " pod="openshift-insights/insights-operator-585dfdc468-5jzh4" Apr 17 20:46:20.420621 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.420605 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/18ff00e6-c4f7-4852-b689-94ec992eb400-tmp\") pod \"insights-operator-585dfdc468-5jzh4\" (UID: \"18ff00e6-c4f7-4852-b689-94ec992eb400\") " pod="openshift-insights/insights-operator-585dfdc468-5jzh4" Apr 17 20:46:20.420773 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.420748 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jz2j9\" (UniqueName: \"kubernetes.io/projected/18ff00e6-c4f7-4852-b689-94ec992eb400-kube-api-access-jz2j9\") pod \"insights-operator-585dfdc468-5jzh4\" (UID: \"18ff00e6-c4f7-4852-b689-94ec992eb400\") " pod="openshift-insights/insights-operator-585dfdc468-5jzh4" Apr 17 20:46:20.421013 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.420975 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"snapshots\" (UniqueName: \"kubernetes.io/empty-dir/18ff00e6-c4f7-4852-b689-94ec992eb400-snapshots\") pod \"insights-operator-585dfdc468-5jzh4\" (UID: \"18ff00e6-c4f7-4852-b689-94ec992eb400\") " pod="openshift-insights/insights-operator-585dfdc468-5jzh4" Apr 17 20:46:20.421203 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.421185 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/18ff00e6-c4f7-4852-b689-94ec992eb400-service-ca-bundle\") pod \"insights-operator-585dfdc468-5jzh4\" (UID: \"18ff00e6-c4f7-4852-b689-94ec992eb400\") " pod="openshift-insights/insights-operator-585dfdc468-5jzh4" Apr 17 20:46:20.421381 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.421353 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bm92\" (UniqueName: \"kubernetes.io/projected/5f4b3fb0-6561-4993-a933-390f93a354ea-kube-api-access-8bm92\") pod \"cluster-samples-operator-6dc5bdb6b4-rz8d2\" (UID: \"5f4b3fb0-6561-4993-a933-390f93a354ea\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-rz8d2" Apr 17 20:46:20.421503 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.421485 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-9d4b6777b-9rmnh"] Apr 17 20:46:20.421716 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.421683 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"trusted-ca\"" Apr 17 20:46:20.522309 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.522284 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/18ff00e6-c4f7-4852-b689-94ec992eb400-service-ca-bundle\") pod \"insights-operator-585dfdc468-5jzh4\" (UID: \"18ff00e6-c4f7-4852-b689-94ec992eb400\") " pod="openshift-insights/insights-operator-585dfdc468-5jzh4" Apr 17 20:46:20.522406 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.522314 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1981af6-ac30-4f6a-b971-bb5e2e4f8b1e-config\") pod \"service-ca-operator-d6fc45fc5-rxj8l\" (UID: \"f1981af6-ac30-4f6a-b971-bb5e2e4f8b1e\") " pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-rxj8l" Apr 17 20:46:20.522406 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.522332 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/5506563e-216d-4b61-97dc-8cea5be3f474-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-7cqn6\" (UID: \"5506563e-216d-4b61-97dc-8cea5be3f474\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-7cqn6" Apr 17 20:46:20.522406 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.522393 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8bm92\" (UniqueName: \"kubernetes.io/projected/5f4b3fb0-6561-4993-a933-390f93a354ea-kube-api-access-8bm92\") pod \"cluster-samples-operator-6dc5bdb6b4-rz8d2\" (UID: \"5f4b3fb0-6561-4993-a933-390f93a354ea\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-rz8d2" Apr 17 20:46:20.522528 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.522428 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hkk7\" (UniqueName: \"kubernetes.io/projected/f1981af6-ac30-4f6a-b971-bb5e2e4f8b1e-kube-api-access-8hkk7\") pod \"service-ca-operator-d6fc45fc5-rxj8l\" (UID: \"f1981af6-ac30-4f6a-b971-bb5e2e4f8b1e\") " pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-rxj8l" Apr 17 20:46:20.522528 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.522467 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18ff00e6-c4f7-4852-b689-94ec992eb400-serving-cert\") pod \"insights-operator-585dfdc468-5jzh4\" (UID: \"18ff00e6-c4f7-4852-b689-94ec992eb400\") " pod="openshift-insights/insights-operator-585dfdc468-5jzh4" Apr 17 20:46:20.522528 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.522487 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5f4b3fb0-6561-4993-a933-390f93a354ea-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-rz8d2\" (UID: \"5f4b3fb0-6561-4993-a933-390f93a354ea\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-rz8d2" Apr 17 20:46:20.522528 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.522517 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/5506563e-216d-4b61-97dc-8cea5be3f474-telemetry-config\") pod \"cluster-monitoring-operator-75587bd455-7cqn6\" (UID: \"5506563e-216d-4b61-97dc-8cea5be3f474\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-7cqn6" Apr 17 20:46:20.522715 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.522547 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-2pkmh\" (UniqueName: \"kubernetes.io/projected/2cfc1b29-dfac-4fb3-878c-0961d357f28e-kube-api-access-2pkmh\") pod \"volume-data-source-validator-7c6cbb6c87-dhmj2\" (UID: \"2cfc1b29-dfac-4fb3-878c-0961d357f28e\") " pod="openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-dhmj2" Apr 17 20:46:20.522715 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.522578 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05c5418a-43f4-476b-8e92-b7b56513d5af-config\") pod \"kube-storage-version-migrator-operator-6769c5d45-fs4ts\" (UID: \"05c5418a-43f4-476b-8e92-b7b56513d5af\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-fs4ts" Apr 17 20:46:20.522715 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:20.522598 2577 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Apr 17 20:46:20.522715 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:20.522671 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f4b3fb0-6561-4993-a933-390f93a354ea-samples-operator-tls podName:5f4b3fb0-6561-4993-a933-390f93a354ea nodeName:}" failed. No retries permitted until 2026-04-17 20:46:21.022651705 +0000 UTC m=+106.496429987 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/5f4b3fb0-6561-4993-a933-390f93a354ea-samples-operator-tls") pod "cluster-samples-operator-6dc5bdb6b4-rz8d2" (UID: "5f4b3fb0-6561-4993-a933-390f93a354ea") : secret "samples-operator-tls" not found Apr 17 20:46:20.522715 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.522603 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr7w2\" (UniqueName: \"kubernetes.io/projected/5506563e-216d-4b61-97dc-8cea5be3f474-kube-api-access-wr7w2\") pod \"cluster-monitoring-operator-75587bd455-7cqn6\" (UID: \"5506563e-216d-4b61-97dc-8cea5be3f474\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-7cqn6" Apr 17 20:46:20.523055 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.522792 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e562e939-99f0-4bee-952d-7af46132c860-serving-cert\") pod \"console-operator-9d4b6777b-9rmnh\" (UID: \"e562e939-99f0-4bee-952d-7af46132c860\") " pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" Apr 17 20:46:20.523055 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.522847 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/18ff00e6-c4f7-4852-b689-94ec992eb400-trusted-ca-bundle\") pod \"insights-operator-585dfdc468-5jzh4\" (UID: \"18ff00e6-c4f7-4852-b689-94ec992eb400\") " pod="openshift-insights/insights-operator-585dfdc468-5jzh4" Apr 17 20:46:20.523055 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.522864 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/18ff00e6-c4f7-4852-b689-94ec992eb400-service-ca-bundle\") pod \"insights-operator-585dfdc468-5jzh4\" (UID: \"18ff00e6-c4f7-4852-b689-94ec992eb400\") " pod="openshift-insights/insights-operator-585dfdc468-5jzh4" Apr 17 20:46:20.523055 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.522884 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e562e939-99f0-4bee-952d-7af46132c860-trusted-ca\") pod \"console-operator-9d4b6777b-9rmnh\" (UID: \"e562e939-99f0-4bee-952d-7af46132c860\") " pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" Apr 17 20:46:20.523055 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.522912 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77cdc\" (UniqueName: \"kubernetes.io/projected/05c5418a-43f4-476b-8e92-b7b56513d5af-kube-api-access-77cdc\") pod \"kube-storage-version-migrator-operator-6769c5d45-fs4ts\" (UID: \"05c5418a-43f4-476b-8e92-b7b56513d5af\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-fs4ts" Apr 17 20:46:20.523055 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.522938 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e562e939-99f0-4bee-952d-7af46132c860-config\") pod \"console-operator-9d4b6777b-9rmnh\" (UID: \"e562e939-99f0-4bee-952d-7af46132c860\") " pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" Apr 17 20:46:20.523055 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.522969 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1981af6-ac30-4f6a-b971-bb5e2e4f8b1e-serving-cert\") pod \"service-ca-operator-d6fc45fc5-rxj8l\" (UID: \"f1981af6-ac30-4f6a-b971-bb5e2e4f8b1e\") " pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-rxj8l" Apr 17 20:46:20.523055 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.523001 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/18ff00e6-c4f7-4852-b689-94ec992eb400-tmp\") pod \"insights-operator-585dfdc468-5jzh4\" (UID: \"18ff00e6-c4f7-4852-b689-94ec992eb400\") " pod="openshift-insights/insights-operator-585dfdc468-5jzh4" Apr 17 20:46:20.523055 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.523026 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-jz2j9\" (UniqueName: \"kubernetes.io/projected/18ff00e6-c4f7-4852-b689-94ec992eb400-kube-api-access-jz2j9\") pod \"insights-operator-585dfdc468-5jzh4\" (UID: \"18ff00e6-c4f7-4852-b689-94ec992eb400\") " pod="openshift-insights/insights-operator-585dfdc468-5jzh4" Apr 17 20:46:20.523055 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.523054 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvjct\" (UniqueName: \"kubernetes.io/projected/e562e939-99f0-4bee-952d-7af46132c860-kube-api-access-vvjct\") pod \"console-operator-9d4b6777b-9rmnh\" (UID: \"e562e939-99f0-4bee-952d-7af46132c860\") " pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" Apr 17 20:46:20.523496 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.523102 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"snapshots\" (UniqueName: \"kubernetes.io/empty-dir/18ff00e6-c4f7-4852-b689-94ec992eb400-snapshots\") pod \"insights-operator-585dfdc468-5jzh4\" (UID: \"18ff00e6-c4f7-4852-b689-94ec992eb400\") " pod="openshift-insights/insights-operator-585dfdc468-5jzh4" Apr 17 20:46:20.523496 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.523257 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05c5418a-43f4-476b-8e92-b7b56513d5af-serving-cert\") pod \"kube-storage-version-migrator-operator-6769c5d45-fs4ts\" (UID: \"05c5418a-43f4-476b-8e92-b7b56513d5af\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-fs4ts" Apr 17 20:46:20.523496 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.523473 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/18ff00e6-c4f7-4852-b689-94ec992eb400-tmp\") pod \"insights-operator-585dfdc468-5jzh4\" (UID: \"18ff00e6-c4f7-4852-b689-94ec992eb400\") " pod="openshift-insights/insights-operator-585dfdc468-5jzh4" Apr 17 20:46:20.523627 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.523557 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"snapshots\" (UniqueName: \"kubernetes.io/empty-dir/18ff00e6-c4f7-4852-b689-94ec992eb400-snapshots\") pod \"insights-operator-585dfdc468-5jzh4\" (UID: \"18ff00e6-c4f7-4852-b689-94ec992eb400\") " pod="openshift-insights/insights-operator-585dfdc468-5jzh4" Apr 17 20:46:20.523627 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.523596 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/18ff00e6-c4f7-4852-b689-94ec992eb400-trusted-ca-bundle\") pod \"insights-operator-585dfdc468-5jzh4\" (UID: \"18ff00e6-c4f7-4852-b689-94ec992eb400\") " pod="openshift-insights/insights-operator-585dfdc468-5jzh4" Apr 17 20:46:20.524907 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.524887 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18ff00e6-c4f7-4852-b689-94ec992eb400-serving-cert\") pod \"insights-operator-585dfdc468-5jzh4\" (UID: \"18ff00e6-c4f7-4852-b689-94ec992eb400\") " pod="openshift-insights/insights-operator-585dfdc468-5jzh4" Apr 17 20:46:20.530256 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.530238 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bm92\" (UniqueName: \"kubernetes.io/projected/5f4b3fb0-6561-4993-a933-390f93a354ea-kube-api-access-8bm92\") pod \"cluster-samples-operator-6dc5bdb6b4-rz8d2\" (UID: \"5f4b3fb0-6561-4993-a933-390f93a354ea\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-rz8d2" Apr 17 20:46:20.530479 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.530457 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pkmh\" (UniqueName: \"kubernetes.io/projected/2cfc1b29-dfac-4fb3-878c-0961d357f28e-kube-api-access-2pkmh\") pod \"volume-data-source-validator-7c6cbb6c87-dhmj2\" (UID: \"2cfc1b29-dfac-4fb3-878c-0961d357f28e\") " pod="openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-dhmj2" Apr 17 20:46:20.530519 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.530497 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-jz2j9\" (UniqueName: \"kubernetes.io/projected/18ff00e6-c4f7-4852-b689-94ec992eb400-kube-api-access-jz2j9\") pod \"insights-operator-585dfdc468-5jzh4\" (UID: \"18ff00e6-c4f7-4852-b689-94ec992eb400\") " pod="openshift-insights/insights-operator-585dfdc468-5jzh4" Apr 17 20:46:20.623519 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.623469 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/5506563e-216d-4b61-97dc-8cea5be3f474-telemetry-config\") pod \"cluster-monitoring-operator-75587bd455-7cqn6\" (UID: \"5506563e-216d-4b61-97dc-8cea5be3f474\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-7cqn6" Apr 17 20:46:20.623519 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.623500 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05c5418a-43f4-476b-8e92-b7b56513d5af-config\") pod \"kube-storage-version-migrator-operator-6769c5d45-fs4ts\" (UID: \"05c5418a-43f4-476b-8e92-b7b56513d5af\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-fs4ts" Apr 17 20:46:20.623519 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.623516 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-wr7w2\" (UniqueName: \"kubernetes.io/projected/5506563e-216d-4b61-97dc-8cea5be3f474-kube-api-access-wr7w2\") pod \"cluster-monitoring-operator-75587bd455-7cqn6\" (UID: \"5506563e-216d-4b61-97dc-8cea5be3f474\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-7cqn6" Apr 17 20:46:20.623713 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.623532 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e562e939-99f0-4bee-952d-7af46132c860-serving-cert\") pod \"console-operator-9d4b6777b-9rmnh\" (UID: \"e562e939-99f0-4bee-952d-7af46132c860\") " pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" Apr 17 20:46:20.623713 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.623550 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e562e939-99f0-4bee-952d-7af46132c860-trusted-ca\") pod \"console-operator-9d4b6777b-9rmnh\" (UID: \"e562e939-99f0-4bee-952d-7af46132c860\") " pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" Apr 17 20:46:20.623713 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.623566 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-77cdc\" (UniqueName: \"kubernetes.io/projected/05c5418a-43f4-476b-8e92-b7b56513d5af-kube-api-access-77cdc\") pod \"kube-storage-version-migrator-operator-6769c5d45-fs4ts\" (UID: \"05c5418a-43f4-476b-8e92-b7b56513d5af\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-fs4ts" Apr 17 20:46:20.623713 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.623583 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e562e939-99f0-4bee-952d-7af46132c860-config\") pod \"console-operator-9d4b6777b-9rmnh\" (UID: \"e562e939-99f0-4bee-952d-7af46132c860\") " pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" Apr 17 20:46:20.623713 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.623608 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1981af6-ac30-4f6a-b971-bb5e2e4f8b1e-serving-cert\") pod \"service-ca-operator-d6fc45fc5-rxj8l\" (UID: \"f1981af6-ac30-4f6a-b971-bb5e2e4f8b1e\") " pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-rxj8l" Apr 17 20:46:20.623713 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.623639 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vvjct\" (UniqueName: \"kubernetes.io/projected/e562e939-99f0-4bee-952d-7af46132c860-kube-api-access-vvjct\") pod \"console-operator-9d4b6777b-9rmnh\" (UID: \"e562e939-99f0-4bee-952d-7af46132c860\") " pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" Apr 17 20:46:20.623713 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.623681 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05c5418a-43f4-476b-8e92-b7b56513d5af-serving-cert\") pod \"kube-storage-version-migrator-operator-6769c5d45-fs4ts\" (UID: \"05c5418a-43f4-476b-8e92-b7b56513d5af\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-fs4ts" Apr 17 20:46:20.623713 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.623709 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1981af6-ac30-4f6a-b971-bb5e2e4f8b1e-config\") pod \"service-ca-operator-d6fc45fc5-rxj8l\" (UID: \"f1981af6-ac30-4f6a-b971-bb5e2e4f8b1e\") " pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-rxj8l" Apr 17 20:46:20.624097 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.623732 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/5506563e-216d-4b61-97dc-8cea5be3f474-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-7cqn6\" (UID: \"5506563e-216d-4b61-97dc-8cea5be3f474\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-7cqn6" Apr 17 20:46:20.624097 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.623773 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8hkk7\" (UniqueName: \"kubernetes.io/projected/f1981af6-ac30-4f6a-b971-bb5e2e4f8b1e-kube-api-access-8hkk7\") pod \"service-ca-operator-d6fc45fc5-rxj8l\" (UID: \"f1981af6-ac30-4f6a-b971-bb5e2e4f8b1e\") " pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-rxj8l" Apr 17 20:46:20.624191 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:20.624097 2577 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Apr 17 20:46:20.624191 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:20.624162 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5506563e-216d-4b61-97dc-8cea5be3f474-cluster-monitoring-operator-tls podName:5506563e-216d-4b61-97dc-8cea5be3f474 nodeName:}" failed. No retries permitted until 2026-04-17 20:46:21.124144483 +0000 UTC m=+106.597922753 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/5506563e-216d-4b61-97dc-8cea5be3f474-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-75587bd455-7cqn6" (UID: "5506563e-216d-4b61-97dc-8cea5be3f474") : secret "cluster-monitoring-operator-tls" not found Apr 17 20:46:20.624191 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.624165 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05c5418a-43f4-476b-8e92-b7b56513d5af-config\") pod \"kube-storage-version-migrator-operator-6769c5d45-fs4ts\" (UID: \"05c5418a-43f4-476b-8e92-b7b56513d5af\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-fs4ts" Apr 17 20:46:20.624346 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.624298 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/5506563e-216d-4b61-97dc-8cea5be3f474-telemetry-config\") pod \"cluster-monitoring-operator-75587bd455-7cqn6\" (UID: \"5506563e-216d-4b61-97dc-8cea5be3f474\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-7cqn6" Apr 17 20:46:20.624440 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.624421 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f1981af6-ac30-4f6a-b971-bb5e2e4f8b1e-config\") pod \"service-ca-operator-d6fc45fc5-rxj8l\" (UID: \"f1981af6-ac30-4f6a-b971-bb5e2e4f8b1e\") " pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-rxj8l" Apr 17 20:46:20.624651 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.624629 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e562e939-99f0-4bee-952d-7af46132c860-config\") pod \"console-operator-9d4b6777b-9rmnh\" (UID: \"e562e939-99f0-4bee-952d-7af46132c860\") " pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" Apr 17 20:46:20.625579 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.625558 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e562e939-99f0-4bee-952d-7af46132c860-trusted-ca\") pod \"console-operator-9d4b6777b-9rmnh\" (UID: \"e562e939-99f0-4bee-952d-7af46132c860\") " pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" Apr 17 20:46:20.626418 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.626396 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e562e939-99f0-4bee-952d-7af46132c860-serving-cert\") pod \"console-operator-9d4b6777b-9rmnh\" (UID: \"e562e939-99f0-4bee-952d-7af46132c860\") " pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" Apr 17 20:46:20.626486 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.626450 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05c5418a-43f4-476b-8e92-b7b56513d5af-serving-cert\") pod \"kube-storage-version-migrator-operator-6769c5d45-fs4ts\" (UID: \"05c5418a-43f4-476b-8e92-b7b56513d5af\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-fs4ts" Apr 17 20:46:20.626562 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.626541 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f1981af6-ac30-4f6a-b971-bb5e2e4f8b1e-serving-cert\") pod \"service-ca-operator-d6fc45fc5-rxj8l\" (UID: \"f1981af6-ac30-4f6a-b971-bb5e2e4f8b1e\") " pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-rxj8l" Apr 17 20:46:20.627434 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.627419 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-dhmj2" Apr 17 20:46:20.631876 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.631851 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr7w2\" (UniqueName: \"kubernetes.io/projected/5506563e-216d-4b61-97dc-8cea5be3f474-kube-api-access-wr7w2\") pod \"cluster-monitoring-operator-75587bd455-7cqn6\" (UID: \"5506563e-216d-4b61-97dc-8cea5be3f474\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-7cqn6" Apr 17 20:46:20.632311 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.632290 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hkk7\" (UniqueName: \"kubernetes.io/projected/f1981af6-ac30-4f6a-b971-bb5e2e4f8b1e-kube-api-access-8hkk7\") pod \"service-ca-operator-d6fc45fc5-rxj8l\" (UID: \"f1981af6-ac30-4f6a-b971-bb5e2e4f8b1e\") " pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-rxj8l" Apr 17 20:46:20.632412 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.632363 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-77cdc\" (UniqueName: \"kubernetes.io/projected/05c5418a-43f4-476b-8e92-b7b56513d5af-kube-api-access-77cdc\") pod \"kube-storage-version-migrator-operator-6769c5d45-fs4ts\" (UID: \"05c5418a-43f4-476b-8e92-b7b56513d5af\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-fs4ts" Apr 17 20:46:20.632477 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.632416 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvjct\" (UniqueName: \"kubernetes.io/projected/e562e939-99f0-4bee-952d-7af46132c860-kube-api-access-vvjct\") pod \"console-operator-9d4b6777b-9rmnh\" (UID: \"e562e939-99f0-4bee-952d-7af46132c860\") " pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" Apr 17 20:46:20.646823 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.642933 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-operator-585dfdc468-5jzh4" Apr 17 20:46:20.717125 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.717099 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-rxj8l" Apr 17 20:46:20.731431 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.731402 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" Apr 17 20:46:20.735120 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.735080 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-fs4ts" Apr 17 20:46:20.765396 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.765271 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-dhmj2"] Apr 17 20:46:20.769873 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:46:20.769841 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2cfc1b29_dfac_4fb3_878c_0961d357f28e.slice/crio-434e7cff1f78e767e00abc10001e77a92cf8103a4c7495104eb9429d8b2b7e04 WatchSource:0}: Error finding container 434e7cff1f78e767e00abc10001e77a92cf8103a4c7495104eb9429d8b2b7e04: Status 404 returned error can't find the container with id 434e7cff1f78e767e00abc10001e77a92cf8103a4c7495104eb9429d8b2b7e04 Apr 17 20:46:20.788225 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.788161 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-operator-585dfdc468-5jzh4"] Apr 17 20:46:20.794264 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:46:20.794212 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18ff00e6_c4f7_4852_b689_94ec992eb400.slice/crio-6d8d9279a71c4a20581aebb6d781801d55e05c64db7e486f0ce6b113727ff3aa WatchSource:0}: Error finding container 6d8d9279a71c4a20581aebb6d781801d55e05c64db7e486f0ce6b113727ff3aa: Status 404 returned error can't find the container with id 6d8d9279a71c4a20581aebb6d781801d55e05c64db7e486f0ce6b113727ff3aa Apr 17 20:46:20.865545 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:20.865473 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-d6fc45fc5-rxj8l"] Apr 17 20:46:20.869298 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:46:20.869258 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1981af6_ac30_4f6a_b971_bb5e2e4f8b1e.slice/crio-7641b53b041e6c51e9827ec097e1d1605a63ca7884013f81314f49ebd0ab8016 WatchSource:0}: Error finding container 7641b53b041e6c51e9827ec097e1d1605a63ca7884013f81314f49ebd0ab8016: Status 404 returned error can't find the container with id 7641b53b041e6c51e9827ec097e1d1605a63ca7884013f81314f49ebd0ab8016 Apr 17 20:46:21.027467 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:21.027442 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5f4b3fb0-6561-4993-a933-390f93a354ea-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-rz8d2\" (UID: \"5f4b3fb0-6561-4993-a933-390f93a354ea\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-rz8d2" Apr 17 20:46:21.027589 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:21.027579 2577 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Apr 17 20:46:21.027647 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:21.027637 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f4b3fb0-6561-4993-a933-390f93a354ea-samples-operator-tls podName:5f4b3fb0-6561-4993-a933-390f93a354ea nodeName:}" failed. No retries permitted until 2026-04-17 20:46:22.027622689 +0000 UTC m=+107.501400954 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/5f4b3fb0-6561-4993-a933-390f93a354ea-samples-operator-tls") pod "cluster-samples-operator-6dc5bdb6b4-rz8d2" (UID: "5f4b3fb0-6561-4993-a933-390f93a354ea") : secret "samples-operator-tls" not found Apr 17 20:46:21.084530 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:21.084498 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-fs4ts"] Apr 17 20:46:21.087320 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:21.087294 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-9d4b6777b-9rmnh"] Apr 17 20:46:21.087653 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:46:21.087627 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05c5418a_43f4_476b_8e92_b7b56513d5af.slice/crio-8d193b44b639906a9f6ca3065ae9b44d2f1e7689a781c58909b3f7f514b7fdb3 WatchSource:0}: Error finding container 8d193b44b639906a9f6ca3065ae9b44d2f1e7689a781c58909b3f7f514b7fdb3: Status 404 returned error can't find the container with id 8d193b44b639906a9f6ca3065ae9b44d2f1e7689a781c58909b3f7f514b7fdb3 Apr 17 20:46:21.091617 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:46:21.091594 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode562e939_99f0_4bee_952d_7af46132c860.slice/crio-e7badc18787522fc79f365dfe89907915cd8ef44215ee701e43a3deba71eb070 WatchSource:0}: Error finding container e7badc18787522fc79f365dfe89907915cd8ef44215ee701e43a3deba71eb070: Status 404 returned error can't find the container with id e7badc18787522fc79f365dfe89907915cd8ef44215ee701e43a3deba71eb070 Apr 17 20:46:21.128049 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:21.128030 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/5506563e-216d-4b61-97dc-8cea5be3f474-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-7cqn6\" (UID: \"5506563e-216d-4b61-97dc-8cea5be3f474\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-7cqn6" Apr 17 20:46:21.128180 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:21.128162 2577 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Apr 17 20:46:21.128229 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:21.128211 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5506563e-216d-4b61-97dc-8cea5be3f474-cluster-monitoring-operator-tls podName:5506563e-216d-4b61-97dc-8cea5be3f474 nodeName:}" failed. No retries permitted until 2026-04-17 20:46:22.128196493 +0000 UTC m=+107.601974757 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/5506563e-216d-4b61-97dc-8cea5be3f474-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-75587bd455-7cqn6" (UID: "5506563e-216d-4b61-97dc-8cea5be3f474") : secret "cluster-monitoring-operator-tls" not found Apr 17 20:46:21.401587 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:21.401506 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-fs4ts" event={"ID":"05c5418a-43f4-476b-8e92-b7b56513d5af","Type":"ContainerStarted","Data":"8d193b44b639906a9f6ca3065ae9b44d2f1e7689a781c58909b3f7f514b7fdb3"} Apr 17 20:46:21.403727 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:21.403699 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-operator-585dfdc468-5jzh4" event={"ID":"18ff00e6-c4f7-4852-b689-94ec992eb400","Type":"ContainerStarted","Data":"6d8d9279a71c4a20581aebb6d781801d55e05c64db7e486f0ce6b113727ff3aa"} Apr 17 20:46:21.404865 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:21.404840 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" event={"ID":"e562e939-99f0-4bee-952d-7af46132c860","Type":"ContainerStarted","Data":"e7badc18787522fc79f365dfe89907915cd8ef44215ee701e43a3deba71eb070"} Apr 17 20:46:21.406205 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:21.406182 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-rxj8l" event={"ID":"f1981af6-ac30-4f6a-b971-bb5e2e4f8b1e","Type":"ContainerStarted","Data":"7641b53b041e6c51e9827ec097e1d1605a63ca7884013f81314f49ebd0ab8016"} Apr 17 20:46:21.407278 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:21.407243 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-dhmj2" event={"ID":"2cfc1b29-dfac-4fb3-878c-0961d357f28e","Type":"ContainerStarted","Data":"434e7cff1f78e767e00abc10001e77a92cf8103a4c7495104eb9429d8b2b7e04"} Apr 17 20:46:22.038200 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:22.037589 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5f4b3fb0-6561-4993-a933-390f93a354ea-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-rz8d2\" (UID: \"5f4b3fb0-6561-4993-a933-390f93a354ea\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-rz8d2" Apr 17 20:46:22.038200 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:22.037722 2577 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Apr 17 20:46:22.038200 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:22.037785 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f4b3fb0-6561-4993-a933-390f93a354ea-samples-operator-tls podName:5f4b3fb0-6561-4993-a933-390f93a354ea nodeName:}" failed. No retries permitted until 2026-04-17 20:46:24.037765406 +0000 UTC m=+109.511543671 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/5f4b3fb0-6561-4993-a933-390f93a354ea-samples-operator-tls") pod "cluster-samples-operator-6dc5bdb6b4-rz8d2" (UID: "5f4b3fb0-6561-4993-a933-390f93a354ea") : secret "samples-operator-tls" not found Apr 17 20:46:22.138691 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:22.138522 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/5506563e-216d-4b61-97dc-8cea5be3f474-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-7cqn6\" (UID: \"5506563e-216d-4b61-97dc-8cea5be3f474\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-7cqn6" Apr 17 20:46:22.138938 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:22.138672 2577 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Apr 17 20:46:22.138938 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:22.138794 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5506563e-216d-4b61-97dc-8cea5be3f474-cluster-monitoring-operator-tls podName:5506563e-216d-4b61-97dc-8cea5be3f474 nodeName:}" failed. No retries permitted until 2026-04-17 20:46:24.138771962 +0000 UTC m=+109.612550229 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/5506563e-216d-4b61-97dc-8cea5be3f474-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-75587bd455-7cqn6" (UID: "5506563e-216d-4b61-97dc-8cea5be3f474") : secret "cluster-monitoring-operator-tls" not found Apr 17 20:46:23.414252 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:23.414204 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-dhmj2" event={"ID":"2cfc1b29-dfac-4fb3-878c-0961d357f28e","Type":"ContainerStarted","Data":"e67856ef0b84f97a1b1320254c9de5d88c2a560040bdb2018fd31b512c1e01de"} Apr 17 20:46:23.429183 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:23.429133 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-storage-operator/volume-data-source-validator-7c6cbb6c87-dhmj2" podStartSLOduration=1.828532213 podStartE2EDuration="3.429119081s" podCreationTimestamp="2026-04-17 20:46:20 +0000 UTC" firstStartedPulling="2026-04-17 20:46:20.772099229 +0000 UTC m=+106.245877497" lastFinishedPulling="2026-04-17 20:46:22.372686087 +0000 UTC m=+107.846464365" observedRunningTime="2026-04-17 20:46:23.428087843 +0000 UTC m=+108.901866141" watchObservedRunningTime="2026-04-17 20:46:23.429119081 +0000 UTC m=+108.902897367" Apr 17 20:46:23.676593 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:23.676502 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-8894fc9bd-z74x2"] Apr 17 20:46:23.680973 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:23.680947 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-8894fc9bd-z74x2" Apr 17 20:46:23.682865 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:23.682664 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-diagnostics\"/\"network-diagnostics-dockercfg-cj84s\"" Apr 17 20:46:23.687674 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:23.687653 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-source-8894fc9bd-z74x2"] Apr 17 20:46:23.755380 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:23.755354 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbv2x\" (UniqueName: \"kubernetes.io/projected/16649afe-84e4-4bed-b4c3-a24055c6af4d-kube-api-access-jbv2x\") pod \"network-check-source-8894fc9bd-z74x2\" (UID: \"16649afe-84e4-4bed-b4c3-a24055c6af4d\") " pod="openshift-network-diagnostics/network-check-source-8894fc9bd-z74x2" Apr 17 20:46:23.855843 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:23.855791 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-jbv2x\" (UniqueName: \"kubernetes.io/projected/16649afe-84e4-4bed-b4c3-a24055c6af4d-kube-api-access-jbv2x\") pod \"network-check-source-8894fc9bd-z74x2\" (UID: \"16649afe-84e4-4bed-b4c3-a24055c6af4d\") " pod="openshift-network-diagnostics/network-check-source-8894fc9bd-z74x2" Apr 17 20:46:23.863452 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:23.863419 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbv2x\" (UniqueName: \"kubernetes.io/projected/16649afe-84e4-4bed-b4c3-a24055c6af4d-kube-api-access-jbv2x\") pod \"network-check-source-8894fc9bd-z74x2\" (UID: \"16649afe-84e4-4bed-b4c3-a24055c6af4d\") " pod="openshift-network-diagnostics/network-check-source-8894fc9bd-z74x2" Apr 17 20:46:23.992636 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:23.992602 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-8894fc9bd-z74x2" Apr 17 20:46:24.056890 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:24.056861 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5f4b3fb0-6561-4993-a933-390f93a354ea-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-rz8d2\" (UID: \"5f4b3fb0-6561-4993-a933-390f93a354ea\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-rz8d2" Apr 17 20:46:24.057015 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:24.056971 2577 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Apr 17 20:46:24.057059 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:24.057025 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f4b3fb0-6561-4993-a933-390f93a354ea-samples-operator-tls podName:5f4b3fb0-6561-4993-a933-390f93a354ea nodeName:}" failed. No retries permitted until 2026-04-17 20:46:28.057010624 +0000 UTC m=+113.530788888 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/5f4b3fb0-6561-4993-a933-390f93a354ea-samples-operator-tls") pod "cluster-samples-operator-6dc5bdb6b4-rz8d2" (UID: "5f4b3fb0-6561-4993-a933-390f93a354ea") : secret "samples-operator-tls" not found Apr 17 20:46:24.157362 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:24.157326 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/5506563e-216d-4b61-97dc-8cea5be3f474-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-7cqn6\" (UID: \"5506563e-216d-4b61-97dc-8cea5be3f474\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-7cqn6" Apr 17 20:46:24.157531 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:24.157473 2577 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Apr 17 20:46:24.157579 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:24.157549 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5506563e-216d-4b61-97dc-8cea5be3f474-cluster-monitoring-operator-tls podName:5506563e-216d-4b61-97dc-8cea5be3f474 nodeName:}" failed. No retries permitted until 2026-04-17 20:46:28.157529668 +0000 UTC m=+113.631307949 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/5506563e-216d-4b61-97dc-8cea5be3f474-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-75587bd455-7cqn6" (UID: "5506563e-216d-4b61-97dc-8cea5be3f474") : secret "cluster-monitoring-operator-tls" not found Apr 17 20:46:24.709642 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:24.709541 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-source-8894fc9bd-z74x2"] Apr 17 20:46:24.712485 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:46:24.712453 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16649afe_84e4_4bed_b4c3_a24055c6af4d.slice/crio-4a074d3119e31444c3836d9f7b737d1978b50f45f4dc481333d93852b79b383d WatchSource:0}: Error finding container 4a074d3119e31444c3836d9f7b737d1978b50f45f4dc481333d93852b79b383d: Status 404 returned error can't find the container with id 4a074d3119e31444c3836d9f7b737d1978b50f45f4dc481333d93852b79b383d Apr 17 20:46:25.419478 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:25.419436 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-8894fc9bd-z74x2" event={"ID":"16649afe-84e4-4bed-b4c3-a24055c6af4d","Type":"ContainerStarted","Data":"35ed28652b862e07b1bfc9f525895f3997867de9e58b8aa5d3c33854e911d1dd"} Apr 17 20:46:25.419478 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:25.419478 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-8894fc9bd-z74x2" event={"ID":"16649afe-84e4-4bed-b4c3-a24055c6af4d","Type":"ContainerStarted","Data":"4a074d3119e31444c3836d9f7b737d1978b50f45f4dc481333d93852b79b383d"} Apr 17 20:46:25.421541 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:25.421517 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-9rmnh_e562e939-99f0-4bee-952d-7af46132c860/console-operator/0.log" Apr 17 20:46:25.421661 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:25.421562 2577 generic.go:358] "Generic (PLEG): container finished" podID="e562e939-99f0-4bee-952d-7af46132c860" containerID="d25d96b9ec5844494361be48f4d361d7ae4b66c611b7f4773ec95157962e6ad0" exitCode=255 Apr 17 20:46:25.421661 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:25.421628 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" event={"ID":"e562e939-99f0-4bee-952d-7af46132c860","Type":"ContainerDied","Data":"d25d96b9ec5844494361be48f4d361d7ae4b66c611b7f4773ec95157962e6ad0"} Apr 17 20:46:25.421955 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:25.421932 2577 scope.go:117] "RemoveContainer" containerID="d25d96b9ec5844494361be48f4d361d7ae4b66c611b7f4773ec95157962e6ad0" Apr 17 20:46:25.425292 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:25.425259 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-rxj8l" event={"ID":"f1981af6-ac30-4f6a-b971-bb5e2e4f8b1e","Type":"ContainerStarted","Data":"dbfffb563d8a142bdf535ab5b834de395ded0178d8c9f349295907b6d10ba249"} Apr 17 20:46:25.426945 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:25.426919 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-fs4ts" event={"ID":"05c5418a-43f4-476b-8e92-b7b56513d5af","Type":"ContainerStarted","Data":"4ff09f85efdf190efc51e4034ea41febf5532eaa4dd2f653ea6b0df5d2253af7"} Apr 17 20:46:25.428434 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:25.428411 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-operator-585dfdc468-5jzh4" event={"ID":"18ff00e6-c4f7-4852-b689-94ec992eb400","Type":"ContainerStarted","Data":"68bc86d5b22339444a41cf053918b6e5a22adaeafd4d59636a98e77c5f5b3ff5"} Apr 17 20:46:25.432692 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:25.432650 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-diagnostics/network-check-source-8894fc9bd-z74x2" podStartSLOduration=2.432638028 podStartE2EDuration="2.432638028s" podCreationTimestamp="2026-04-17 20:46:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 20:46:25.432249726 +0000 UTC m=+110.906028016" watchObservedRunningTime="2026-04-17 20:46:25.432638028 +0000 UTC m=+110.906416310" Apr 17 20:46:25.446054 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:25.446010 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-rxj8l" podStartSLOduration=1.75193596 podStartE2EDuration="5.445995616s" podCreationTimestamp="2026-04-17 20:46:20 +0000 UTC" firstStartedPulling="2026-04-17 20:46:20.871219194 +0000 UTC m=+106.344997459" lastFinishedPulling="2026-04-17 20:46:24.565278835 +0000 UTC m=+110.039057115" observedRunningTime="2026-04-17 20:46:25.445044757 +0000 UTC m=+110.918823044" watchObservedRunningTime="2026-04-17 20:46:25.445995616 +0000 UTC m=+110.919773908" Apr 17 20:46:25.459126 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:25.459083 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-insights/insights-operator-585dfdc468-5jzh4" podStartSLOduration=1.6895413700000002 podStartE2EDuration="5.459069061s" podCreationTimestamp="2026-04-17 20:46:20 +0000 UTC" firstStartedPulling="2026-04-17 20:46:20.796114284 +0000 UTC m=+106.269892557" lastFinishedPulling="2026-04-17 20:46:24.565641967 +0000 UTC m=+110.039420248" observedRunningTime="2026-04-17 20:46:25.458438897 +0000 UTC m=+110.932217185" watchObservedRunningTime="2026-04-17 20:46:25.459069061 +0000 UTC m=+110.932847349" Apr 17 20:46:25.472357 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:25.472314 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-fs4ts" podStartSLOduration=1.99105147 podStartE2EDuration="5.472296427s" podCreationTimestamp="2026-04-17 20:46:20 +0000 UTC" firstStartedPulling="2026-04-17 20:46:21.089725167 +0000 UTC m=+106.563503432" lastFinishedPulling="2026-04-17 20:46:24.570970117 +0000 UTC m=+110.044748389" observedRunningTime="2026-04-17 20:46:25.47164166 +0000 UTC m=+110.945419948" watchObservedRunningTime="2026-04-17 20:46:25.472296427 +0000 UTC m=+110.946074717" Apr 17 20:46:25.946534 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:25.946504 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-74bb7799d9-wjmrh"] Apr 17 20:46:25.949425 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:25.949409 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-wjmrh" Apr 17 20:46:25.951228 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:25.951202 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator\"/\"kube-root-ca.crt\"" Apr 17 20:46:25.951369 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:25.951203 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator\"/\"openshift-service-ca.crt\"" Apr 17 20:46:25.951754 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:25.951739 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator\"/\"kube-storage-version-migrator-sa-dockercfg-55kr2\"" Apr 17 20:46:25.957316 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:25.957297 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-74bb7799d9-wjmrh"] Apr 17 20:46:26.072098 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:26.072066 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrbqq\" (UniqueName: \"kubernetes.io/projected/944799f6-55a3-4bc4-a6e0-3a72a00ae117-kube-api-access-wrbqq\") pod \"migrator-74bb7799d9-wjmrh\" (UID: \"944799f6-55a3-4bc4-a6e0-3a72a00ae117\") " pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-wjmrh" Apr 17 20:46:26.172619 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:26.172582 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-wrbqq\" (UniqueName: \"kubernetes.io/projected/944799f6-55a3-4bc4-a6e0-3a72a00ae117-kube-api-access-wrbqq\") pod \"migrator-74bb7799d9-wjmrh\" (UID: \"944799f6-55a3-4bc4-a6e0-3a72a00ae117\") " pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-wjmrh" Apr 17 20:46:26.179454 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:26.179433 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrbqq\" (UniqueName: \"kubernetes.io/projected/944799f6-55a3-4bc4-a6e0-3a72a00ae117-kube-api-access-wrbqq\") pod \"migrator-74bb7799d9-wjmrh\" (UID: \"944799f6-55a3-4bc4-a6e0-3a72a00ae117\") " pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-wjmrh" Apr 17 20:46:26.259533 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:26.259471 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-wjmrh" Apr 17 20:46:26.373517 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:26.373488 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-74bb7799d9-wjmrh"] Apr 17 20:46:26.377384 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:46:26.377359 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod944799f6_55a3_4bc4_a6e0_3a72a00ae117.slice/crio-4714bc9af4c2d0f0a13582a3371383337a4887f1021b1a58e7f90a03d8ba9fde WatchSource:0}: Error finding container 4714bc9af4c2d0f0a13582a3371383337a4887f1021b1a58e7f90a03d8ba9fde: Status 404 returned error can't find the container with id 4714bc9af4c2d0f0a13582a3371383337a4887f1021b1a58e7f90a03d8ba9fde Apr 17 20:46:26.432402 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:26.432381 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-9rmnh_e562e939-99f0-4bee-952d-7af46132c860/console-operator/1.log" Apr 17 20:46:26.432739 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:26.432725 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-9rmnh_e562e939-99f0-4bee-952d-7af46132c860/console-operator/0.log" Apr 17 20:46:26.432821 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:26.432758 2577 generic.go:358] "Generic (PLEG): container finished" podID="e562e939-99f0-4bee-952d-7af46132c860" containerID="95b531c2ef89963de74a0b8dfe8273f56bbb121833267ad8f9c22be8938385a5" exitCode=255 Apr 17 20:46:26.432874 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:26.432859 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" event={"ID":"e562e939-99f0-4bee-952d-7af46132c860","Type":"ContainerDied","Data":"95b531c2ef89963de74a0b8dfe8273f56bbb121833267ad8f9c22be8938385a5"} Apr 17 20:46:26.432920 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:26.432901 2577 scope.go:117] "RemoveContainer" containerID="d25d96b9ec5844494361be48f4d361d7ae4b66c611b7f4773ec95157962e6ad0" Apr 17 20:46:26.433160 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:26.433143 2577 scope.go:117] "RemoveContainer" containerID="95b531c2ef89963de74a0b8dfe8273f56bbb121833267ad8f9c22be8938385a5" Apr 17 20:46:26.433366 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:26.433345 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"console-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=console-operator pod=console-operator-9d4b6777b-9rmnh_openshift-console-operator(e562e939-99f0-4bee-952d-7af46132c860)\"" pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" podUID="e562e939-99f0-4bee-952d-7af46132c860" Apr 17 20:46:26.434164 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:26.433999 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-wjmrh" event={"ID":"944799f6-55a3-4bc4-a6e0-3a72a00ae117","Type":"ContainerStarted","Data":"4714bc9af4c2d0f0a13582a3371383337a4887f1021b1a58e7f90a03d8ba9fde"} Apr 17 20:46:27.437163 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:27.437137 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-9rmnh_e562e939-99f0-4bee-952d-7af46132c860/console-operator/1.log" Apr 17 20:46:27.437512 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:27.437481 2577 scope.go:117] "RemoveContainer" containerID="95b531c2ef89963de74a0b8dfe8273f56bbb121833267ad8f9c22be8938385a5" Apr 17 20:46:27.437700 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:27.437682 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"console-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=console-operator pod=console-operator-9d4b6777b-9rmnh_openshift-console-operator(e562e939-99f0-4bee-952d-7af46132c860)\"" pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" podUID="e562e939-99f0-4bee-952d-7af46132c860" Apr 17 20:46:27.657982 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:27.657963 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_node-resolver-cwp9d_c9f9af16-3b9f-4ebb-8025-9b9162c7eaa1/dns-node-resolver/0.log" Apr 17 20:46:28.087046 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:28.087016 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5f4b3fb0-6561-4993-a933-390f93a354ea-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-rz8d2\" (UID: \"5f4b3fb0-6561-4993-a933-390f93a354ea\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-rz8d2" Apr 17 20:46:28.087194 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:28.087126 2577 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Apr 17 20:46:28.087194 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:28.087182 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5f4b3fb0-6561-4993-a933-390f93a354ea-samples-operator-tls podName:5f4b3fb0-6561-4993-a933-390f93a354ea nodeName:}" failed. No retries permitted until 2026-04-17 20:46:36.08716409 +0000 UTC m=+121.560942371 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/5f4b3fb0-6561-4993-a933-390f93a354ea-samples-operator-tls") pod "cluster-samples-operator-6dc5bdb6b4-rz8d2" (UID: "5f4b3fb0-6561-4993-a933-390f93a354ea") : secret "samples-operator-tls" not found Apr 17 20:46:28.187317 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:28.187290 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/5506563e-216d-4b61-97dc-8cea5be3f474-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-7cqn6\" (UID: \"5506563e-216d-4b61-97dc-8cea5be3f474\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-7cqn6" Apr 17 20:46:28.187475 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:28.187453 2577 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Apr 17 20:46:28.187552 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:28.187532 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5506563e-216d-4b61-97dc-8cea5be3f474-cluster-monitoring-operator-tls podName:5506563e-216d-4b61-97dc-8cea5be3f474 nodeName:}" failed. No retries permitted until 2026-04-17 20:46:36.187512905 +0000 UTC m=+121.661291172 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/5506563e-216d-4b61-97dc-8cea5be3f474-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-75587bd455-7cqn6" (UID: "5506563e-216d-4b61-97dc-8cea5be3f474") : secret "cluster-monitoring-operator-tls" not found Apr 17 20:46:28.441139 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:28.441111 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-wjmrh" event={"ID":"944799f6-55a3-4bc4-a6e0-3a72a00ae117","Type":"ContainerStarted","Data":"d3f3e7ec411398740903b0528e135d7bd3a6fb595926c8c165a0cda6f7115645"} Apr 17 20:46:28.441455 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:28.441146 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-wjmrh" event={"ID":"944799f6-55a3-4bc4-a6e0-3a72a00ae117","Type":"ContainerStarted","Data":"701c0d395bb616740c79fe6273989b04c0004ef11741f2047063e584be2f321d"} Apr 17 20:46:28.454281 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:28.454230 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-74bb7799d9-wjmrh" podStartSLOduration=2.377415076 podStartE2EDuration="3.454216334s" podCreationTimestamp="2026-04-17 20:46:25 +0000 UTC" firstStartedPulling="2026-04-17 20:46:26.379275991 +0000 UTC m=+111.853054256" lastFinishedPulling="2026-04-17 20:46:27.456077246 +0000 UTC m=+112.929855514" observedRunningTime="2026-04-17 20:46:28.453504618 +0000 UTC m=+113.927282905" watchObservedRunningTime="2026-04-17 20:46:28.454216334 +0000 UTC m=+113.927994622" Apr 17 20:46:28.657673 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:28.657649 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_node-ca-v8fbw_5fa81115-694f-4f2f-8797-8962d102b7e7/node-ca/0.log" Apr 17 20:46:30.732370 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:30.732335 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" Apr 17 20:46:30.732370 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:30.732366 2577 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" Apr 17 20:46:30.732774 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:30.732713 2577 scope.go:117] "RemoveContainer" containerID="95b531c2ef89963de74a0b8dfe8273f56bbb121833267ad8f9c22be8938385a5" Apr 17 20:46:30.732924 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:30.732905 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"console-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=console-operator pod=console-operator-9d4b6777b-9rmnh_openshift-console-operator(e562e939-99f0-4bee-952d-7af46132c860)\"" pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" podUID="e562e939-99f0-4bee-952d-7af46132c860" Apr 17 20:46:36.141832 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:36.141770 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5f4b3fb0-6561-4993-a933-390f93a354ea-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-rz8d2\" (UID: \"5f4b3fb0-6561-4993-a933-390f93a354ea\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-rz8d2" Apr 17 20:46:36.144498 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:36.144466 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/5f4b3fb0-6561-4993-a933-390f93a354ea-samples-operator-tls\") pod \"cluster-samples-operator-6dc5bdb6b4-rz8d2\" (UID: \"5f4b3fb0-6561-4993-a933-390f93a354ea\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-rz8d2" Apr 17 20:46:36.234786 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:36.234763 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"cluster-samples-operator-dockercfg-g4s6r\"" Apr 17 20:46:36.242369 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:36.242347 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/5506563e-216d-4b61-97dc-8cea5be3f474-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-7cqn6\" (UID: \"5506563e-216d-4b61-97dc-8cea5be3f474\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-7cqn6" Apr 17 20:46:36.242523 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:36.242506 2577 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Apr 17 20:46:36.242584 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:36.242574 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5506563e-216d-4b61-97dc-8cea5be3f474-cluster-monitoring-operator-tls podName:5506563e-216d-4b61-97dc-8cea5be3f474 nodeName:}" failed. No retries permitted until 2026-04-17 20:46:52.242558485 +0000 UTC m=+137.716336749 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/5506563e-216d-4b61-97dc-8cea5be3f474-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-75587bd455-7cqn6" (UID: "5506563e-216d-4b61-97dc-8cea5be3f474") : secret "cluster-monitoring-operator-tls" not found Apr 17 20:46:36.243327 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:36.243302 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-rz8d2" Apr 17 20:46:36.355740 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:36.355645 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-rz8d2"] Apr 17 20:46:36.466739 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:36.466710 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-rz8d2" event={"ID":"5f4b3fb0-6561-4993-a933-390f93a354ea","Type":"ContainerStarted","Data":"ce4920bda58ae3193702b81bfea837d1d1eb8bf8c6b7cca870d4b314cdde5495"} Apr 17 20:46:38.473065 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:38.473025 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-rz8d2" event={"ID":"5f4b3fb0-6561-4993-a933-390f93a354ea","Type":"ContainerStarted","Data":"117c6cdbd1b70d4f83411458e4c3c60d2afce4d91483125bf964cdd0fca833bf"} Apr 17 20:46:38.473065 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:38.473066 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-rz8d2" event={"ID":"5f4b3fb0-6561-4993-a933-390f93a354ea","Type":"ContainerStarted","Data":"ccce2d834599c73f3ba06524f61777a41b61f8c6632ba21a9ed3adaeb52bc928"} Apr 17 20:46:38.487818 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:38.487753 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-6dc5bdb6b4-rz8d2" podStartSLOduration=16.610257606 podStartE2EDuration="18.487740867s" podCreationTimestamp="2026-04-17 20:46:20 +0000 UTC" firstStartedPulling="2026-04-17 20:46:36.399785146 +0000 UTC m=+121.873563411" lastFinishedPulling="2026-04-17 20:46:38.277268393 +0000 UTC m=+123.751046672" observedRunningTime="2026-04-17 20:46:38.486566135 +0000 UTC m=+123.960344435" watchObservedRunningTime="2026-04-17 20:46:38.487740867 +0000 UTC m=+123.961519150" Apr 17 20:46:44.810468 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:44.810426 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1a6721ad-2d96-4d01-a9db-b8654256e62f-metrics-certs\") pod \"network-metrics-daemon-7ddt5\" (UID: \"1a6721ad-2d96-4d01-a9db-b8654256e62f\") " pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:46:44.813132 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:44.813111 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1a6721ad-2d96-4d01-a9db-b8654256e62f-metrics-certs\") pod \"network-metrics-daemon-7ddt5\" (UID: \"1a6721ad-2d96-4d01-a9db-b8654256e62f\") " pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:46:44.837862 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:44.837834 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-jzn7k\"" Apr 17 20:46:44.846227 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:44.846207 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-7ddt5" Apr 17 20:46:44.962055 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:44.959513 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-7ddt5"] Apr 17 20:46:45.117727 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:45.117666 2577 scope.go:117] "RemoveContainer" containerID="95b531c2ef89963de74a0b8dfe8273f56bbb121833267ad8f9c22be8938385a5" Apr 17 20:46:45.490228 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:45.490191 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7ddt5" event={"ID":"1a6721ad-2d96-4d01-a9db-b8654256e62f","Type":"ContainerStarted","Data":"e86282d1e842a1c9e6ee34782f1f249bdc924f1326de31100ea53a950e35ac7f"} Apr 17 20:46:45.491584 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:45.491561 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-9rmnh_e562e939-99f0-4bee-952d-7af46132c860/console-operator/2.log" Apr 17 20:46:45.491966 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:45.491952 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-9rmnh_e562e939-99f0-4bee-952d-7af46132c860/console-operator/1.log" Apr 17 20:46:45.492039 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:45.491983 2577 generic.go:358] "Generic (PLEG): container finished" podID="e562e939-99f0-4bee-952d-7af46132c860" containerID="593b07a09262e1a26741d3fa8f78da22e06081a698ef09a33fef086a57e4c88c" exitCode=255 Apr 17 20:46:45.492039 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:45.492012 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" event={"ID":"e562e939-99f0-4bee-952d-7af46132c860","Type":"ContainerDied","Data":"593b07a09262e1a26741d3fa8f78da22e06081a698ef09a33fef086a57e4c88c"} Apr 17 20:46:45.492039 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:45.492036 2577 scope.go:117] "RemoveContainer" containerID="95b531c2ef89963de74a0b8dfe8273f56bbb121833267ad8f9c22be8938385a5" Apr 17 20:46:45.492393 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:45.492369 2577 scope.go:117] "RemoveContainer" containerID="593b07a09262e1a26741d3fa8f78da22e06081a698ef09a33fef086a57e4c88c" Apr 17 20:46:45.492582 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:45.492562 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"console-operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=console-operator pod=console-operator-9d4b6777b-9rmnh_openshift-console-operator(e562e939-99f0-4bee-952d-7af46132c860)\"" pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" podUID="e562e939-99f0-4bee-952d-7af46132c860" Apr 17 20:46:46.495624 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:46.495600 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-9rmnh_e562e939-99f0-4bee-952d-7af46132c860/console-operator/2.log" Apr 17 20:46:46.497136 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:46.497110 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7ddt5" event={"ID":"1a6721ad-2d96-4d01-a9db-b8654256e62f","Type":"ContainerStarted","Data":"7c2d05b625278498d72be9da471ed9aff8a1d7eafe0307bc656d58719445702f"} Apr 17 20:46:46.497238 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:46.497142 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-7ddt5" event={"ID":"1a6721ad-2d96-4d01-a9db-b8654256e62f","Type":"ContainerStarted","Data":"bb6d85930057238f7a4d54628a4f1f8b8c7f18db72d1225437ce1010061dafe9"} Apr 17 20:46:46.510501 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:46.510453 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-7ddt5" podStartSLOduration=130.419098279 podStartE2EDuration="2m11.510437417s" podCreationTimestamp="2026-04-17 20:44:35 +0000 UTC" firstStartedPulling="2026-04-17 20:46:44.963493921 +0000 UTC m=+130.437272186" lastFinishedPulling="2026-04-17 20:46:46.054833046 +0000 UTC m=+131.528611324" observedRunningTime="2026-04-17 20:46:46.510163915 +0000 UTC m=+131.983942205" watchObservedRunningTime="2026-04-17 20:46:46.510437417 +0000 UTC m=+131.984215705" Apr 17 20:46:48.686877 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.686842 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-insights/insights-runtime-extractor-rnnj4"] Apr 17 20:46:48.690533 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.690510 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-runtime-extractor-rnnj4" Apr 17 20:46:48.692789 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.692770 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"insights-runtime-extractor-sa-dockercfg-jrrs2\"" Apr 17 20:46:48.693160 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.693126 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"kube-rbac-proxy\"" Apr 17 20:46:48.693561 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.693543 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"insights-runtime-extractor-tls\"" Apr 17 20:46:48.697715 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.697694 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-runtime-extractor-rnnj4"] Apr 17 20:46:48.757310 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.757280 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-57fdff8cb7-8qxnn"] Apr 17 20:46:48.760283 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.760268 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:46:48.762350 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.762318 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-private-configuration\"" Apr 17 20:46:48.762450 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.762373 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"registry-dockercfg-6qbmr\"" Apr 17 20:46:48.762506 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.762445 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"installation-pull-secrets\"" Apr 17 20:46:48.762548 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.762522 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-tls\"" Apr 17 20:46:48.769485 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.769467 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"trusted-ca\"" Apr 17 20:46:48.774456 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.774435 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-57fdff8cb7-8qxnn"] Apr 17 20:46:48.840165 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.840142 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/4e8e0c9a-eac8-4af9-bd4c-927f0b406f02-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-rnnj4\" (UID: \"4e8e0c9a-eac8-4af9-bd4c-927f0b406f02\") " pod="openshift-insights/insights-runtime-extractor-rnnj4" Apr 17 20:46:48.840251 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.840174 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9wn2\" (UniqueName: \"kubernetes.io/projected/d0f3d88e-5398-4272-bf50-7675902545aa-kube-api-access-h9wn2\") pod \"image-registry-57fdff8cb7-8qxnn\" (UID: \"d0f3d88e-5398-4272-bf50-7675902545aa\") " pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:46:48.840251 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.840204 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d0f3d88e-5398-4272-bf50-7675902545aa-bound-sa-token\") pod \"image-registry-57fdff8cb7-8qxnn\" (UID: \"d0f3d88e-5398-4272-bf50-7675902545aa\") " pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:46:48.840251 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.840246 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d0f3d88e-5398-4272-bf50-7675902545aa-installation-pull-secrets\") pod \"image-registry-57fdff8cb7-8qxnn\" (UID: \"d0f3d88e-5398-4272-bf50-7675902545aa\") " pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:46:48.840346 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.840274 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d0f3d88e-5398-4272-bf50-7675902545aa-registry-tls\") pod \"image-registry-57fdff8cb7-8qxnn\" (UID: \"d0f3d88e-5398-4272-bf50-7675902545aa\") " pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:46:48.840346 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.840317 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/4e8e0c9a-eac8-4af9-bd4c-927f0b406f02-data-volume\") pod \"insights-runtime-extractor-rnnj4\" (UID: \"4e8e0c9a-eac8-4af9-bd4c-927f0b406f02\") " pod="openshift-insights/insights-runtime-extractor-rnnj4" Apr 17 20:46:48.840346 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.840332 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/4e8e0c9a-eac8-4af9-bd4c-927f0b406f02-crio-socket\") pod \"insights-runtime-extractor-rnnj4\" (UID: \"4e8e0c9a-eac8-4af9-bd4c-927f0b406f02\") " pod="openshift-insights/insights-runtime-extractor-rnnj4" Apr 17 20:46:48.840429 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.840347 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmtkn\" (UniqueName: \"kubernetes.io/projected/4e8e0c9a-eac8-4af9-bd4c-927f0b406f02-kube-api-access-vmtkn\") pod \"insights-runtime-extractor-rnnj4\" (UID: \"4e8e0c9a-eac8-4af9-bd4c-927f0b406f02\") " pod="openshift-insights/insights-runtime-extractor-rnnj4" Apr 17 20:46:48.840429 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.840375 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d0f3d88e-5398-4272-bf50-7675902545aa-registry-certificates\") pod \"image-registry-57fdff8cb7-8qxnn\" (UID: \"d0f3d88e-5398-4272-bf50-7675902545aa\") " pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:46:48.840429 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.840399 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d0f3d88e-5398-4272-bf50-7675902545aa-trusted-ca\") pod \"image-registry-57fdff8cb7-8qxnn\" (UID: \"d0f3d88e-5398-4272-bf50-7675902545aa\") " pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:46:48.840429 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.840418 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/d0f3d88e-5398-4272-bf50-7675902545aa-image-registry-private-configuration\") pod \"image-registry-57fdff8cb7-8qxnn\" (UID: \"d0f3d88e-5398-4272-bf50-7675902545aa\") " pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:46:48.840553 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.840436 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d0f3d88e-5398-4272-bf50-7675902545aa-ca-trust-extracted\") pod \"image-registry-57fdff8cb7-8qxnn\" (UID: \"d0f3d88e-5398-4272-bf50-7675902545aa\") " pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:46:48.840553 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.840467 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/4e8e0c9a-eac8-4af9-bd4c-927f0b406f02-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-rnnj4\" (UID: \"4e8e0c9a-eac8-4af9-bd4c-927f0b406f02\") " pod="openshift-insights/insights-runtime-extractor-rnnj4" Apr 17 20:46:48.941537 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.941517 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-h9wn2\" (UniqueName: \"kubernetes.io/projected/d0f3d88e-5398-4272-bf50-7675902545aa-kube-api-access-h9wn2\") pod \"image-registry-57fdff8cb7-8qxnn\" (UID: \"d0f3d88e-5398-4272-bf50-7675902545aa\") " pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:46:48.941623 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.941548 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d0f3d88e-5398-4272-bf50-7675902545aa-bound-sa-token\") pod \"image-registry-57fdff8cb7-8qxnn\" (UID: \"d0f3d88e-5398-4272-bf50-7675902545aa\") " pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:46:48.941623 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.941567 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d0f3d88e-5398-4272-bf50-7675902545aa-installation-pull-secrets\") pod \"image-registry-57fdff8cb7-8qxnn\" (UID: \"d0f3d88e-5398-4272-bf50-7675902545aa\") " pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:46:48.941887 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.941868 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d0f3d88e-5398-4272-bf50-7675902545aa-registry-tls\") pod \"image-registry-57fdff8cb7-8qxnn\" (UID: \"d0f3d88e-5398-4272-bf50-7675902545aa\") " pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:46:48.941933 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.941918 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/4e8e0c9a-eac8-4af9-bd4c-927f0b406f02-data-volume\") pod \"insights-runtime-extractor-rnnj4\" (UID: \"4e8e0c9a-eac8-4af9-bd4c-927f0b406f02\") " pod="openshift-insights/insights-runtime-extractor-rnnj4" Apr 17 20:46:48.941978 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.941938 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/4e8e0c9a-eac8-4af9-bd4c-927f0b406f02-crio-socket\") pod \"insights-runtime-extractor-rnnj4\" (UID: \"4e8e0c9a-eac8-4af9-bd4c-927f0b406f02\") " pod="openshift-insights/insights-runtime-extractor-rnnj4" Apr 17 20:46:48.942016 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.941992 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/4e8e0c9a-eac8-4af9-bd4c-927f0b406f02-crio-socket\") pod \"insights-runtime-extractor-rnnj4\" (UID: \"4e8e0c9a-eac8-4af9-bd4c-927f0b406f02\") " pod="openshift-insights/insights-runtime-extractor-rnnj4" Apr 17 20:46:48.942066 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.942018 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vmtkn\" (UniqueName: \"kubernetes.io/projected/4e8e0c9a-eac8-4af9-bd4c-927f0b406f02-kube-api-access-vmtkn\") pod \"insights-runtime-extractor-rnnj4\" (UID: \"4e8e0c9a-eac8-4af9-bd4c-927f0b406f02\") " pod="openshift-insights/insights-runtime-extractor-rnnj4" Apr 17 20:46:48.942066 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.942057 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d0f3d88e-5398-4272-bf50-7675902545aa-registry-certificates\") pod \"image-registry-57fdff8cb7-8qxnn\" (UID: \"d0f3d88e-5398-4272-bf50-7675902545aa\") " pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:46:48.942156 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.942074 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d0f3d88e-5398-4272-bf50-7675902545aa-trusted-ca\") pod \"image-registry-57fdff8cb7-8qxnn\" (UID: \"d0f3d88e-5398-4272-bf50-7675902545aa\") " pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:46:48.942156 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.942118 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/d0f3d88e-5398-4272-bf50-7675902545aa-image-registry-private-configuration\") pod \"image-registry-57fdff8cb7-8qxnn\" (UID: \"d0f3d88e-5398-4272-bf50-7675902545aa\") " pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:46:48.942156 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.942148 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d0f3d88e-5398-4272-bf50-7675902545aa-ca-trust-extracted\") pod \"image-registry-57fdff8cb7-8qxnn\" (UID: \"d0f3d88e-5398-4272-bf50-7675902545aa\") " pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:46:48.942312 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.942221 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/4e8e0c9a-eac8-4af9-bd4c-927f0b406f02-data-volume\") pod \"insights-runtime-extractor-rnnj4\" (UID: \"4e8e0c9a-eac8-4af9-bd4c-927f0b406f02\") " pod="openshift-insights/insights-runtime-extractor-rnnj4" Apr 17 20:46:48.942312 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.942280 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/4e8e0c9a-eac8-4af9-bd4c-927f0b406f02-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-rnnj4\" (UID: \"4e8e0c9a-eac8-4af9-bd4c-927f0b406f02\") " pod="openshift-insights/insights-runtime-extractor-rnnj4" Apr 17 20:46:48.942388 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.942333 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/4e8e0c9a-eac8-4af9-bd4c-927f0b406f02-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-rnnj4\" (UID: \"4e8e0c9a-eac8-4af9-bd4c-927f0b406f02\") " pod="openshift-insights/insights-runtime-extractor-rnnj4" Apr 17 20:46:48.942716 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.942694 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d0f3d88e-5398-4272-bf50-7675902545aa-ca-trust-extracted\") pod \"image-registry-57fdff8cb7-8qxnn\" (UID: \"d0f3d88e-5398-4272-bf50-7675902545aa\") " pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:46:48.943171 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.943081 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/4e8e0c9a-eac8-4af9-bd4c-927f0b406f02-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-rnnj4\" (UID: \"4e8e0c9a-eac8-4af9-bd4c-927f0b406f02\") " pod="openshift-insights/insights-runtime-extractor-rnnj4" Apr 17 20:46:48.943171 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.943111 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d0f3d88e-5398-4272-bf50-7675902545aa-registry-certificates\") pod \"image-registry-57fdff8cb7-8qxnn\" (UID: \"d0f3d88e-5398-4272-bf50-7675902545aa\") " pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:46:48.943329 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.943292 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d0f3d88e-5398-4272-bf50-7675902545aa-trusted-ca\") pod \"image-registry-57fdff8cb7-8qxnn\" (UID: \"d0f3d88e-5398-4272-bf50-7675902545aa\") " pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:46:48.944525 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.944503 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d0f3d88e-5398-4272-bf50-7675902545aa-registry-tls\") pod \"image-registry-57fdff8cb7-8qxnn\" (UID: \"d0f3d88e-5398-4272-bf50-7675902545aa\") " pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:46:48.944858 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.944839 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d0f3d88e-5398-4272-bf50-7675902545aa-installation-pull-secrets\") pod \"image-registry-57fdff8cb7-8qxnn\" (UID: \"d0f3d88e-5398-4272-bf50-7675902545aa\") " pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:46:48.945028 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.945008 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/4e8e0c9a-eac8-4af9-bd4c-927f0b406f02-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-rnnj4\" (UID: \"4e8e0c9a-eac8-4af9-bd4c-927f0b406f02\") " pod="openshift-insights/insights-runtime-extractor-rnnj4" Apr 17 20:46:48.945120 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.945102 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/d0f3d88e-5398-4272-bf50-7675902545aa-image-registry-private-configuration\") pod \"image-registry-57fdff8cb7-8qxnn\" (UID: \"d0f3d88e-5398-4272-bf50-7675902545aa\") " pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:46:48.948603 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.948574 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmtkn\" (UniqueName: \"kubernetes.io/projected/4e8e0c9a-eac8-4af9-bd4c-927f0b406f02-kube-api-access-vmtkn\") pod \"insights-runtime-extractor-rnnj4\" (UID: \"4e8e0c9a-eac8-4af9-bd4c-927f0b406f02\") " pod="openshift-insights/insights-runtime-extractor-rnnj4" Apr 17 20:46:48.948953 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.948930 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d0f3d88e-5398-4272-bf50-7675902545aa-bound-sa-token\") pod \"image-registry-57fdff8cb7-8qxnn\" (UID: \"d0f3d88e-5398-4272-bf50-7675902545aa\") " pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:46:48.949146 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:48.949126 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9wn2\" (UniqueName: \"kubernetes.io/projected/d0f3d88e-5398-4272-bf50-7675902545aa-kube-api-access-h9wn2\") pod \"image-registry-57fdff8cb7-8qxnn\" (UID: \"d0f3d88e-5398-4272-bf50-7675902545aa\") " pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:46:49.002136 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:49.002111 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-runtime-extractor-rnnj4" Apr 17 20:46:49.072825 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:49.072686 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:46:49.139236 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:49.139210 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-runtime-extractor-rnnj4"] Apr 17 20:46:49.141748 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:46:49.141722 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e8e0c9a_eac8_4af9_bd4c_927f0b406f02.slice/crio-46d1018acf7ea9f2ad893fa934f65b1ea24cf3ffeea9cf56b0c8755cae68b4c7 WatchSource:0}: Error finding container 46d1018acf7ea9f2ad893fa934f65b1ea24cf3ffeea9cf56b0c8755cae68b4c7: Status 404 returned error can't find the container with id 46d1018acf7ea9f2ad893fa934f65b1ea24cf3ffeea9cf56b0c8755cae68b4c7 Apr 17 20:46:49.197094 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:49.197073 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-57fdff8cb7-8qxnn"] Apr 17 20:46:49.199083 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:46:49.199063 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0f3d88e_5398_4272_bf50_7675902545aa.slice/crio-8bbca3e2281946c20396ebe798e87ccf4a7f15e7ef6f2954611192335ee38e4f WatchSource:0}: Error finding container 8bbca3e2281946c20396ebe798e87ccf4a7f15e7ef6f2954611192335ee38e4f: Status 404 returned error can't find the container with id 8bbca3e2281946c20396ebe798e87ccf4a7f15e7ef6f2954611192335ee38e4f Apr 17 20:46:49.506825 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:49.506735 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-rnnj4" event={"ID":"4e8e0c9a-eac8-4af9-bd4c-927f0b406f02","Type":"ContainerStarted","Data":"12c37e7a85de993208738fd62697fb8400b740f835bc627ab9a9acf7a45fea2e"} Apr 17 20:46:49.506825 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:49.506770 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-rnnj4" event={"ID":"4e8e0c9a-eac8-4af9-bd4c-927f0b406f02","Type":"ContainerStarted","Data":"46d1018acf7ea9f2ad893fa934f65b1ea24cf3ffeea9cf56b0c8755cae68b4c7"} Apr 17 20:46:49.508072 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:49.508048 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" event={"ID":"d0f3d88e-5398-4272-bf50-7675902545aa","Type":"ContainerStarted","Data":"cdbf8013f0887df1ebf1c0f1e19d600dcfbafbf1ca4bf72b8960dd7e7fbefe40"} Apr 17 20:46:49.508185 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:49.508078 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" event={"ID":"d0f3d88e-5398-4272-bf50-7675902545aa","Type":"ContainerStarted","Data":"8bbca3e2281946c20396ebe798e87ccf4a7f15e7ef6f2954611192335ee38e4f"} Apr 17 20:46:49.508185 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:49.508169 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:46:49.527235 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:49.527196 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" podStartSLOduration=1.52718391 podStartE2EDuration="1.52718391s" podCreationTimestamp="2026-04-17 20:46:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 20:46:49.526114993 +0000 UTC m=+134.999893280" watchObservedRunningTime="2026-04-17 20:46:49.52718391 +0000 UTC m=+135.000962220" Apr 17 20:46:50.512231 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:50.512197 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-rnnj4" event={"ID":"4e8e0c9a-eac8-4af9-bd4c-927f0b406f02","Type":"ContainerStarted","Data":"9e172ed0ab9c59a120e8622c1a5086b7087cc9d0fadda7fb6f5e8dae20c58561"} Apr 17 20:46:50.732552 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:50.732515 2577 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" Apr 17 20:46:50.732552 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:50.732558 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" Apr 17 20:46:50.732932 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:50.732916 2577 scope.go:117] "RemoveContainer" containerID="593b07a09262e1a26741d3fa8f78da22e06081a698ef09a33fef086a57e4c88c" Apr 17 20:46:50.733123 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:46:50.733103 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"console-operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=console-operator pod=console-operator-9d4b6777b-9rmnh_openshift-console-operator(e562e939-99f0-4bee-952d-7af46132c860)\"" pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" podUID="e562e939-99f0-4bee-952d-7af46132c860" Apr 17 20:46:51.516600 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:51.516570 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-rnnj4" event={"ID":"4e8e0c9a-eac8-4af9-bd4c-927f0b406f02","Type":"ContainerStarted","Data":"a277b3a526ab8a7c480f195082abbb4258031f64cd35057f8481e97259506c39"} Apr 17 20:46:51.531934 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:51.531891 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-insights/insights-runtime-extractor-rnnj4" podStartSLOduration=1.332120219 podStartE2EDuration="3.531880509s" podCreationTimestamp="2026-04-17 20:46:48 +0000 UTC" firstStartedPulling="2026-04-17 20:46:49.201925577 +0000 UTC m=+134.675703846" lastFinishedPulling="2026-04-17 20:46:51.401685868 +0000 UTC m=+136.875464136" observedRunningTime="2026-04-17 20:46:51.530752427 +0000 UTC m=+137.004530713" watchObservedRunningTime="2026-04-17 20:46:51.531880509 +0000 UTC m=+137.005658796" Apr 17 20:46:52.268847 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:52.268788 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/5506563e-216d-4b61-97dc-8cea5be3f474-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-7cqn6\" (UID: \"5506563e-216d-4b61-97dc-8cea5be3f474\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-7cqn6" Apr 17 20:46:52.271282 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:52.271254 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/5506563e-216d-4b61-97dc-8cea5be3f474-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-75587bd455-7cqn6\" (UID: \"5506563e-216d-4b61-97dc-8cea5be3f474\") " pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-7cqn6" Apr 17 20:46:52.525710 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:52.525643 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"cluster-monitoring-operator-dockercfg-kx4g8\"" Apr 17 20:46:52.534277 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:52.534257 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-7cqn6" Apr 17 20:46:52.653345 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:52.653311 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-75587bd455-7cqn6"] Apr 17 20:46:52.655713 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:46:52.655686 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5506563e_216d_4b61_97dc_8cea5be3f474.slice/crio-57d87558825cd1add6a2391ec6f094b4ff78e4b2e35ebf0f226d4dda9741d692 WatchSource:0}: Error finding container 57d87558825cd1add6a2391ec6f094b4ff78e4b2e35ebf0f226d4dda9741d692: Status 404 returned error can't find the container with id 57d87558825cd1add6a2391ec6f094b4ff78e4b2e35ebf0f226d4dda9741d692 Apr 17 20:46:53.523334 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:53.523303 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-7cqn6" event={"ID":"5506563e-216d-4b61-97dc-8cea5be3f474","Type":"ContainerStarted","Data":"57d87558825cd1add6a2391ec6f094b4ff78e4b2e35ebf0f226d4dda9741d692"} Apr 17 20:46:54.527722 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:54.527682 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-7cqn6" event={"ID":"5506563e-216d-4b61-97dc-8cea5be3f474","Type":"ContainerStarted","Data":"602ea8857e99c63fe414e87e3336a0776bf4218e05976351865a490972e405c8"} Apr 17 20:46:54.541038 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:46:54.540998 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/cluster-monitoring-operator-75587bd455-7cqn6" podStartSLOduration=32.813942545 podStartE2EDuration="34.540986058s" podCreationTimestamp="2026-04-17 20:46:20 +0000 UTC" firstStartedPulling="2026-04-17 20:46:52.65751251 +0000 UTC m=+138.131290774" lastFinishedPulling="2026-04-17 20:46:54.384556017 +0000 UTC m=+139.858334287" observedRunningTime="2026-04-17 20:46:54.540373971 +0000 UTC m=+140.014152258" watchObservedRunningTime="2026-04-17 20:46:54.540986058 +0000 UTC m=+140.014764345" Apr 17 20:47:02.295174 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.295139 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/node-exporter-lpkzb"] Apr 17 20:47:02.327870 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.327844 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:02.330364 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.330330 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-tls\"" Apr 17 20:47:02.333525 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.333503 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-kube-rbac-proxy-config\"" Apr 17 20:47:02.334297 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.334016 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"node-exporter-accelerators-collector-config\"" Apr 17 20:47:02.334297 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.334039 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-dockercfg-s8fns\"" Apr 17 20:47:02.334297 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.334161 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"metrics-client-ca\"" Apr 17 20:47:02.443947 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.443919 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsmrg\" (UniqueName: \"kubernetes.io/projected/a044b44f-06c8-425f-9f9f-24aaeddea464-kube-api-access-bsmrg\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:02.444075 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.443957 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/a044b44f-06c8-425f-9f9f-24aaeddea464-node-exporter-textfile\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:02.444075 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.443978 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/a044b44f-06c8-425f-9f9f-24aaeddea464-node-exporter-accelerators-collector-config\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:02.444075 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.444060 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/a044b44f-06c8-425f-9f9f-24aaeddea464-root\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:02.444239 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.444080 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a044b44f-06c8-425f-9f9f-24aaeddea464-sys\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:02.444239 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.444182 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/a044b44f-06c8-425f-9f9f-24aaeddea464-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:02.444239 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.444223 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/a044b44f-06c8-425f-9f9f-24aaeddea464-node-exporter-wtmp\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:02.444351 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.444286 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/a044b44f-06c8-425f-9f9f-24aaeddea464-metrics-client-ca\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:02.444351 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.444318 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/a044b44f-06c8-425f-9f9f-24aaeddea464-node-exporter-tls\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:02.544875 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.544846 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/a044b44f-06c8-425f-9f9f-24aaeddea464-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:02.544997 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.544881 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/a044b44f-06c8-425f-9f9f-24aaeddea464-node-exporter-wtmp\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:02.545059 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.545012 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/a044b44f-06c8-425f-9f9f-24aaeddea464-metrics-client-ca\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:02.545059 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.545034 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/a044b44f-06c8-425f-9f9f-24aaeddea464-node-exporter-wtmp\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:02.545059 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.545047 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/a044b44f-06c8-425f-9f9f-24aaeddea464-node-exporter-tls\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:02.545290 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:47:02.545149 2577 secret.go:189] Couldn't get secret openshift-monitoring/node-exporter-tls: secret "node-exporter-tls" not found Apr 17 20:47:02.545290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.545172 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-bsmrg\" (UniqueName: \"kubernetes.io/projected/a044b44f-06c8-425f-9f9f-24aaeddea464-kube-api-access-bsmrg\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:02.545290 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:47:02.545222 2577 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a044b44f-06c8-425f-9f9f-24aaeddea464-node-exporter-tls podName:a044b44f-06c8-425f-9f9f-24aaeddea464 nodeName:}" failed. No retries permitted until 2026-04-17 20:47:03.045202326 +0000 UTC m=+148.518980594 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-exporter-tls" (UniqueName: "kubernetes.io/secret/a044b44f-06c8-425f-9f9f-24aaeddea464-node-exporter-tls") pod "node-exporter-lpkzb" (UID: "a044b44f-06c8-425f-9f9f-24aaeddea464") : secret "node-exporter-tls" not found Apr 17 20:47:02.545290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.545244 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/a044b44f-06c8-425f-9f9f-24aaeddea464-node-exporter-textfile\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:02.545290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.545276 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/a044b44f-06c8-425f-9f9f-24aaeddea464-node-exporter-accelerators-collector-config\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:02.545571 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.545307 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/a044b44f-06c8-425f-9f9f-24aaeddea464-root\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:02.545571 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.545335 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a044b44f-06c8-425f-9f9f-24aaeddea464-sys\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:02.545571 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.545435 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a044b44f-06c8-425f-9f9f-24aaeddea464-sys\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:02.545571 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.545503 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/a044b44f-06c8-425f-9f9f-24aaeddea464-root\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:02.545765 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.545603 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/a044b44f-06c8-425f-9f9f-24aaeddea464-metrics-client-ca\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:02.545765 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.545742 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/a044b44f-06c8-425f-9f9f-24aaeddea464-node-exporter-textfile\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:02.545862 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.545797 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/a044b44f-06c8-425f-9f9f-24aaeddea464-node-exporter-accelerators-collector-config\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:02.547553 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.547534 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/a044b44f-06c8-425f-9f9f-24aaeddea464-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:02.554734 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:02.554714 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsmrg\" (UniqueName: \"kubernetes.io/projected/a044b44f-06c8-425f-9f9f-24aaeddea464-kube-api-access-bsmrg\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:03.050467 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:03.050433 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/a044b44f-06c8-425f-9f9f-24aaeddea464-node-exporter-tls\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:03.052824 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:03.052778 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/a044b44f-06c8-425f-9f9f-24aaeddea464-node-exporter-tls\") pod \"node-exporter-lpkzb\" (UID: \"a044b44f-06c8-425f-9f9f-24aaeddea464\") " pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:03.236437 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:03.236407 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-lpkzb" Apr 17 20:47:03.244788 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:47:03.244758 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda044b44f_06c8_425f_9f9f_24aaeddea464.slice/crio-deec2e291dfe078dc72d12be4d77275dbcb466d1ac9fa0ea40a8ae6f191466a2 WatchSource:0}: Error finding container deec2e291dfe078dc72d12be4d77275dbcb466d1ac9fa0ea40a8ae6f191466a2: Status 404 returned error can't find the container with id deec2e291dfe078dc72d12be4d77275dbcb466d1ac9fa0ea40a8ae6f191466a2 Apr 17 20:47:03.550564 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:03.550530 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-lpkzb" event={"ID":"a044b44f-06c8-425f-9f9f-24aaeddea464","Type":"ContainerStarted","Data":"deec2e291dfe078dc72d12be4d77275dbcb466d1ac9fa0ea40a8ae6f191466a2"} Apr 17 20:47:04.554371 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:04.554339 2577 generic.go:358] "Generic (PLEG): container finished" podID="a044b44f-06c8-425f-9f9f-24aaeddea464" containerID="65306f5ed91911cbef3d51cbf3f5d5cfc4fce0788e182359af34269673f1a114" exitCode=0 Apr 17 20:47:04.554818 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:04.554410 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-lpkzb" event={"ID":"a044b44f-06c8-425f-9f9f-24aaeddea464","Type":"ContainerDied","Data":"65306f5ed91911cbef3d51cbf3f5d5cfc4fce0788e182359af34269673f1a114"} Apr 17 20:47:05.117941 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:05.117911 2577 scope.go:117] "RemoveContainer" containerID="593b07a09262e1a26741d3fa8f78da22e06081a698ef09a33fef086a57e4c88c" Apr 17 20:47:05.118124 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:47:05.118106 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"console-operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=console-operator pod=console-operator-9d4b6777b-9rmnh_openshift-console-operator(e562e939-99f0-4bee-952d-7af46132c860)\"" pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" podUID="e562e939-99f0-4bee-952d-7af46132c860" Apr 17 20:47:05.558527 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:05.558487 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-lpkzb" event={"ID":"a044b44f-06c8-425f-9f9f-24aaeddea464","Type":"ContainerStarted","Data":"7e79f46b49bc51c6556b59ca167ccc8738c06c80aecdf28c267ca981ab21934f"} Apr 17 20:47:05.558527 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:05.558529 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-lpkzb" event={"ID":"a044b44f-06c8-425f-9f9f-24aaeddea464","Type":"ContainerStarted","Data":"1f3e2a80aa3384d3a151a62abc9ccb7ba5b524af19bef45fab368dec2a0ef3fc"} Apr 17 20:47:05.575705 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:05.575655 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/node-exporter-lpkzb" podStartSLOduration=2.718678244 podStartE2EDuration="3.575589339s" podCreationTimestamp="2026-04-17 20:47:02 +0000 UTC" firstStartedPulling="2026-04-17 20:47:03.246722634 +0000 UTC m=+148.720500900" lastFinishedPulling="2026-04-17 20:47:04.103633725 +0000 UTC m=+149.577411995" observedRunningTime="2026-04-17 20:47:05.573925855 +0000 UTC m=+151.047704139" watchObservedRunningTime="2026-04-17 20:47:05.575589339 +0000 UTC m=+151.049367626" Apr 17 20:47:06.782745 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.782714 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/metrics-server-5c998b9cc6-kqcc5"] Apr 17 20:47:06.785986 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.785965 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" Apr 17 20:47:06.788472 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.788451 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-server-a78euhtn4djlc\"" Apr 17 20:47:06.788578 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.788508 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"kubelet-serving-ca-bundle\"" Apr 17 20:47:06.788578 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.788562 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-server-tls\"" Apr 17 20:47:06.788707 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.788575 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"metrics-server-audit-profiles\"" Apr 17 20:47:06.788707 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.788453 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-server-client-certs\"" Apr 17 20:47:06.788846 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.788832 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-server-dockercfg-sbb5r\"" Apr 17 20:47:06.793244 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.793222 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-5c998b9cc6-kqcc5"] Apr 17 20:47:06.874339 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.874313 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/595b0fbe-52fe-4bf3-be9d-8249e0ca08aa-client-ca-bundle\") pod \"metrics-server-5c998b9cc6-kqcc5\" (UID: \"595b0fbe-52fe-4bf3-be9d-8249e0ca08aa\") " pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" Apr 17 20:47:06.874442 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.874368 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/595b0fbe-52fe-4bf3-be9d-8249e0ca08aa-secret-metrics-server-tls\") pod \"metrics-server-5c998b9cc6-kqcc5\" (UID: \"595b0fbe-52fe-4bf3-be9d-8249e0ca08aa\") " pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" Apr 17 20:47:06.874442 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.874405 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-server-client-certs\" (UniqueName: \"kubernetes.io/secret/595b0fbe-52fe-4bf3-be9d-8249e0ca08aa-secret-metrics-server-client-certs\") pod \"metrics-server-5c998b9cc6-kqcc5\" (UID: \"595b0fbe-52fe-4bf3-be9d-8249e0ca08aa\") " pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" Apr 17 20:47:06.874442 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.874424 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/595b0fbe-52fe-4bf3-be9d-8249e0ca08aa-metrics-server-audit-profiles\") pod \"metrics-server-5c998b9cc6-kqcc5\" (UID: \"595b0fbe-52fe-4bf3-be9d-8249e0ca08aa\") " pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" Apr 17 20:47:06.874552 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.874447 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/595b0fbe-52fe-4bf3-be9d-8249e0ca08aa-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-5c998b9cc6-kqcc5\" (UID: \"595b0fbe-52fe-4bf3-be9d-8249e0ca08aa\") " pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" Apr 17 20:47:06.874552 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.874526 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgd74\" (UniqueName: \"kubernetes.io/projected/595b0fbe-52fe-4bf3-be9d-8249e0ca08aa-kube-api-access-mgd74\") pod \"metrics-server-5c998b9cc6-kqcc5\" (UID: \"595b0fbe-52fe-4bf3-be9d-8249e0ca08aa\") " pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" Apr 17 20:47:06.874613 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.874570 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/595b0fbe-52fe-4bf3-be9d-8249e0ca08aa-audit-log\") pod \"metrics-server-5c998b9cc6-kqcc5\" (UID: \"595b0fbe-52fe-4bf3-be9d-8249e0ca08aa\") " pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" Apr 17 20:47:06.975203 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.975177 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-metrics-server-client-certs\" (UniqueName: \"kubernetes.io/secret/595b0fbe-52fe-4bf3-be9d-8249e0ca08aa-secret-metrics-server-client-certs\") pod \"metrics-server-5c998b9cc6-kqcc5\" (UID: \"595b0fbe-52fe-4bf3-be9d-8249e0ca08aa\") " pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" Apr 17 20:47:06.975321 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.975211 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/595b0fbe-52fe-4bf3-be9d-8249e0ca08aa-metrics-server-audit-profiles\") pod \"metrics-server-5c998b9cc6-kqcc5\" (UID: \"595b0fbe-52fe-4bf3-be9d-8249e0ca08aa\") " pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" Apr 17 20:47:06.975321 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.975236 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/595b0fbe-52fe-4bf3-be9d-8249e0ca08aa-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-5c998b9cc6-kqcc5\" (UID: \"595b0fbe-52fe-4bf3-be9d-8249e0ca08aa\") " pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" Apr 17 20:47:06.975321 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.975278 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-mgd74\" (UniqueName: \"kubernetes.io/projected/595b0fbe-52fe-4bf3-be9d-8249e0ca08aa-kube-api-access-mgd74\") pod \"metrics-server-5c998b9cc6-kqcc5\" (UID: \"595b0fbe-52fe-4bf3-be9d-8249e0ca08aa\") " pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" Apr 17 20:47:06.975321 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.975309 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/595b0fbe-52fe-4bf3-be9d-8249e0ca08aa-audit-log\") pod \"metrics-server-5c998b9cc6-kqcc5\" (UID: \"595b0fbe-52fe-4bf3-be9d-8249e0ca08aa\") " pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" Apr 17 20:47:06.975514 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.975359 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/595b0fbe-52fe-4bf3-be9d-8249e0ca08aa-client-ca-bundle\") pod \"metrics-server-5c998b9cc6-kqcc5\" (UID: \"595b0fbe-52fe-4bf3-be9d-8249e0ca08aa\") " pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" Apr 17 20:47:06.975571 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.975543 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/595b0fbe-52fe-4bf3-be9d-8249e0ca08aa-secret-metrics-server-tls\") pod \"metrics-server-5c998b9cc6-kqcc5\" (UID: \"595b0fbe-52fe-4bf3-be9d-8249e0ca08aa\") " pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" Apr 17 20:47:06.975753 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.975725 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/595b0fbe-52fe-4bf3-be9d-8249e0ca08aa-audit-log\") pod \"metrics-server-5c998b9cc6-kqcc5\" (UID: \"595b0fbe-52fe-4bf3-be9d-8249e0ca08aa\") " pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" Apr 17 20:47:06.976089 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.976060 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/595b0fbe-52fe-4bf3-be9d-8249e0ca08aa-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-5c998b9cc6-kqcc5\" (UID: \"595b0fbe-52fe-4bf3-be9d-8249e0ca08aa\") " pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" Apr 17 20:47:06.976249 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.976204 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/595b0fbe-52fe-4bf3-be9d-8249e0ca08aa-metrics-server-audit-profiles\") pod \"metrics-server-5c998b9cc6-kqcc5\" (UID: \"595b0fbe-52fe-4bf3-be9d-8249e0ca08aa\") " pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" Apr 17 20:47:06.978009 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.977989 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/595b0fbe-52fe-4bf3-be9d-8249e0ca08aa-secret-metrics-server-tls\") pod \"metrics-server-5c998b9cc6-kqcc5\" (UID: \"595b0fbe-52fe-4bf3-be9d-8249e0ca08aa\") " pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" Apr 17 20:47:06.978398 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.978379 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-metrics-server-client-certs\" (UniqueName: \"kubernetes.io/secret/595b0fbe-52fe-4bf3-be9d-8249e0ca08aa-secret-metrics-server-client-certs\") pod \"metrics-server-5c998b9cc6-kqcc5\" (UID: \"595b0fbe-52fe-4bf3-be9d-8249e0ca08aa\") " pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" Apr 17 20:47:06.978542 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.978520 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/595b0fbe-52fe-4bf3-be9d-8249e0ca08aa-client-ca-bundle\") pod \"metrics-server-5c998b9cc6-kqcc5\" (UID: \"595b0fbe-52fe-4bf3-be9d-8249e0ca08aa\") " pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" Apr 17 20:47:06.982706 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:06.982681 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgd74\" (UniqueName: \"kubernetes.io/projected/595b0fbe-52fe-4bf3-be9d-8249e0ca08aa-kube-api-access-mgd74\") pod \"metrics-server-5c998b9cc6-kqcc5\" (UID: \"595b0fbe-52fe-4bf3-be9d-8249e0ca08aa\") " pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" Apr 17 20:47:07.095650 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:07.095596 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" Apr 17 20:47:07.209633 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:07.209610 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-5c998b9cc6-kqcc5"] Apr 17 20:47:07.212221 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:47:07.212188 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod595b0fbe_52fe_4bf3_be9d_8249e0ca08aa.slice/crio-839f5c8e82d06f1e879f21e37649de0a7da4ba75f9ebf01a8ff014147424e9d4 WatchSource:0}: Error finding container 839f5c8e82d06f1e879f21e37649de0a7da4ba75f9ebf01a8ff014147424e9d4: Status 404 returned error can't find the container with id 839f5c8e82d06f1e879f21e37649de0a7da4ba75f9ebf01a8ff014147424e9d4 Apr 17 20:47:07.565092 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:07.565061 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" event={"ID":"595b0fbe-52fe-4bf3-be9d-8249e0ca08aa","Type":"ContainerStarted","Data":"839f5c8e82d06f1e879f21e37649de0a7da4ba75f9ebf01a8ff014147424e9d4"} Apr 17 20:47:08.486313 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.486285 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Apr 17 20:47:08.493018 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.492995 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.495172 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.495147 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-tls-assets-0\"" Apr 17 20:47:08.496004 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.495980 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-thanos-sidecar-tls\"" Apr 17 20:47:08.496004 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.495998 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-thanos-prometheus-http-client-file\"" Apr 17 20:47:08.496183 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.496103 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"serving-certs-ca-bundle\"" Apr 17 20:47:08.496183 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.496136 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-grpc-tls-9hk0s6bkcm5i8\"" Apr 17 20:47:08.496290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.496278 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-web-config\"" Apr 17 20:47:08.497605 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.497587 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-client-certs\"" Apr 17 20:47:08.497830 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.497795 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s\"" Apr 17 20:47:08.497921 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.497636 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-tls\"" Apr 17 20:47:08.497979 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.497591 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"kube-rbac-proxy\"" Apr 17 20:47:08.498116 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.498098 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-dockercfg-kwhx4\"" Apr 17 20:47:08.498116 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.498108 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-kube-rbac-proxy-web\"" Apr 17 20:47:08.500281 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.500246 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-rulefiles-0\"" Apr 17 20:47:08.502781 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.502661 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Apr 17 20:47:08.504603 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.504579 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"prometheus-trusted-ca-bundle\"" Apr 17 20:47:08.588425 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.588388 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.588577 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.588445 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.588577 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.588487 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.588577 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.588521 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-web-config\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.588577 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.588548 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.588577 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.588566 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.588870 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.588639 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.588870 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.588696 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2505b439-b489-4c27-a41c-3cacc6db2776-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.588870 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.588717 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2505b439-b489-4c27-a41c-3cacc6db2776-config-out\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.588870 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.588742 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-config\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.588870 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.588855 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.589129 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.588908 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.589129 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.588943 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.589129 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.588972 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/2505b439-b489-4c27-a41c-3cacc6db2776-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.589129 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.589009 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.589129 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.589035 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cb5vb\" (UniqueName: \"kubernetes.io/projected/2505b439-b489-4c27-a41c-3cacc6db2776-kube-api-access-cb5vb\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.589129 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.589086 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.589324 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.589130 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.689634 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.689605 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.689739 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.689651 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.689739 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.689673 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.689739 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.689690 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-web-config\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.689739 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.689712 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.689739 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.689734 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.690979 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.690163 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.690979 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.690217 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2505b439-b489-4c27-a41c-3cacc6db2776-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.690979 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.690244 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2505b439-b489-4c27-a41c-3cacc6db2776-config-out\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.690979 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.690270 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-config\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.690979 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.690361 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.690979 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.690426 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.690979 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.690540 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.690979 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.690572 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/2505b439-b489-4c27-a41c-3cacc6db2776-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.690979 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.690613 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.690979 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.690641 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-cb5vb\" (UniqueName: \"kubernetes.io/projected/2505b439-b489-4c27-a41c-3cacc6db2776-kube-api-access-cb5vb\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.690979 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.690690 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.690979 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.690721 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.690979 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.690948 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.693162 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.691521 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.693162 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.692149 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.693162 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.692756 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.693162 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.692944 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.693162 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.692957 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-web-config\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.693428 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.693240 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.693428 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.693295 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/2505b439-b489-4c27-a41c-3cacc6db2776-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.694507 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.693786 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.694507 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.694025 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.694507 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.694471 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.695274 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.695248 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2505b439-b489-4c27-a41c-3cacc6db2776-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.695760 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.695743 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.696070 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.696048 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.696644 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.696629 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-config\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.697128 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.697107 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2505b439-b489-4c27-a41c-3cacc6db2776-config-out\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.698418 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.698392 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.700925 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.700907 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-cb5vb\" (UniqueName: \"kubernetes.io/projected/2505b439-b489-4c27-a41c-3cacc6db2776-kube-api-access-cb5vb\") pod \"prometheus-k8s-0\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.806121 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.806096 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:08.933830 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:08.933697 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Apr 17 20:47:08.935845 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:47:08.935818 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2505b439_b489_4c27_a41c_3cacc6db2776.slice/crio-71254b6ec62c33bd4dd03ff557729030f823662f77e8da946574970e6bbc4526 WatchSource:0}: Error finding container 71254b6ec62c33bd4dd03ff557729030f823662f77e8da946574970e6bbc4526: Status 404 returned error can't find the container with id 71254b6ec62c33bd4dd03ff557729030f823662f77e8da946574970e6bbc4526 Apr 17 20:47:09.077036 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:09.076955 2577 patch_prober.go:28] interesting pod/image-registry-57fdff8cb7-8qxnn container/registry namespace/openshift-image-registry: Liveness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body={"errors":[{"code":"UNAVAILABLE","message":"service unavailable","detail":"health check failed: please see /debug/health"}]} Apr 17 20:47:09.077155 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:09.077023 2577 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" podUID="d0f3d88e-5398-4272-bf50-7675902545aa" containerName="registry" probeResult="failure" output="HTTP probe failed with statuscode: 503" Apr 17 20:47:09.571486 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:09.571450 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"2505b439-b489-4c27-a41c-3cacc6db2776","Type":"ContainerStarted","Data":"71254b6ec62c33bd4dd03ff557729030f823662f77e8da946574970e6bbc4526"} Apr 17 20:47:09.572723 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:09.572699 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" event={"ID":"595b0fbe-52fe-4bf3-be9d-8249e0ca08aa","Type":"ContainerStarted","Data":"e41180493897e08bd732cbb018ab6f005358354bd14dc8dca7ef81a3bca854c4"} Apr 17 20:47:09.587265 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:09.587221 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" podStartSLOduration=2.143064946 podStartE2EDuration="3.587207686s" podCreationTimestamp="2026-04-17 20:47:06 +0000 UTC" firstStartedPulling="2026-04-17 20:47:07.214501876 +0000 UTC m=+152.688280141" lastFinishedPulling="2026-04-17 20:47:08.658644616 +0000 UTC m=+154.132422881" observedRunningTime="2026-04-17 20:47:09.586362328 +0000 UTC m=+155.060140615" watchObservedRunningTime="2026-04-17 20:47:09.587207686 +0000 UTC m=+155.060985973" Apr 17 20:47:10.516603 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:10.516574 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-57fdff8cb7-8qxnn" Apr 17 20:47:10.577552 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:10.577516 2577 generic.go:358] "Generic (PLEG): container finished" podID="2505b439-b489-4c27-a41c-3cacc6db2776" containerID="77783870ae1deea2166d5abe8ddfbea7b14497deb4bfc7b3e499413f6f7e1c4c" exitCode=0 Apr 17 20:47:10.577944 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:10.577609 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"2505b439-b489-4c27-a41c-3cacc6db2776","Type":"ContainerDied","Data":"77783870ae1deea2166d5abe8ddfbea7b14497deb4bfc7b3e499413f6f7e1c4c"} Apr 17 20:47:11.434648 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:47:11.434605 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[metrics-tls], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-dns/dns-default-bc6wf" podUID="821ba3d4-448f-45b9-bcec-701682592a0f" Apr 17 20:47:11.439746 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:47:11.439718 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[cert], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-ingress-canary/ingress-canary-q59g7" podUID="79b05e51-04fa-4473-b612-c762aa0e3856" Apr 17 20:47:11.580079 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:11.580048 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-q59g7" Apr 17 20:47:11.580495 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:11.580055 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-bc6wf" Apr 17 20:47:13.588499 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:13.588426 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"2505b439-b489-4c27-a41c-3cacc6db2776","Type":"ContainerStarted","Data":"a7db2bf1e97eae79efc8f9a1887ab095b2d2ad4cd5aaec2f7795fa8b4f4934db"} Apr 17 20:47:13.588499 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:13.588458 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"2505b439-b489-4c27-a41c-3cacc6db2776","Type":"ContainerStarted","Data":"1eea24c6739c5904842a5d0c0c5a820e903433aaf22a2642e06fd8995c1dcd1a"} Apr 17 20:47:15.597157 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:15.597126 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"2505b439-b489-4c27-a41c-3cacc6db2776","Type":"ContainerStarted","Data":"f9df59a5ded69de59ac6100a894265062e32e38a0a3bd44901cbbe4c535f07b5"} Apr 17 20:47:15.597497 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:15.597161 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"2505b439-b489-4c27-a41c-3cacc6db2776","Type":"ContainerStarted","Data":"73bb05bea309ada12836a23f856d5e6dd52c4fb597ea35970d2092c727d510d5"} Apr 17 20:47:15.597497 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:15.597172 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"2505b439-b489-4c27-a41c-3cacc6db2776","Type":"ContainerStarted","Data":"2bc856f7e8d2ceacf86758e494ef3f3fccf228327419e5bc907a74bc455bbad8"} Apr 17 20:47:15.597497 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:15.597180 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"2505b439-b489-4c27-a41c-3cacc6db2776","Type":"ContainerStarted","Data":"7a699d681b62bc5e090c3b6d9f919fa04f05e19dcbfa6f9ae31d251d3ec3bf77"} Apr 17 20:47:15.621183 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:15.621140 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-k8s-0" podStartSLOduration=1.43006378 podStartE2EDuration="7.621125162s" podCreationTimestamp="2026-04-17 20:47:08 +0000 UTC" firstStartedPulling="2026-04-17 20:47:08.937782787 +0000 UTC m=+154.411561052" lastFinishedPulling="2026-04-17 20:47:15.128844168 +0000 UTC m=+160.602622434" observedRunningTime="2026-04-17 20:47:15.619696753 +0000 UTC m=+161.093475040" watchObservedRunningTime="2026-04-17 20:47:15.621125162 +0000 UTC m=+161.094903477" Apr 17 20:47:16.356472 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:16.356437 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79b05e51-04fa-4473-b612-c762aa0e3856-cert\") pod \"ingress-canary-q59g7\" (UID: \"79b05e51-04fa-4473-b612-c762aa0e3856\") " pod="openshift-ingress-canary/ingress-canary-q59g7" Apr 17 20:47:16.356666 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:16.356493 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/821ba3d4-448f-45b9-bcec-701682592a0f-metrics-tls\") pod \"dns-default-bc6wf\" (UID: \"821ba3d4-448f-45b9-bcec-701682592a0f\") " pod="openshift-dns/dns-default-bc6wf" Apr 17 20:47:16.359113 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:16.359085 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/821ba3d4-448f-45b9-bcec-701682592a0f-metrics-tls\") pod \"dns-default-bc6wf\" (UID: \"821ba3d4-448f-45b9-bcec-701682592a0f\") " pod="openshift-dns/dns-default-bc6wf" Apr 17 20:47:16.359223 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:16.359131 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/79b05e51-04fa-4473-b612-c762aa0e3856-cert\") pod \"ingress-canary-q59g7\" (UID: \"79b05e51-04fa-4473-b612-c762aa0e3856\") " pod="openshift-ingress-canary/ingress-canary-q59g7" Apr 17 20:47:16.383042 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:16.383016 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"default-dockercfg-pn5wd\"" Apr 17 20:47:16.383042 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:16.383022 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-dockercfg-8pvd4\"" Apr 17 20:47:16.391416 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:16.391392 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-bc6wf" Apr 17 20:47:16.391517 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:16.391478 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-q59g7" Apr 17 20:47:16.521075 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:16.521051 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-q59g7"] Apr 17 20:47:16.522908 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:47:16.522876 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod79b05e51_04fa_4473_b612_c762aa0e3856.slice/crio-848cbafd41a1df2979d91a296ab551169731790361328090db9452e71162f37a WatchSource:0}: Error finding container 848cbafd41a1df2979d91a296ab551169731790361328090db9452e71162f37a: Status 404 returned error can't find the container with id 848cbafd41a1df2979d91a296ab551169731790361328090db9452e71162f37a Apr 17 20:47:16.534915 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:16.534893 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-bc6wf"] Apr 17 20:47:16.537012 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:47:16.536990 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod821ba3d4_448f_45b9_bcec_701682592a0f.slice/crio-16ff1b02e73634031a7a6e8f98d8fd455c6aaaf14e6ffbc5cf29d5b9905cab52 WatchSource:0}: Error finding container 16ff1b02e73634031a7a6e8f98d8fd455c6aaaf14e6ffbc5cf29d5b9905cab52: Status 404 returned error can't find the container with id 16ff1b02e73634031a7a6e8f98d8fd455c6aaaf14e6ffbc5cf29d5b9905cab52 Apr 17 20:47:16.600366 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:16.600343 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-q59g7" event={"ID":"79b05e51-04fa-4473-b612-c762aa0e3856","Type":"ContainerStarted","Data":"848cbafd41a1df2979d91a296ab551169731790361328090db9452e71162f37a"} Apr 17 20:47:16.601417 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:16.601395 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-bc6wf" event={"ID":"821ba3d4-448f-45b9-bcec-701682592a0f","Type":"ContainerStarted","Data":"16ff1b02e73634031a7a6e8f98d8fd455c6aaaf14e6ffbc5cf29d5b9905cab52"} Apr 17 20:47:18.116239 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:18.116205 2577 scope.go:117] "RemoveContainer" containerID="593b07a09262e1a26741d3fa8f78da22e06081a698ef09a33fef086a57e4c88c" Apr 17 20:47:18.806662 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:18.806636 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:47:19.613978 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:19.613937 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-q59g7" event={"ID":"79b05e51-04fa-4473-b612-c762aa0e3856","Type":"ContainerStarted","Data":"2c6d0c95eeb53e0add2b20472710d24d2f94026bebcad9a2249a946759e00d24"} Apr 17 20:47:19.615556 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:19.615537 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-9rmnh_e562e939-99f0-4bee-952d-7af46132c860/console-operator/2.log" Apr 17 20:47:19.615686 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:19.615650 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" event={"ID":"e562e939-99f0-4bee-952d-7af46132c860","Type":"ContainerStarted","Data":"1d3fbae5ba90a7f266768cceb958a60519c187ae4dab18eb2cde88986df5c00d"} Apr 17 20:47:19.615937 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:19.615916 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" Apr 17 20:47:19.617376 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:19.617357 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-bc6wf" event={"ID":"821ba3d4-448f-45b9-bcec-701682592a0f","Type":"ContainerStarted","Data":"19640c137cad68e78919bf44fa30c4708268c0df94a44402c794db51a2d3293d"} Apr 17 20:47:19.617474 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:19.617381 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-bc6wf" event={"ID":"821ba3d4-448f-45b9-bcec-701682592a0f","Type":"ContainerStarted","Data":"8b14f15db4cf27e36d9b24b1cc1e7ad9c49a90dda0d8f3cc898bcb0c149830e4"} Apr 17 20:47:19.617545 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:19.617533 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-dns/dns-default-bc6wf" Apr 17 20:47:19.620915 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:19.620893 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" Apr 17 20:47:19.627687 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:19.627648 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-q59g7" podStartSLOduration=129.595981879 podStartE2EDuration="2m11.627635127s" podCreationTimestamp="2026-04-17 20:45:08 +0000 UTC" firstStartedPulling="2026-04-17 20:47:16.524938828 +0000 UTC m=+161.998717092" lastFinishedPulling="2026-04-17 20:47:18.556592061 +0000 UTC m=+164.030370340" observedRunningTime="2026-04-17 20:47:19.626369868 +0000 UTC m=+165.100148155" watchObservedRunningTime="2026-04-17 20:47:19.627635127 +0000 UTC m=+165.101413415" Apr 17 20:47:19.640557 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:19.640517 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-bc6wf" podStartSLOduration=129.624942332 podStartE2EDuration="2m11.640506624s" podCreationTimestamp="2026-04-17 20:45:08 +0000 UTC" firstStartedPulling="2026-04-17 20:47:16.538693239 +0000 UTC m=+162.012471504" lastFinishedPulling="2026-04-17 20:47:18.554257531 +0000 UTC m=+164.028035796" observedRunningTime="2026-04-17 20:47:19.639240983 +0000 UTC m=+165.113019270" watchObservedRunningTime="2026-04-17 20:47:19.640506624 +0000 UTC m=+165.114284911" Apr 17 20:47:19.653645 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:19.653609 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-9d4b6777b-9rmnh" podStartSLOduration=56.181472442 podStartE2EDuration="59.653593132s" podCreationTimestamp="2026-04-17 20:46:20 +0000 UTC" firstStartedPulling="2026-04-17 20:46:21.093358251 +0000 UTC m=+106.567136516" lastFinishedPulling="2026-04-17 20:46:24.56547894 +0000 UTC m=+110.039257206" observedRunningTime="2026-04-17 20:47:19.652704818 +0000 UTC m=+165.126483105" watchObservedRunningTime="2026-04-17 20:47:19.653593132 +0000 UTC m=+165.127371422" Apr 17 20:47:27.096623 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:27.096542 2577 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" Apr 17 20:47:27.096623 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:27.096580 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" Apr 17 20:47:29.622717 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:29.622689 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-bc6wf" Apr 17 20:47:36.670686 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:36.670649 2577 generic.go:358] "Generic (PLEG): container finished" podID="f1981af6-ac30-4f6a-b971-bb5e2e4f8b1e" containerID="dbfffb563d8a142bdf535ab5b834de395ded0178d8c9f349295907b6d10ba249" exitCode=0 Apr 17 20:47:36.671079 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:36.670730 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-rxj8l" event={"ID":"f1981af6-ac30-4f6a-b971-bb5e2e4f8b1e","Type":"ContainerDied","Data":"dbfffb563d8a142bdf535ab5b834de395ded0178d8c9f349295907b6d10ba249"} Apr 17 20:47:36.671079 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:36.671057 2577 scope.go:117] "RemoveContainer" containerID="dbfffb563d8a142bdf535ab5b834de395ded0178d8c9f349295907b6d10ba249" Apr 17 20:47:37.674886 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:37.674851 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-d6fc45fc5-rxj8l" event={"ID":"f1981af6-ac30-4f6a-b971-bb5e2e4f8b1e","Type":"ContainerStarted","Data":"fefde5427ab9d98521115ee03b1fda87200428be3b40a5ca52a3e8b6d2112f5b"} Apr 17 20:47:45.695998 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:45.695965 2577 generic.go:358] "Generic (PLEG): container finished" podID="05c5418a-43f4-476b-8e92-b7b56513d5af" containerID="4ff09f85efdf190efc51e4034ea41febf5532eaa4dd2f653ea6b0df5d2253af7" exitCode=0 Apr 17 20:47:45.696358 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:45.696010 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-fs4ts" event={"ID":"05c5418a-43f4-476b-8e92-b7b56513d5af","Type":"ContainerDied","Data":"4ff09f85efdf190efc51e4034ea41febf5532eaa4dd2f653ea6b0df5d2253af7"} Apr 17 20:47:45.696358 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:45.696280 2577 scope.go:117] "RemoveContainer" containerID="4ff09f85efdf190efc51e4034ea41febf5532eaa4dd2f653ea6b0df5d2253af7" Apr 17 20:47:46.705033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:46.704990 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-6769c5d45-fs4ts" event={"ID":"05c5418a-43f4-476b-8e92-b7b56513d5af","Type":"ContainerStarted","Data":"9bfe15103a0571c8d387b643a2a74282e2a59ce66ffc210ec87b0161bb2dd1ab"} Apr 17 20:47:47.102221 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:47.102141 2577 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" Apr 17 20:47:47.106368 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:47:47.106342 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/metrics-server-5c998b9cc6-kqcc5" Apr 17 20:48:00.747001 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:00.746966 2577 generic.go:358] "Generic (PLEG): container finished" podID="18ff00e6-c4f7-4852-b689-94ec992eb400" containerID="68bc86d5b22339444a41cf053918b6e5a22adaeafd4d59636a98e77c5f5b3ff5" exitCode=0 Apr 17 20:48:00.747522 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:00.747048 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-operator-585dfdc468-5jzh4" event={"ID":"18ff00e6-c4f7-4852-b689-94ec992eb400","Type":"ContainerDied","Data":"68bc86d5b22339444a41cf053918b6e5a22adaeafd4d59636a98e77c5f5b3ff5"} Apr 17 20:48:00.747522 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:00.747498 2577 scope.go:117] "RemoveContainer" containerID="68bc86d5b22339444a41cf053918b6e5a22adaeafd4d59636a98e77c5f5b3ff5" Apr 17 20:48:01.751786 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:01.751750 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-operator-585dfdc468-5jzh4" event={"ID":"18ff00e6-c4f7-4852-b689-94ec992eb400","Type":"ContainerStarted","Data":"540a7dc9ce26123a0d5066754d36424a2e78091f7f654a743152237fdea1e9ad"} Apr 17 20:48:08.807124 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:08.807082 2577 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:08.828421 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:08.828393 2577 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:09.790026 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:09.790000 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:26.811750 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:26.811704 2577 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Apr 17 20:48:26.812367 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:26.812204 2577 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-0" podUID="2505b439-b489-4c27-a41c-3cacc6db2776" containerName="prometheus" containerID="cri-o://1eea24c6739c5904842a5d0c0c5a820e903433aaf22a2642e06fd8995c1dcd1a" gracePeriod=600 Apr 17 20:48:26.812367 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:26.812222 2577 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-0" podUID="2505b439-b489-4c27-a41c-3cacc6db2776" containerName="kube-rbac-proxy" containerID="cri-o://73bb05bea309ada12836a23f856d5e6dd52c4fb597ea35970d2092c727d510d5" gracePeriod=600 Apr 17 20:48:26.812367 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:26.812247 2577 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-0" podUID="2505b439-b489-4c27-a41c-3cacc6db2776" containerName="thanos-sidecar" containerID="cri-o://7a699d681b62bc5e090c3b6d9f919fa04f05e19dcbfa6f9ae31d251d3ec3bf77" gracePeriod=600 Apr 17 20:48:26.812367 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:26.812267 2577 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-0" podUID="2505b439-b489-4c27-a41c-3cacc6db2776" containerName="kube-rbac-proxy-web" containerID="cri-o://2bc856f7e8d2ceacf86758e494ef3f3fccf228327419e5bc907a74bc455bbad8" gracePeriod=600 Apr 17 20:48:26.812367 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:26.812295 2577 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-0" podUID="2505b439-b489-4c27-a41c-3cacc6db2776" containerName="kube-rbac-proxy-thanos" containerID="cri-o://f9df59a5ded69de59ac6100a894265062e32e38a0a3bd44901cbbe4c535f07b5" gracePeriod=600 Apr 17 20:48:26.812367 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:26.812305 2577 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-0" podUID="2505b439-b489-4c27-a41c-3cacc6db2776" containerName="config-reloader" containerID="cri-o://a7db2bf1e97eae79efc8f9a1887ab095b2d2ad4cd5aaec2f7795fa8b4f4934db" gracePeriod=600 Apr 17 20:48:27.843471 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:27.843440 2577 generic.go:358] "Generic (PLEG): container finished" podID="2505b439-b489-4c27-a41c-3cacc6db2776" containerID="f9df59a5ded69de59ac6100a894265062e32e38a0a3bd44901cbbe4c535f07b5" exitCode=0 Apr 17 20:48:27.843471 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:27.843465 2577 generic.go:358] "Generic (PLEG): container finished" podID="2505b439-b489-4c27-a41c-3cacc6db2776" containerID="73bb05bea309ada12836a23f856d5e6dd52c4fb597ea35970d2092c727d510d5" exitCode=0 Apr 17 20:48:27.843471 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:27.843472 2577 generic.go:358] "Generic (PLEG): container finished" podID="2505b439-b489-4c27-a41c-3cacc6db2776" containerID="7a699d681b62bc5e090c3b6d9f919fa04f05e19dcbfa6f9ae31d251d3ec3bf77" exitCode=0 Apr 17 20:48:27.843471 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:27.843478 2577 generic.go:358] "Generic (PLEG): container finished" podID="2505b439-b489-4c27-a41c-3cacc6db2776" containerID="a7db2bf1e97eae79efc8f9a1887ab095b2d2ad4cd5aaec2f7795fa8b4f4934db" exitCode=0 Apr 17 20:48:27.843937 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:27.843483 2577 generic.go:358] "Generic (PLEG): container finished" podID="2505b439-b489-4c27-a41c-3cacc6db2776" containerID="1eea24c6739c5904842a5d0c0c5a820e903433aaf22a2642e06fd8995c1dcd1a" exitCode=0 Apr 17 20:48:27.843937 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:27.843509 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"2505b439-b489-4c27-a41c-3cacc6db2776","Type":"ContainerDied","Data":"f9df59a5ded69de59ac6100a894265062e32e38a0a3bd44901cbbe4c535f07b5"} Apr 17 20:48:27.843937 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:27.843539 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"2505b439-b489-4c27-a41c-3cacc6db2776","Type":"ContainerDied","Data":"73bb05bea309ada12836a23f856d5e6dd52c4fb597ea35970d2092c727d510d5"} Apr 17 20:48:27.843937 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:27.843550 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"2505b439-b489-4c27-a41c-3cacc6db2776","Type":"ContainerDied","Data":"7a699d681b62bc5e090c3b6d9f919fa04f05e19dcbfa6f9ae31d251d3ec3bf77"} Apr 17 20:48:27.843937 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:27.843559 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"2505b439-b489-4c27-a41c-3cacc6db2776","Type":"ContainerDied","Data":"a7db2bf1e97eae79efc8f9a1887ab095b2d2ad4cd5aaec2f7795fa8b4f4934db"} Apr 17 20:48:27.843937 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:27.843567 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"2505b439-b489-4c27-a41c-3cacc6db2776","Type":"ContainerDied","Data":"1eea24c6739c5904842a5d0c0c5a820e903433aaf22a2642e06fd8995c1dcd1a"} Apr 17 20:48:28.152746 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.152724 2577 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:28.294011 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.293983 2577 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-config\") pod \"2505b439-b489-4c27-a41c-3cacc6db2776\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " Apr 17 20:48:28.294168 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.294032 2577 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-configmap-kubelet-serving-ca-bundle\") pod \"2505b439-b489-4c27-a41c-3cacc6db2776\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " Apr 17 20:48:28.294168 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.294058 2577 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/2505b439-b489-4c27-a41c-3cacc6db2776-prometheus-k8s-db\") pod \"2505b439-b489-4c27-a41c-3cacc6db2776\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " Apr 17 20:48:28.294168 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.294080 2577 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-prometheus-k8s-tls\") pod \"2505b439-b489-4c27-a41c-3cacc6db2776\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " Apr 17 20:48:28.294330 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.294191 2577 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cb5vb\" (UniqueName: \"kubernetes.io/projected/2505b439-b489-4c27-a41c-3cacc6db2776-kube-api-access-cb5vb\") pod \"2505b439-b489-4c27-a41c-3cacc6db2776\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " Apr 17 20:48:28.294330 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.294226 2577 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-prometheus-k8s-rulefiles-0\") pod \"2505b439-b489-4c27-a41c-3cacc6db2776\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " Apr 17 20:48:28.294330 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.294254 2577 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-kube-rbac-proxy\") pod \"2505b439-b489-4c27-a41c-3cacc6db2776\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " Apr 17 20:48:28.294330 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.294297 2577 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-configmap-metrics-client-ca\") pod \"2505b439-b489-4c27-a41c-3cacc6db2776\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " Apr 17 20:48:28.294515 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.294343 2577 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"2505b439-b489-4c27-a41c-3cacc6db2776\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " Apr 17 20:48:28.294515 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.294373 2577 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-metrics-client-certs\") pod \"2505b439-b489-4c27-a41c-3cacc6db2776\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " Apr 17 20:48:28.294515 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.294403 2577 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2505b439-b489-4c27-a41c-3cacc6db2776-config-out\") pod \"2505b439-b489-4c27-a41c-3cacc6db2776\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " Apr 17 20:48:28.294515 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.294439 2577 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-grpc-tls\") pod \"2505b439-b489-4c27-a41c-3cacc6db2776\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " Apr 17 20:48:28.294515 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.294473 2577 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"2505b439-b489-4c27-a41c-3cacc6db2776\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " Apr 17 20:48:28.294515 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.294503 2577 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-thanos-prometheus-http-client-file\") pod \"2505b439-b489-4c27-a41c-3cacc6db2776\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " Apr 17 20:48:28.294792 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.294532 2577 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2505b439-b489-4c27-a41c-3cacc6db2776-tls-assets\") pod \"2505b439-b489-4c27-a41c-3cacc6db2776\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " Apr 17 20:48:28.294792 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.294572 2577 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-prometheus-trusted-ca-bundle\") pod \"2505b439-b489-4c27-a41c-3cacc6db2776\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " Apr 17 20:48:28.294792 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.294576 2577 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-configmap-kubelet-serving-ca-bundle" (OuterVolumeSpecName: "configmap-kubelet-serving-ca-bundle") pod "2505b439-b489-4c27-a41c-3cacc6db2776" (UID: "2505b439-b489-4c27-a41c-3cacc6db2776"). InnerVolumeSpecName "configmap-kubelet-serving-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 17 20:48:28.294792 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.294602 2577 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-web-config\") pod \"2505b439-b489-4c27-a41c-3cacc6db2776\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " Apr 17 20:48:28.294792 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.294628 2577 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-configmap-serving-certs-ca-bundle\") pod \"2505b439-b489-4c27-a41c-3cacc6db2776\" (UID: \"2505b439-b489-4c27-a41c-3cacc6db2776\") " Apr 17 20:48:28.295106 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.294911 2577 reconciler_common.go:299] "Volume detached for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-configmap-kubelet-serving-ca-bundle\") on node \"ip-10-0-137-110.ec2.internal\" DevicePath \"\"" Apr 17 20:48:28.295707 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.295444 2577 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-configmap-serving-certs-ca-bundle" (OuterVolumeSpecName: "configmap-serving-certs-ca-bundle") pod "2505b439-b489-4c27-a41c-3cacc6db2776" (UID: "2505b439-b489-4c27-a41c-3cacc6db2776"). InnerVolumeSpecName "configmap-serving-certs-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 17 20:48:28.295707 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.295527 2577 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2505b439-b489-4c27-a41c-3cacc6db2776-prometheus-k8s-db" (OuterVolumeSpecName: "prometheus-k8s-db") pod "2505b439-b489-4c27-a41c-3cacc6db2776" (UID: "2505b439-b489-4c27-a41c-3cacc6db2776"). InnerVolumeSpecName "prometheus-k8s-db". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 17 20:48:28.296124 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.296096 2577 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-prometheus-trusted-ca-bundle" (OuterVolumeSpecName: "prometheus-trusted-ca-bundle") pod "2505b439-b489-4c27-a41c-3cacc6db2776" (UID: "2505b439-b489-4c27-a41c-3cacc6db2776"). InnerVolumeSpecName "prometheus-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 17 20:48:28.296196 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.296129 2577 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-configmap-metrics-client-ca" (OuterVolumeSpecName: "configmap-metrics-client-ca") pod "2505b439-b489-4c27-a41c-3cacc6db2776" (UID: "2505b439-b489-4c27-a41c-3cacc6db2776"). InnerVolumeSpecName "configmap-metrics-client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 17 20:48:28.297258 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.297192 2577 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-config" (OuterVolumeSpecName: "config") pod "2505b439-b489-4c27-a41c-3cacc6db2776" (UID: "2505b439-b489-4c27-a41c-3cacc6db2776"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 17 20:48:28.297517 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.297487 2577 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-prometheus-k8s-rulefiles-0" (OuterVolumeSpecName: "prometheus-k8s-rulefiles-0") pod "2505b439-b489-4c27-a41c-3cacc6db2776" (UID: "2505b439-b489-4c27-a41c-3cacc6db2776"). InnerVolumeSpecName "prometheus-k8s-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 17 20:48:28.297782 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.297707 2577 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-grpc-tls" (OuterVolumeSpecName: "secret-grpc-tls") pod "2505b439-b489-4c27-a41c-3cacc6db2776" (UID: "2505b439-b489-4c27-a41c-3cacc6db2776"). InnerVolumeSpecName "secret-grpc-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 17 20:48:28.299272 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.299153 2577 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2505b439-b489-4c27-a41c-3cacc6db2776-kube-api-access-cb5vb" (OuterVolumeSpecName: "kube-api-access-cb5vb") pod "2505b439-b489-4c27-a41c-3cacc6db2776" (UID: "2505b439-b489-4c27-a41c-3cacc6db2776"). InnerVolumeSpecName "kube-api-access-cb5vb". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 17 20:48:28.299272 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.299232 2577 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-prometheus-k8s-thanos-sidecar-tls" (OuterVolumeSpecName: "secret-prometheus-k8s-thanos-sidecar-tls") pod "2505b439-b489-4c27-a41c-3cacc6db2776" (UID: "2505b439-b489-4c27-a41c-3cacc6db2776"). InnerVolumeSpecName "secret-prometheus-k8s-thanos-sidecar-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 17 20:48:28.299430 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.299359 2577 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-prometheus-k8s-tls" (OuterVolumeSpecName: "secret-prometheus-k8s-tls") pod "2505b439-b489-4c27-a41c-3cacc6db2776" (UID: "2505b439-b489-4c27-a41c-3cacc6db2776"). InnerVolumeSpecName "secret-prometheus-k8s-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 17 20:48:28.299539 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.299488 2577 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-kube-rbac-proxy" (OuterVolumeSpecName: "secret-kube-rbac-proxy") pod "2505b439-b489-4c27-a41c-3cacc6db2776" (UID: "2505b439-b489-4c27-a41c-3cacc6db2776"). InnerVolumeSpecName "secret-kube-rbac-proxy". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 17 20:48:28.299657 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.299588 2577 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-prometheus-k8s-kube-rbac-proxy-web" (OuterVolumeSpecName: "secret-prometheus-k8s-kube-rbac-proxy-web") pod "2505b439-b489-4c27-a41c-3cacc6db2776" (UID: "2505b439-b489-4c27-a41c-3cacc6db2776"). InnerVolumeSpecName "secret-prometheus-k8s-kube-rbac-proxy-web". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 17 20:48:28.300090 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.300063 2577 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2505b439-b489-4c27-a41c-3cacc6db2776-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "2505b439-b489-4c27-a41c-3cacc6db2776" (UID: "2505b439-b489-4c27-a41c-3cacc6db2776"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 17 20:48:28.300286 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.300248 2577 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2505b439-b489-4c27-a41c-3cacc6db2776-config-out" (OuterVolumeSpecName: "config-out") pod "2505b439-b489-4c27-a41c-3cacc6db2776" (UID: "2505b439-b489-4c27-a41c-3cacc6db2776"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Apr 17 20:48:28.300391 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.300320 2577 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-metrics-client-certs" (OuterVolumeSpecName: "secret-metrics-client-certs") pod "2505b439-b489-4c27-a41c-3cacc6db2776" (UID: "2505b439-b489-4c27-a41c-3cacc6db2776"). InnerVolumeSpecName "secret-metrics-client-certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 17 20:48:28.300841 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.300791 2577 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "2505b439-b489-4c27-a41c-3cacc6db2776" (UID: "2505b439-b489-4c27-a41c-3cacc6db2776"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 17 20:48:28.309721 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.309673 2577 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-web-config" (OuterVolumeSpecName: "web-config") pod "2505b439-b489-4c27-a41c-3cacc6db2776" (UID: "2505b439-b489-4c27-a41c-3cacc6db2776"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Apr 17 20:48:28.396282 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.396233 2577 reconciler_common.go:299] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2505b439-b489-4c27-a41c-3cacc6db2776-config-out\") on node \"ip-10-0-137-110.ec2.internal\" DevicePath \"\"" Apr 17 20:48:28.396282 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.396254 2577 reconciler_common.go:299] "Volume detached for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-grpc-tls\") on node \"ip-10-0-137-110.ec2.internal\" DevicePath \"\"" Apr 17 20:48:28.396282 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.396265 2577 reconciler_common.go:299] "Volume detached for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-prometheus-k8s-thanos-sidecar-tls\") on node \"ip-10-0-137-110.ec2.internal\" DevicePath \"\"" Apr 17 20:48:28.396282 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.396274 2577 reconciler_common.go:299] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-thanos-prometheus-http-client-file\") on node \"ip-10-0-137-110.ec2.internal\" DevicePath \"\"" Apr 17 20:48:28.396282 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.396283 2577 reconciler_common.go:299] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2505b439-b489-4c27-a41c-3cacc6db2776-tls-assets\") on node \"ip-10-0-137-110.ec2.internal\" DevicePath \"\"" Apr 17 20:48:28.396471 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.396294 2577 reconciler_common.go:299] "Volume detached for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-prometheus-trusted-ca-bundle\") on node \"ip-10-0-137-110.ec2.internal\" DevicePath \"\"" Apr 17 20:48:28.396471 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.396303 2577 reconciler_common.go:299] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-web-config\") on node \"ip-10-0-137-110.ec2.internal\" DevicePath \"\"" Apr 17 20:48:28.396471 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.396313 2577 reconciler_common.go:299] "Volume detached for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-configmap-serving-certs-ca-bundle\") on node \"ip-10-0-137-110.ec2.internal\" DevicePath \"\"" Apr 17 20:48:28.396471 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.396321 2577 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-config\") on node \"ip-10-0-137-110.ec2.internal\" DevicePath \"\"" Apr 17 20:48:28.396471 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.396329 2577 reconciler_common.go:299] "Volume detached for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/2505b439-b489-4c27-a41c-3cacc6db2776-prometheus-k8s-db\") on node \"ip-10-0-137-110.ec2.internal\" DevicePath \"\"" Apr 17 20:48:28.396471 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.396343 2577 reconciler_common.go:299] "Volume detached for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-prometheus-k8s-tls\") on node \"ip-10-0-137-110.ec2.internal\" DevicePath \"\"" Apr 17 20:48:28.396471 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.396352 2577 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-cb5vb\" (UniqueName: \"kubernetes.io/projected/2505b439-b489-4c27-a41c-3cacc6db2776-kube-api-access-cb5vb\") on node \"ip-10-0-137-110.ec2.internal\" DevicePath \"\"" Apr 17 20:48:28.396471 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.396360 2577 reconciler_common.go:299] "Volume detached for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-prometheus-k8s-rulefiles-0\") on node \"ip-10-0-137-110.ec2.internal\" DevicePath \"\"" Apr 17 20:48:28.396471 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.396369 2577 reconciler_common.go:299] "Volume detached for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-kube-rbac-proxy\") on node \"ip-10-0-137-110.ec2.internal\" DevicePath \"\"" Apr 17 20:48:28.396471 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.396377 2577 reconciler_common.go:299] "Volume detached for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2505b439-b489-4c27-a41c-3cacc6db2776-configmap-metrics-client-ca\") on node \"ip-10-0-137-110.ec2.internal\" DevicePath \"\"" Apr 17 20:48:28.396471 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.396385 2577 reconciler_common.go:299] "Volume detached for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-prometheus-k8s-kube-rbac-proxy-web\") on node \"ip-10-0-137-110.ec2.internal\" DevicePath \"\"" Apr 17 20:48:28.396471 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.396394 2577 reconciler_common.go:299] "Volume detached for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/2505b439-b489-4c27-a41c-3cacc6db2776-secret-metrics-client-certs\") on node \"ip-10-0-137-110.ec2.internal\" DevicePath \"\"" Apr 17 20:48:28.849648 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.849613 2577 generic.go:358] "Generic (PLEG): container finished" podID="2505b439-b489-4c27-a41c-3cacc6db2776" containerID="2bc856f7e8d2ceacf86758e494ef3f3fccf228327419e5bc907a74bc455bbad8" exitCode=0 Apr 17 20:48:28.850038 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.849667 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"2505b439-b489-4c27-a41c-3cacc6db2776","Type":"ContainerDied","Data":"2bc856f7e8d2ceacf86758e494ef3f3fccf228327419e5bc907a74bc455bbad8"} Apr 17 20:48:28.850038 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.849692 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"2505b439-b489-4c27-a41c-3cacc6db2776","Type":"ContainerDied","Data":"71254b6ec62c33bd4dd03ff557729030f823662f77e8da946574970e6bbc4526"} Apr 17 20:48:28.850038 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.849709 2577 scope.go:117] "RemoveContainer" containerID="f9df59a5ded69de59ac6100a894265062e32e38a0a3bd44901cbbe4c535f07b5" Apr 17 20:48:28.850038 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.849772 2577 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:28.862144 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.862127 2577 scope.go:117] "RemoveContainer" containerID="73bb05bea309ada12836a23f856d5e6dd52c4fb597ea35970d2092c727d510d5" Apr 17 20:48:28.869301 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.869278 2577 scope.go:117] "RemoveContainer" containerID="2bc856f7e8d2ceacf86758e494ef3f3fccf228327419e5bc907a74bc455bbad8" Apr 17 20:48:28.873645 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.873616 2577 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Apr 17 20:48:28.876544 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.876522 2577 scope.go:117] "RemoveContainer" containerID="7a699d681b62bc5e090c3b6d9f919fa04f05e19dcbfa6f9ae31d251d3ec3bf77" Apr 17 20:48:28.881265 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.881243 2577 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Apr 17 20:48:28.884048 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.884021 2577 scope.go:117] "RemoveContainer" containerID="a7db2bf1e97eae79efc8f9a1887ab095b2d2ad4cd5aaec2f7795fa8b4f4934db" Apr 17 20:48:28.890563 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.890543 2577 scope.go:117] "RemoveContainer" containerID="1eea24c6739c5904842a5d0c0c5a820e903433aaf22a2642e06fd8995c1dcd1a" Apr 17 20:48:28.897388 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.897372 2577 scope.go:117] "RemoveContainer" containerID="77783870ae1deea2166d5abe8ddfbea7b14497deb4bfc7b3e499413f6f7e1c4c" Apr 17 20:48:28.904073 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.904055 2577 scope.go:117] "RemoveContainer" containerID="f9df59a5ded69de59ac6100a894265062e32e38a0a3bd44901cbbe4c535f07b5" Apr 17 20:48:28.904385 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:48:28.904355 2577 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9df59a5ded69de59ac6100a894265062e32e38a0a3bd44901cbbe4c535f07b5\": container with ID starting with f9df59a5ded69de59ac6100a894265062e32e38a0a3bd44901cbbe4c535f07b5 not found: ID does not exist" containerID="f9df59a5ded69de59ac6100a894265062e32e38a0a3bd44901cbbe4c535f07b5" Apr 17 20:48:28.904477 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.904396 2577 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9df59a5ded69de59ac6100a894265062e32e38a0a3bd44901cbbe4c535f07b5"} err="failed to get container status \"f9df59a5ded69de59ac6100a894265062e32e38a0a3bd44901cbbe4c535f07b5\": rpc error: code = NotFound desc = could not find container \"f9df59a5ded69de59ac6100a894265062e32e38a0a3bd44901cbbe4c535f07b5\": container with ID starting with f9df59a5ded69de59ac6100a894265062e32e38a0a3bd44901cbbe4c535f07b5 not found: ID does not exist" Apr 17 20:48:28.904477 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.904439 2577 scope.go:117] "RemoveContainer" containerID="73bb05bea309ada12836a23f856d5e6dd52c4fb597ea35970d2092c727d510d5" Apr 17 20:48:28.904708 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.904689 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Apr 17 20:48:28.904764 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:48:28.904710 2577 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73bb05bea309ada12836a23f856d5e6dd52c4fb597ea35970d2092c727d510d5\": container with ID starting with 73bb05bea309ada12836a23f856d5e6dd52c4fb597ea35970d2092c727d510d5 not found: ID does not exist" containerID="73bb05bea309ada12836a23f856d5e6dd52c4fb597ea35970d2092c727d510d5" Apr 17 20:48:28.904764 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.904730 2577 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73bb05bea309ada12836a23f856d5e6dd52c4fb597ea35970d2092c727d510d5"} err="failed to get container status \"73bb05bea309ada12836a23f856d5e6dd52c4fb597ea35970d2092c727d510d5\": rpc error: code = NotFound desc = could not find container \"73bb05bea309ada12836a23f856d5e6dd52c4fb597ea35970d2092c727d510d5\": container with ID starting with 73bb05bea309ada12836a23f856d5e6dd52c4fb597ea35970d2092c727d510d5 not found: ID does not exist" Apr 17 20:48:28.904764 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.904746 2577 scope.go:117] "RemoveContainer" containerID="2bc856f7e8d2ceacf86758e494ef3f3fccf228327419e5bc907a74bc455bbad8" Apr 17 20:48:28.905009 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:48:28.904992 2577 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bc856f7e8d2ceacf86758e494ef3f3fccf228327419e5bc907a74bc455bbad8\": container with ID starting with 2bc856f7e8d2ceacf86758e494ef3f3fccf228327419e5bc907a74bc455bbad8 not found: ID does not exist" containerID="2bc856f7e8d2ceacf86758e494ef3f3fccf228327419e5bc907a74bc455bbad8" Apr 17 20:48:28.905054 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905013 2577 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bc856f7e8d2ceacf86758e494ef3f3fccf228327419e5bc907a74bc455bbad8"} err="failed to get container status \"2bc856f7e8d2ceacf86758e494ef3f3fccf228327419e5bc907a74bc455bbad8\": rpc error: code = NotFound desc = could not find container \"2bc856f7e8d2ceacf86758e494ef3f3fccf228327419e5bc907a74bc455bbad8\": container with ID starting with 2bc856f7e8d2ceacf86758e494ef3f3fccf228327419e5bc907a74bc455bbad8 not found: ID does not exist" Apr 17 20:48:28.905054 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905027 2577 scope.go:117] "RemoveContainer" containerID="7a699d681b62bc5e090c3b6d9f919fa04f05e19dcbfa6f9ae31d251d3ec3bf77" Apr 17 20:48:28.905054 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905038 2577 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="2505b439-b489-4c27-a41c-3cacc6db2776" containerName="config-reloader" Apr 17 20:48:28.905054 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905052 2577 state_mem.go:107] "Deleted CPUSet assignment" podUID="2505b439-b489-4c27-a41c-3cacc6db2776" containerName="config-reloader" Apr 17 20:48:28.905198 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905060 2577 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="2505b439-b489-4c27-a41c-3cacc6db2776" containerName="kube-rbac-proxy" Apr 17 20:48:28.905198 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905065 2577 state_mem.go:107] "Deleted CPUSet assignment" podUID="2505b439-b489-4c27-a41c-3cacc6db2776" containerName="kube-rbac-proxy" Apr 17 20:48:28.905198 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905072 2577 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="2505b439-b489-4c27-a41c-3cacc6db2776" containerName="init-config-reloader" Apr 17 20:48:28.905198 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905077 2577 state_mem.go:107] "Deleted CPUSet assignment" podUID="2505b439-b489-4c27-a41c-3cacc6db2776" containerName="init-config-reloader" Apr 17 20:48:28.905198 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905085 2577 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="2505b439-b489-4c27-a41c-3cacc6db2776" containerName="prometheus" Apr 17 20:48:28.905198 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905093 2577 state_mem.go:107] "Deleted CPUSet assignment" podUID="2505b439-b489-4c27-a41c-3cacc6db2776" containerName="prometheus" Apr 17 20:48:28.905198 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905102 2577 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="2505b439-b489-4c27-a41c-3cacc6db2776" containerName="kube-rbac-proxy-thanos" Apr 17 20:48:28.905198 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905107 2577 state_mem.go:107] "Deleted CPUSet assignment" podUID="2505b439-b489-4c27-a41c-3cacc6db2776" containerName="kube-rbac-proxy-thanos" Apr 17 20:48:28.905198 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905120 2577 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="2505b439-b489-4c27-a41c-3cacc6db2776" containerName="thanos-sidecar" Apr 17 20:48:28.905198 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905127 2577 state_mem.go:107] "Deleted CPUSet assignment" podUID="2505b439-b489-4c27-a41c-3cacc6db2776" containerName="thanos-sidecar" Apr 17 20:48:28.905198 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905155 2577 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="2505b439-b489-4c27-a41c-3cacc6db2776" containerName="kube-rbac-proxy-web" Apr 17 20:48:28.905198 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905164 2577 state_mem.go:107] "Deleted CPUSet assignment" podUID="2505b439-b489-4c27-a41c-3cacc6db2776" containerName="kube-rbac-proxy-web" Apr 17 20:48:28.905638 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905236 2577 memory_manager.go:356] "RemoveStaleState removing state" podUID="2505b439-b489-4c27-a41c-3cacc6db2776" containerName="config-reloader" Apr 17 20:48:28.905638 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905248 2577 memory_manager.go:356] "RemoveStaleState removing state" podUID="2505b439-b489-4c27-a41c-3cacc6db2776" containerName="kube-rbac-proxy-web" Apr 17 20:48:28.905638 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:48:28.905246 2577 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a699d681b62bc5e090c3b6d9f919fa04f05e19dcbfa6f9ae31d251d3ec3bf77\": container with ID starting with 7a699d681b62bc5e090c3b6d9f919fa04f05e19dcbfa6f9ae31d251d3ec3bf77 not found: ID does not exist" containerID="7a699d681b62bc5e090c3b6d9f919fa04f05e19dcbfa6f9ae31d251d3ec3bf77" Apr 17 20:48:28.905638 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905274 2577 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a699d681b62bc5e090c3b6d9f919fa04f05e19dcbfa6f9ae31d251d3ec3bf77"} err="failed to get container status \"7a699d681b62bc5e090c3b6d9f919fa04f05e19dcbfa6f9ae31d251d3ec3bf77\": rpc error: code = NotFound desc = could not find container \"7a699d681b62bc5e090c3b6d9f919fa04f05e19dcbfa6f9ae31d251d3ec3bf77\": container with ID starting with 7a699d681b62bc5e090c3b6d9f919fa04f05e19dcbfa6f9ae31d251d3ec3bf77 not found: ID does not exist" Apr 17 20:48:28.905638 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905292 2577 scope.go:117] "RemoveContainer" containerID="a7db2bf1e97eae79efc8f9a1887ab095b2d2ad4cd5aaec2f7795fa8b4f4934db" Apr 17 20:48:28.905638 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905257 2577 memory_manager.go:356] "RemoveStaleState removing state" podUID="2505b439-b489-4c27-a41c-3cacc6db2776" containerName="thanos-sidecar" Apr 17 20:48:28.905638 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905345 2577 memory_manager.go:356] "RemoveStaleState removing state" podUID="2505b439-b489-4c27-a41c-3cacc6db2776" containerName="prometheus" Apr 17 20:48:28.905638 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905352 2577 memory_manager.go:356] "RemoveStaleState removing state" podUID="2505b439-b489-4c27-a41c-3cacc6db2776" containerName="kube-rbac-proxy" Apr 17 20:48:28.905638 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905360 2577 memory_manager.go:356] "RemoveStaleState removing state" podUID="2505b439-b489-4c27-a41c-3cacc6db2776" containerName="kube-rbac-proxy-thanos" Apr 17 20:48:28.905638 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:48:28.905555 2577 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7db2bf1e97eae79efc8f9a1887ab095b2d2ad4cd5aaec2f7795fa8b4f4934db\": container with ID starting with a7db2bf1e97eae79efc8f9a1887ab095b2d2ad4cd5aaec2f7795fa8b4f4934db not found: ID does not exist" containerID="a7db2bf1e97eae79efc8f9a1887ab095b2d2ad4cd5aaec2f7795fa8b4f4934db" Apr 17 20:48:28.905638 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905578 2577 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7db2bf1e97eae79efc8f9a1887ab095b2d2ad4cd5aaec2f7795fa8b4f4934db"} err="failed to get container status \"a7db2bf1e97eae79efc8f9a1887ab095b2d2ad4cd5aaec2f7795fa8b4f4934db\": rpc error: code = NotFound desc = could not find container \"a7db2bf1e97eae79efc8f9a1887ab095b2d2ad4cd5aaec2f7795fa8b4f4934db\": container with ID starting with a7db2bf1e97eae79efc8f9a1887ab095b2d2ad4cd5aaec2f7795fa8b4f4934db not found: ID does not exist" Apr 17 20:48:28.905638 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905596 2577 scope.go:117] "RemoveContainer" containerID="1eea24c6739c5904842a5d0c0c5a820e903433aaf22a2642e06fd8995c1dcd1a" Apr 17 20:48:28.906173 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:48:28.905851 2577 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1eea24c6739c5904842a5d0c0c5a820e903433aaf22a2642e06fd8995c1dcd1a\": container with ID starting with 1eea24c6739c5904842a5d0c0c5a820e903433aaf22a2642e06fd8995c1dcd1a not found: ID does not exist" containerID="1eea24c6739c5904842a5d0c0c5a820e903433aaf22a2642e06fd8995c1dcd1a" Apr 17 20:48:28.906173 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905871 2577 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1eea24c6739c5904842a5d0c0c5a820e903433aaf22a2642e06fd8995c1dcd1a"} err="failed to get container status \"1eea24c6739c5904842a5d0c0c5a820e903433aaf22a2642e06fd8995c1dcd1a\": rpc error: code = NotFound desc = could not find container \"1eea24c6739c5904842a5d0c0c5a820e903433aaf22a2642e06fd8995c1dcd1a\": container with ID starting with 1eea24c6739c5904842a5d0c0c5a820e903433aaf22a2642e06fd8995c1dcd1a not found: ID does not exist" Apr 17 20:48:28.906173 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.905884 2577 scope.go:117] "RemoveContainer" containerID="77783870ae1deea2166d5abe8ddfbea7b14497deb4bfc7b3e499413f6f7e1c4c" Apr 17 20:48:28.906173 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:48:28.906115 2577 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77783870ae1deea2166d5abe8ddfbea7b14497deb4bfc7b3e499413f6f7e1c4c\": container with ID starting with 77783870ae1deea2166d5abe8ddfbea7b14497deb4bfc7b3e499413f6f7e1c4c not found: ID does not exist" containerID="77783870ae1deea2166d5abe8ddfbea7b14497deb4bfc7b3e499413f6f7e1c4c" Apr 17 20:48:28.906173 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.906131 2577 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77783870ae1deea2166d5abe8ddfbea7b14497deb4bfc7b3e499413f6f7e1c4c"} err="failed to get container status \"77783870ae1deea2166d5abe8ddfbea7b14497deb4bfc7b3e499413f6f7e1c4c\": rpc error: code = NotFound desc = could not find container \"77783870ae1deea2166d5abe8ddfbea7b14497deb4bfc7b3e499413f6f7e1c4c\": container with ID starting with 77783870ae1deea2166d5abe8ddfbea7b14497deb4bfc7b3e499413f6f7e1c4c not found: ID does not exist" Apr 17 20:48:28.910601 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.910585 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:28.912280 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.912262 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-tls-assets-0\"" Apr 17 20:48:28.912416 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.912401 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-tls\"" Apr 17 20:48:28.912479 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.912471 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-client-certs\"" Apr 17 20:48:28.912652 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.912638 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-kube-rbac-proxy-web\"" Apr 17 20:48:28.912696 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.912642 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-dockercfg-kwhx4\"" Apr 17 20:48:28.912780 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.912759 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"serving-certs-ca-bundle\"" Apr 17 20:48:28.912864 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.912847 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-thanos-prometheus-http-client-file\"" Apr 17 20:48:28.913365 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.913343 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-grpc-tls-9hk0s6bkcm5i8\"" Apr 17 20:48:28.913469 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.913343 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-web-config\"" Apr 17 20:48:28.913469 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.913349 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-thanos-sidecar-tls\"" Apr 17 20:48:28.913575 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.913480 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s\"" Apr 17 20:48:28.913575 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.913558 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"kube-rbac-proxy\"" Apr 17 20:48:28.915958 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.915940 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-rulefiles-0\"" Apr 17 20:48:28.919227 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.919208 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"prometheus-trusted-ca-bundle\"" Apr 17 20:48:28.919998 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.919974 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Apr 17 20:48:29.000010 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:28.999988 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/356df3f8-dff9-4963-a883-fe17c0c75d95-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.000097 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.000016 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/356df3f8-dff9-4963-a883-fe17c0c75d95-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.000097 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.000035 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/356df3f8-dff9-4963-a883-fe17c0c75d95-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.000097 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.000058 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/356df3f8-dff9-4963-a883-fe17c0c75d95-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.000251 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.000132 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/356df3f8-dff9-4963-a883-fe17c0c75d95-web-config\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.000251 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.000163 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/356df3f8-dff9-4963-a883-fe17c0c75d95-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.000251 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.000199 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/356df3f8-dff9-4963-a883-fe17c0c75d95-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.000251 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.000246 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/356df3f8-dff9-4963-a883-fe17c0c75d95-config-out\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.000361 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.000276 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/356df3f8-dff9-4963-a883-fe17c0c75d95-config\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.000361 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.000301 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/356df3f8-dff9-4963-a883-fe17c0c75d95-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.000361 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.000321 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/356df3f8-dff9-4963-a883-fe17c0c75d95-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.000450 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.000367 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/356df3f8-dff9-4963-a883-fe17c0c75d95-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.000450 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.000402 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/356df3f8-dff9-4963-a883-fe17c0c75d95-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.000450 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.000421 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/356df3f8-dff9-4963-a883-fe17c0c75d95-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.000450 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.000442 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/356df3f8-dff9-4963-a883-fe17c0c75d95-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.000593 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.000462 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/356df3f8-dff9-4963-a883-fe17c0c75d95-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.000593 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.000506 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/356df3f8-dff9-4963-a883-fe17c0c75d95-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.000593 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.000528 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5q8l\" (UniqueName: \"kubernetes.io/projected/356df3f8-dff9-4963-a883-fe17c0c75d95-kube-api-access-d5q8l\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.101607 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.101533 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/356df3f8-dff9-4963-a883-fe17c0c75d95-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.101607 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.101563 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/356df3f8-dff9-4963-a883-fe17c0c75d95-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.101607 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.101583 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/356df3f8-dff9-4963-a883-fe17c0c75d95-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.101875 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.101717 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-d5q8l\" (UniqueName: \"kubernetes.io/projected/356df3f8-dff9-4963-a883-fe17c0c75d95-kube-api-access-d5q8l\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.101875 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.101759 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/356df3f8-dff9-4963-a883-fe17c0c75d95-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.101875 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.101782 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/356df3f8-dff9-4963-a883-fe17c0c75d95-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.102032 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.101991 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/356df3f8-dff9-4963-a883-fe17c0c75d95-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.102087 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.102042 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/356df3f8-dff9-4963-a883-fe17c0c75d95-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.102136 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.102084 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/356df3f8-dff9-4963-a883-fe17c0c75d95-web-config\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.102136 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.102109 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/356df3f8-dff9-4963-a883-fe17c0c75d95-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.102243 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.102145 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/356df3f8-dff9-4963-a883-fe17c0c75d95-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.102243 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.102172 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/356df3f8-dff9-4963-a883-fe17c0c75d95-config-out\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.102243 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.102200 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/356df3f8-dff9-4963-a883-fe17c0c75d95-config\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.102243 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.102225 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/356df3f8-dff9-4963-a883-fe17c0c75d95-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.102420 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.102252 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/356df3f8-dff9-4963-a883-fe17c0c75d95-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.102420 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.102284 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/356df3f8-dff9-4963-a883-fe17c0c75d95-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.102420 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.102314 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/356df3f8-dff9-4963-a883-fe17c0c75d95-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.102420 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.102338 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/356df3f8-dff9-4963-a883-fe17c0c75d95-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.102608 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.102456 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/356df3f8-dff9-4963-a883-fe17c0c75d95-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.102831 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.102784 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/356df3f8-dff9-4963-a883-fe17c0c75d95-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.104569 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.104403 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/356df3f8-dff9-4963-a883-fe17c0c75d95-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.105224 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.105191 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/356df3f8-dff9-4963-a883-fe17c0c75d95-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.105335 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.105280 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/356df3f8-dff9-4963-a883-fe17c0c75d95-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.105931 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.105899 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/356df3f8-dff9-4963-a883-fe17c0c75d95-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.106035 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.105933 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/356df3f8-dff9-4963-a883-fe17c0c75d95-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.106035 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.105906 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/356df3f8-dff9-4963-a883-fe17c0c75d95-web-config\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.106161 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.106076 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/356df3f8-dff9-4963-a883-fe17c0c75d95-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.106501 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.106475 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/356df3f8-dff9-4963-a883-fe17c0c75d95-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.106626 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.106607 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/356df3f8-dff9-4963-a883-fe17c0c75d95-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.106727 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.106703 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/356df3f8-dff9-4963-a883-fe17c0c75d95-config\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.107712 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.107687 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/356df3f8-dff9-4963-a883-fe17c0c75d95-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.107927 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.107898 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/356df3f8-dff9-4963-a883-fe17c0c75d95-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.107927 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.107905 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/356df3f8-dff9-4963-a883-fe17c0c75d95-config-out\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.107927 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.107914 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/356df3f8-dff9-4963-a883-fe17c0c75d95-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.108342 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.108321 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/356df3f8-dff9-4963-a883-fe17c0c75d95-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.109102 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.109074 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5q8l\" (UniqueName: \"kubernetes.io/projected/356df3f8-dff9-4963-a883-fe17c0c75d95-kube-api-access-d5q8l\") pod \"prometheus-k8s-0\" (UID: \"356df3f8-dff9-4963-a883-fe17c0c75d95\") " pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.120331 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.120305 2577 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2505b439-b489-4c27-a41c-3cacc6db2776" path="/var/lib/kubelet/pods/2505b439-b489-4c27-a41c-3cacc6db2776/volumes" Apr 17 20:48:29.221394 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.221357 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:48:29.346199 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.346169 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Apr 17 20:48:29.348556 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:48:29.348524 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod356df3f8_dff9_4963_a883_fe17c0c75d95.slice/crio-0391d549984b4cce1a71e419581f0a593acb02fff299476829f35e4a7a9c0332 WatchSource:0}: Error finding container 0391d549984b4cce1a71e419581f0a593acb02fff299476829f35e4a7a9c0332: Status 404 returned error can't find the container with id 0391d549984b4cce1a71e419581f0a593acb02fff299476829f35e4a7a9c0332 Apr 17 20:48:29.855551 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.855519 2577 generic.go:358] "Generic (PLEG): container finished" podID="356df3f8-dff9-4963-a883-fe17c0c75d95" containerID="2ec5c93c275db09969b990e569774708a6c56851b79063dda77dbe0bc396cf80" exitCode=0 Apr 17 20:48:29.855941 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.855612 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"356df3f8-dff9-4963-a883-fe17c0c75d95","Type":"ContainerDied","Data":"2ec5c93c275db09969b990e569774708a6c56851b79063dda77dbe0bc396cf80"} Apr 17 20:48:29.855941 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:29.855649 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"356df3f8-dff9-4963-a883-fe17c0c75d95","Type":"ContainerStarted","Data":"0391d549984b4cce1a71e419581f0a593acb02fff299476829f35e4a7a9c0332"} Apr 17 20:48:30.862104 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:30.862069 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"356df3f8-dff9-4963-a883-fe17c0c75d95","Type":"ContainerStarted","Data":"538df3659e2a55acc99ca9a2f039b728c31ef25add7acdbdce5ca13bfc0cef41"} Apr 17 20:48:30.862104 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:30.862107 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"356df3f8-dff9-4963-a883-fe17c0c75d95","Type":"ContainerStarted","Data":"808e6546e8913021fc865f5a32a44d3f37fe39702189317223b43d082d687486"} Apr 17 20:48:30.862498 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:30.862119 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"356df3f8-dff9-4963-a883-fe17c0c75d95","Type":"ContainerStarted","Data":"7dccedba210588a05973585dbb0a854c7a39819e9a19d2748b5df7b646e158c0"} Apr 17 20:48:30.862498 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:30.862127 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"356df3f8-dff9-4963-a883-fe17c0c75d95","Type":"ContainerStarted","Data":"0052defc03be1c4d345d28b0fc2653a10991182945787c2bc067ac07b997a14c"} Apr 17 20:48:30.862498 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:30.862135 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"356df3f8-dff9-4963-a883-fe17c0c75d95","Type":"ContainerStarted","Data":"2053c919c1ef96877830921d25af454a58298a2d97bbcea84b0f31a47819d7f4"} Apr 17 20:48:30.862498 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:30.862143 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"356df3f8-dff9-4963-a883-fe17c0c75d95","Type":"ContainerStarted","Data":"d8feda7180deb7b84057df38d76ae0cee17cf770d816256f8aeae3701f6b2dbb"} Apr 17 20:48:30.886322 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:30.886268 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-k8s-0" podStartSLOduration=2.886252651 podStartE2EDuration="2.886252651s" podCreationTimestamp="2026-04-17 20:48:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 20:48:30.8848697 +0000 UTC m=+236.358647987" watchObservedRunningTime="2026-04-17 20:48:30.886252651 +0000 UTC m=+236.360030939" Apr 17 20:48:34.222085 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:48:34.222052 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:49:29.221555 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:49:29.221518 2577 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:49:29.237015 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:49:29.236994 2577 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:49:30.046518 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:49:30.046491 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-0" Apr 17 20:49:35.017486 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:49:35.017449 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-9rmnh_e562e939-99f0-4bee-952d-7af46132c860/console-operator/2.log" Apr 17 20:49:35.018571 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:49:35.018543 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-9rmnh_e562e939-99f0-4bee-952d-7af46132c860/console-operator/2.log" Apr 17 20:49:35.023483 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:49:35.023462 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dk97h_37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0/ovn-acl-logging/0.log" Apr 17 20:49:35.024566 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:49:35.024546 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dk97h_37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0/ovn-acl-logging/0.log" Apr 17 20:50:07.895496 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:50:07.895461 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kube-system/global-pull-secret-syncer-flcph"] Apr 17 20:50:07.898868 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:50:07.898851 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-flcph" Apr 17 20:50:07.900673 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:50:07.900650 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"original-pull-secret\"" Apr 17 20:50:07.907684 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:50:07.907661 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kube-system/global-pull-secret-syncer-flcph"] Apr 17 20:50:07.977206 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:50:07.977171 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/103da256-764a-4e19-8698-330c81744ce0-original-pull-secret\") pod \"global-pull-secret-syncer-flcph\" (UID: \"103da256-764a-4e19-8698-330c81744ce0\") " pod="kube-system/global-pull-secret-syncer-flcph" Apr 17 20:50:07.977327 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:50:07.977225 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/103da256-764a-4e19-8698-330c81744ce0-kubelet-config\") pod \"global-pull-secret-syncer-flcph\" (UID: \"103da256-764a-4e19-8698-330c81744ce0\") " pod="kube-system/global-pull-secret-syncer-flcph" Apr 17 20:50:07.977327 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:50:07.977246 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/103da256-764a-4e19-8698-330c81744ce0-dbus\") pod \"global-pull-secret-syncer-flcph\" (UID: \"103da256-764a-4e19-8698-330c81744ce0\") " pod="kube-system/global-pull-secret-syncer-flcph" Apr 17 20:50:08.078096 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:50:08.078074 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/103da256-764a-4e19-8698-330c81744ce0-original-pull-secret\") pod \"global-pull-secret-syncer-flcph\" (UID: \"103da256-764a-4e19-8698-330c81744ce0\") " pod="kube-system/global-pull-secret-syncer-flcph" Apr 17 20:50:08.078213 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:50:08.078146 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/103da256-764a-4e19-8698-330c81744ce0-kubelet-config\") pod \"global-pull-secret-syncer-flcph\" (UID: \"103da256-764a-4e19-8698-330c81744ce0\") " pod="kube-system/global-pull-secret-syncer-flcph" Apr 17 20:50:08.078213 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:50:08.078164 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/103da256-764a-4e19-8698-330c81744ce0-dbus\") pod \"global-pull-secret-syncer-flcph\" (UID: \"103da256-764a-4e19-8698-330c81744ce0\") " pod="kube-system/global-pull-secret-syncer-flcph" Apr 17 20:50:08.078283 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:50:08.078268 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/103da256-764a-4e19-8698-330c81744ce0-kubelet-config\") pod \"global-pull-secret-syncer-flcph\" (UID: \"103da256-764a-4e19-8698-330c81744ce0\") " pod="kube-system/global-pull-secret-syncer-flcph" Apr 17 20:50:08.078334 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:50:08.078322 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/103da256-764a-4e19-8698-330c81744ce0-dbus\") pod \"global-pull-secret-syncer-flcph\" (UID: \"103da256-764a-4e19-8698-330c81744ce0\") " pod="kube-system/global-pull-secret-syncer-flcph" Apr 17 20:50:08.080245 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:50:08.080229 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/103da256-764a-4e19-8698-330c81744ce0-original-pull-secret\") pod \"global-pull-secret-syncer-flcph\" (UID: \"103da256-764a-4e19-8698-330c81744ce0\") " pod="kube-system/global-pull-secret-syncer-flcph" Apr 17 20:50:08.209107 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:50:08.209074 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-flcph" Apr 17 20:50:08.327191 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:50:08.327121 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kube-system/global-pull-secret-syncer-flcph"] Apr 17 20:50:08.329857 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:50:08.329829 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod103da256_764a_4e19_8698_330c81744ce0.slice/crio-f4f2239508fc1d164180c0ae2166409f2135e7434f3ca1bb2ce74653d6d88455 WatchSource:0}: Error finding container f4f2239508fc1d164180c0ae2166409f2135e7434f3ca1bb2ce74653d6d88455: Status 404 returned error can't find the container with id f4f2239508fc1d164180c0ae2166409f2135e7434f3ca1bb2ce74653d6d88455 Apr 17 20:50:08.331264 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:50:08.331245 2577 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 17 20:50:09.139229 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:50:09.139200 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/global-pull-secret-syncer-flcph" event={"ID":"103da256-764a-4e19-8698-330c81744ce0","Type":"ContainerStarted","Data":"f4f2239508fc1d164180c0ae2166409f2135e7434f3ca1bb2ce74653d6d88455"} Apr 17 20:50:14.160025 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:50:14.159983 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/global-pull-secret-syncer-flcph" event={"ID":"103da256-764a-4e19-8698-330c81744ce0","Type":"ContainerStarted","Data":"2b24dd005095548503888a993ce4dbaa7136de9ff5b9ad569728c60f50c83bda"} Apr 17 20:50:14.174316 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:50:14.174263 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/global-pull-secret-syncer-flcph" podStartSLOduration=2.369927456 podStartE2EDuration="7.174246158s" podCreationTimestamp="2026-04-17 20:50:07 +0000 UTC" firstStartedPulling="2026-04-17 20:50:08.331384846 +0000 UTC m=+333.805163112" lastFinishedPulling="2026-04-17 20:50:13.135703548 +0000 UTC m=+338.609481814" observedRunningTime="2026-04-17 20:50:14.172166998 +0000 UTC m=+339.645945285" watchObservedRunningTime="2026-04-17 20:50:14.174246158 +0000 UTC m=+339.648024436" Apr 17 20:51:35.085630 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:35.085600 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["opendatahub/opendatahub-operator-controller-manager-6dc4849f89-szktp"] Apr 17 20:51:35.089073 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:35.089045 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/opendatahub-operator-controller-manager-6dc4849f89-szktp" Apr 17 20:51:35.093704 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:35.093677 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"opendatahub\"/\"openshift-service-ca.crt\"" Apr 17 20:51:35.094187 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:35.094160 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"opendatahub\"/\"opendatahub-operator-controller-webhook-cert\"" Apr 17 20:51:35.094187 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:35.094181 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"opendatahub\"/\"opendatahub-operator-controller-manager-service-cert\"" Apr 17 20:51:35.094367 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:35.094218 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"opendatahub\"/\"kube-root-ca.crt\"" Apr 17 20:51:35.094367 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:35.094277 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"opendatahub\"/\"opendatahub-operator-controller-manager-dockercfg-qjnvg\"" Apr 17 20:51:35.102201 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:35.102179 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/opendatahub-operator-controller-manager-6dc4849f89-szktp"] Apr 17 20:51:35.186944 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:35.186907 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fssmg\" (UniqueName: \"kubernetes.io/projected/c5901a00-633e-4ca1-8035-4599be6e70d8-kube-api-access-fssmg\") pod \"opendatahub-operator-controller-manager-6dc4849f89-szktp\" (UID: \"c5901a00-633e-4ca1-8035-4599be6e70d8\") " pod="opendatahub/opendatahub-operator-controller-manager-6dc4849f89-szktp" Apr 17 20:51:35.187116 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:35.186988 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c5901a00-633e-4ca1-8035-4599be6e70d8-webhook-cert\") pod \"opendatahub-operator-controller-manager-6dc4849f89-szktp\" (UID: \"c5901a00-633e-4ca1-8035-4599be6e70d8\") " pod="opendatahub/opendatahub-operator-controller-manager-6dc4849f89-szktp" Apr 17 20:51:35.187116 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:35.187016 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c5901a00-633e-4ca1-8035-4599be6e70d8-apiservice-cert\") pod \"opendatahub-operator-controller-manager-6dc4849f89-szktp\" (UID: \"c5901a00-633e-4ca1-8035-4599be6e70d8\") " pod="opendatahub/opendatahub-operator-controller-manager-6dc4849f89-szktp" Apr 17 20:51:35.288359 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:35.288327 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-fssmg\" (UniqueName: \"kubernetes.io/projected/c5901a00-633e-4ca1-8035-4599be6e70d8-kube-api-access-fssmg\") pod \"opendatahub-operator-controller-manager-6dc4849f89-szktp\" (UID: \"c5901a00-633e-4ca1-8035-4599be6e70d8\") " pod="opendatahub/opendatahub-operator-controller-manager-6dc4849f89-szktp" Apr 17 20:51:35.288502 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:35.288420 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c5901a00-633e-4ca1-8035-4599be6e70d8-webhook-cert\") pod \"opendatahub-operator-controller-manager-6dc4849f89-szktp\" (UID: \"c5901a00-633e-4ca1-8035-4599be6e70d8\") " pod="opendatahub/opendatahub-operator-controller-manager-6dc4849f89-szktp" Apr 17 20:51:35.288502 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:35.288442 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c5901a00-633e-4ca1-8035-4599be6e70d8-apiservice-cert\") pod \"opendatahub-operator-controller-manager-6dc4849f89-szktp\" (UID: \"c5901a00-633e-4ca1-8035-4599be6e70d8\") " pod="opendatahub/opendatahub-operator-controller-manager-6dc4849f89-szktp" Apr 17 20:51:35.290973 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:35.290944 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c5901a00-633e-4ca1-8035-4599be6e70d8-apiservice-cert\") pod \"opendatahub-operator-controller-manager-6dc4849f89-szktp\" (UID: \"c5901a00-633e-4ca1-8035-4599be6e70d8\") " pod="opendatahub/opendatahub-operator-controller-manager-6dc4849f89-szktp" Apr 17 20:51:35.291069 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:35.290989 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c5901a00-633e-4ca1-8035-4599be6e70d8-webhook-cert\") pod \"opendatahub-operator-controller-manager-6dc4849f89-szktp\" (UID: \"c5901a00-633e-4ca1-8035-4599be6e70d8\") " pod="opendatahub/opendatahub-operator-controller-manager-6dc4849f89-szktp" Apr 17 20:51:35.296713 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:35.296694 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-fssmg\" (UniqueName: \"kubernetes.io/projected/c5901a00-633e-4ca1-8035-4599be6e70d8-kube-api-access-fssmg\") pod \"opendatahub-operator-controller-manager-6dc4849f89-szktp\" (UID: \"c5901a00-633e-4ca1-8035-4599be6e70d8\") " pod="opendatahub/opendatahub-operator-controller-manager-6dc4849f89-szktp" Apr 17 20:51:35.399936 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:35.399883 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/opendatahub-operator-controller-manager-6dc4849f89-szktp" Apr 17 20:51:35.518320 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:35.518219 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/opendatahub-operator-controller-manager-6dc4849f89-szktp"] Apr 17 20:51:35.521694 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:51:35.521667 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5901a00_633e_4ca1_8035_4599be6e70d8.slice/crio-73900a78135769bf2e61ab190451e74485a1f413cc57d8b9efb1b8fdc49344ad WatchSource:0}: Error finding container 73900a78135769bf2e61ab190451e74485a1f413cc57d8b9efb1b8fdc49344ad: Status 404 returned error can't find the container with id 73900a78135769bf2e61ab190451e74485a1f413cc57d8b9efb1b8fdc49344ad Apr 17 20:51:36.399057 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:36.399018 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/opendatahub-operator-controller-manager-6dc4849f89-szktp" event={"ID":"c5901a00-633e-4ca1-8035-4599be6e70d8","Type":"ContainerStarted","Data":"73900a78135769bf2e61ab190451e74485a1f413cc57d8b9efb1b8fdc49344ad"} Apr 17 20:51:38.407624 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:38.407542 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/opendatahub-operator-controller-manager-6dc4849f89-szktp" event={"ID":"c5901a00-633e-4ca1-8035-4599be6e70d8","Type":"ContainerStarted","Data":"bb99acd5d65070b4d1dfe670e44fb5efc277de1626a875c869cadd101d575165"} Apr 17 20:51:38.408082 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:38.407714 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="opendatahub/opendatahub-operator-controller-manager-6dc4849f89-szktp" Apr 17 20:51:38.427849 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:38.427784 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="opendatahub/opendatahub-operator-controller-manager-6dc4849f89-szktp" podStartSLOduration=0.927836453 podStartE2EDuration="3.427769635s" podCreationTimestamp="2026-04-17 20:51:35 +0000 UTC" firstStartedPulling="2026-04-17 20:51:35.523347718 +0000 UTC m=+420.997125987" lastFinishedPulling="2026-04-17 20:51:38.023280903 +0000 UTC m=+423.497059169" observedRunningTime="2026-04-17 20:51:38.426665637 +0000 UTC m=+423.900443925" watchObservedRunningTime="2026-04-17 20:51:38.427769635 +0000 UTC m=+423.901547921" Apr 17 20:51:47.509560 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:47.509528 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-lws-operator/lws-controller-manager-7f68665c84-2nrp7"] Apr 17 20:51:47.513066 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:47.513044 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-lws-operator/lws-controller-manager-7f68665c84-2nrp7" Apr 17 20:51:47.515663 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:47.515641 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-lws-operator\"/\"openshift-service-ca.crt\"" Apr 17 20:51:47.515794 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:47.515719 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-lws-operator\"/\"kube-root-ca.crt\"" Apr 17 20:51:47.515794 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:47.515734 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-lws-operator\"/\"lws-manager-config\"" Apr 17 20:51:47.515794 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:47.515771 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-lws-operator\"/\"lws-controller-manager-dockercfg-2rm4j\"" Apr 17 20:51:47.515794 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:47.515773 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-lws-operator\"/\"webhook-server-cert\"" Apr 17 20:51:47.515985 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:47.515836 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-lws-operator\"/\"metrics-server-cert\"" Apr 17 20:51:47.520140 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:47.519779 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-lws-operator/lws-controller-manager-7f68665c84-2nrp7"] Apr 17 20:51:47.590308 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:47.590283 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/926fd6e1-ce48-4774-be2c-5280dfdc85ad-manager-config\") pod \"lws-controller-manager-7f68665c84-2nrp7\" (UID: \"926fd6e1-ce48-4774-be2c-5280dfdc85ad\") " pod="openshift-lws-operator/lws-controller-manager-7f68665c84-2nrp7" Apr 17 20:51:47.590423 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:47.590322 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4pbk\" (UniqueName: \"kubernetes.io/projected/926fd6e1-ce48-4774-be2c-5280dfdc85ad-kube-api-access-j4pbk\") pod \"lws-controller-manager-7f68665c84-2nrp7\" (UID: \"926fd6e1-ce48-4774-be2c-5280dfdc85ad\") " pod="openshift-lws-operator/lws-controller-manager-7f68665c84-2nrp7" Apr 17 20:51:47.590472 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:47.590431 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-cert\" (UniqueName: \"kubernetes.io/secret/926fd6e1-ce48-4774-be2c-5280dfdc85ad-metrics-cert\") pod \"lws-controller-manager-7f68665c84-2nrp7\" (UID: \"926fd6e1-ce48-4774-be2c-5280dfdc85ad\") " pod="openshift-lws-operator/lws-controller-manager-7f68665c84-2nrp7" Apr 17 20:51:47.590514 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:47.590473 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/926fd6e1-ce48-4774-be2c-5280dfdc85ad-cert\") pod \"lws-controller-manager-7f68665c84-2nrp7\" (UID: \"926fd6e1-ce48-4774-be2c-5280dfdc85ad\") " pod="openshift-lws-operator/lws-controller-manager-7f68665c84-2nrp7" Apr 17 20:51:47.691607 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:47.691583 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/926fd6e1-ce48-4774-be2c-5280dfdc85ad-manager-config\") pod \"lws-controller-manager-7f68665c84-2nrp7\" (UID: \"926fd6e1-ce48-4774-be2c-5280dfdc85ad\") " pod="openshift-lws-operator/lws-controller-manager-7f68665c84-2nrp7" Apr 17 20:51:47.691704 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:47.691609 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-j4pbk\" (UniqueName: \"kubernetes.io/projected/926fd6e1-ce48-4774-be2c-5280dfdc85ad-kube-api-access-j4pbk\") pod \"lws-controller-manager-7f68665c84-2nrp7\" (UID: \"926fd6e1-ce48-4774-be2c-5280dfdc85ad\") " pod="openshift-lws-operator/lws-controller-manager-7f68665c84-2nrp7" Apr 17 20:51:47.691704 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:47.691666 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-cert\" (UniqueName: \"kubernetes.io/secret/926fd6e1-ce48-4774-be2c-5280dfdc85ad-metrics-cert\") pod \"lws-controller-manager-7f68665c84-2nrp7\" (UID: \"926fd6e1-ce48-4774-be2c-5280dfdc85ad\") " pod="openshift-lws-operator/lws-controller-manager-7f68665c84-2nrp7" Apr 17 20:51:47.691704 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:47.691691 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/926fd6e1-ce48-4774-be2c-5280dfdc85ad-cert\") pod \"lws-controller-manager-7f68665c84-2nrp7\" (UID: \"926fd6e1-ce48-4774-be2c-5280dfdc85ad\") " pod="openshift-lws-operator/lws-controller-manager-7f68665c84-2nrp7" Apr 17 20:51:47.692247 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:47.692227 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/926fd6e1-ce48-4774-be2c-5280dfdc85ad-manager-config\") pod \"lws-controller-manager-7f68665c84-2nrp7\" (UID: \"926fd6e1-ce48-4774-be2c-5280dfdc85ad\") " pod="openshift-lws-operator/lws-controller-manager-7f68665c84-2nrp7" Apr 17 20:51:47.694288 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:47.694270 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-cert\" (UniqueName: \"kubernetes.io/secret/926fd6e1-ce48-4774-be2c-5280dfdc85ad-metrics-cert\") pod \"lws-controller-manager-7f68665c84-2nrp7\" (UID: \"926fd6e1-ce48-4774-be2c-5280dfdc85ad\") " pod="openshift-lws-operator/lws-controller-manager-7f68665c84-2nrp7" Apr 17 20:51:47.694365 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:47.694323 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/926fd6e1-ce48-4774-be2c-5280dfdc85ad-cert\") pod \"lws-controller-manager-7f68665c84-2nrp7\" (UID: \"926fd6e1-ce48-4774-be2c-5280dfdc85ad\") " pod="openshift-lws-operator/lws-controller-manager-7f68665c84-2nrp7" Apr 17 20:51:47.700065 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:47.700043 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4pbk\" (UniqueName: \"kubernetes.io/projected/926fd6e1-ce48-4774-be2c-5280dfdc85ad-kube-api-access-j4pbk\") pod \"lws-controller-manager-7f68665c84-2nrp7\" (UID: \"926fd6e1-ce48-4774-be2c-5280dfdc85ad\") " pod="openshift-lws-operator/lws-controller-manager-7f68665c84-2nrp7" Apr 17 20:51:47.822331 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:47.822275 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-lws-operator/lws-controller-manager-7f68665c84-2nrp7" Apr 17 20:51:47.955235 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:47.955207 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-lws-operator/lws-controller-manager-7f68665c84-2nrp7"] Apr 17 20:51:47.958853 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:51:47.958825 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod926fd6e1_ce48_4774_be2c_5280dfdc85ad.slice/crio-c43d4fd5871e2ac7f4d2b56a07a2f9fae120fde97276152ea6148dd10e9cbc7c WatchSource:0}: Error finding container c43d4fd5871e2ac7f4d2b56a07a2f9fae120fde97276152ea6148dd10e9cbc7c: Status 404 returned error can't find the container with id c43d4fd5871e2ac7f4d2b56a07a2f9fae120fde97276152ea6148dd10e9cbc7c Apr 17 20:51:48.440522 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:48.440491 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-lws-operator/lws-controller-manager-7f68665c84-2nrp7" event={"ID":"926fd6e1-ce48-4774-be2c-5280dfdc85ad","Type":"ContainerStarted","Data":"c43d4fd5871e2ac7f4d2b56a07a2f9fae120fde97276152ea6148dd10e9cbc7c"} Apr 17 20:51:49.413840 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:49.413784 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="opendatahub/opendatahub-operator-controller-manager-6dc4849f89-szktp" Apr 17 20:51:52.961272 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:52.961224 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress/kube-auth-proxy-674746b5f4-8jnrh"] Apr 17 20:51:52.964819 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:52.964776 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/kube-auth-proxy-674746b5f4-8jnrh" Apr 17 20:51:52.967238 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:52.967215 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"openshift-service-ca.crt\"" Apr 17 20:51:52.967724 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:52.967227 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"kube-root-ca.crt\"" Apr 17 20:51:52.968587 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:52.967249 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"kube-auth-proxy-creds\"" Apr 17 20:51:52.968587 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:52.967295 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"kube-auth-proxy-tls\"" Apr 17 20:51:52.968587 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:52.968564 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"kube-auth-proxy-dockercfg-k8cwx\"" Apr 17 20:51:52.971482 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:52.971461 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress/kube-auth-proxy-674746b5f4-8jnrh"] Apr 17 20:51:53.034642 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:53.034607 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certs\" (UniqueName: \"kubernetes.io/secret/e9f7fbd0-dec2-4736-877f-dde68502910c-tls-certs\") pod \"kube-auth-proxy-674746b5f4-8jnrh\" (UID: \"e9f7fbd0-dec2-4736-877f-dde68502910c\") " pod="openshift-ingress/kube-auth-proxy-674746b5f4-8jnrh" Apr 17 20:51:53.034792 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:53.034744 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/e9f7fbd0-dec2-4736-877f-dde68502910c-tmp\") pod \"kube-auth-proxy-674746b5f4-8jnrh\" (UID: \"e9f7fbd0-dec2-4736-877f-dde68502910c\") " pod="openshift-ingress/kube-auth-proxy-674746b5f4-8jnrh" Apr 17 20:51:53.034887 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:53.034790 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sk84n\" (UniqueName: \"kubernetes.io/projected/e9f7fbd0-dec2-4736-877f-dde68502910c-kube-api-access-sk84n\") pod \"kube-auth-proxy-674746b5f4-8jnrh\" (UID: \"e9f7fbd0-dec2-4736-877f-dde68502910c\") " pod="openshift-ingress/kube-auth-proxy-674746b5f4-8jnrh" Apr 17 20:51:53.135456 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:53.135419 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-sk84n\" (UniqueName: \"kubernetes.io/projected/e9f7fbd0-dec2-4736-877f-dde68502910c-kube-api-access-sk84n\") pod \"kube-auth-proxy-674746b5f4-8jnrh\" (UID: \"e9f7fbd0-dec2-4736-877f-dde68502910c\") " pod="openshift-ingress/kube-auth-proxy-674746b5f4-8jnrh" Apr 17 20:51:53.135628 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:53.135559 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-certs\" (UniqueName: \"kubernetes.io/secret/e9f7fbd0-dec2-4736-877f-dde68502910c-tls-certs\") pod \"kube-auth-proxy-674746b5f4-8jnrh\" (UID: \"e9f7fbd0-dec2-4736-877f-dde68502910c\") " pod="openshift-ingress/kube-auth-proxy-674746b5f4-8jnrh" Apr 17 20:51:53.135694 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:53.135675 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/e9f7fbd0-dec2-4736-877f-dde68502910c-tmp\") pod \"kube-auth-proxy-674746b5f4-8jnrh\" (UID: \"e9f7fbd0-dec2-4736-877f-dde68502910c\") " pod="openshift-ingress/kube-auth-proxy-674746b5f4-8jnrh" Apr 17 20:51:53.138600 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:53.138573 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/e9f7fbd0-dec2-4736-877f-dde68502910c-tmp\") pod \"kube-auth-proxy-674746b5f4-8jnrh\" (UID: \"e9f7fbd0-dec2-4736-877f-dde68502910c\") " pod="openshift-ingress/kube-auth-proxy-674746b5f4-8jnrh" Apr 17 20:51:53.138891 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:53.138865 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-certs\" (UniqueName: \"kubernetes.io/secret/e9f7fbd0-dec2-4736-877f-dde68502910c-tls-certs\") pod \"kube-auth-proxy-674746b5f4-8jnrh\" (UID: \"e9f7fbd0-dec2-4736-877f-dde68502910c\") " pod="openshift-ingress/kube-auth-proxy-674746b5f4-8jnrh" Apr 17 20:51:53.143675 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:53.143635 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-sk84n\" (UniqueName: \"kubernetes.io/projected/e9f7fbd0-dec2-4736-877f-dde68502910c-kube-api-access-sk84n\") pod \"kube-auth-proxy-674746b5f4-8jnrh\" (UID: \"e9f7fbd0-dec2-4736-877f-dde68502910c\") " pod="openshift-ingress/kube-auth-proxy-674746b5f4-8jnrh" Apr 17 20:51:53.279824 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:53.279721 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/kube-auth-proxy-674746b5f4-8jnrh" Apr 17 20:51:53.427573 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:53.427546 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress/kube-auth-proxy-674746b5f4-8jnrh"] Apr 17 20:51:53.430442 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:51:53.430375 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9f7fbd0_dec2_4736_877f_dde68502910c.slice/crio-4dc9945210c7989bb5e52af61400266f2570dd5ace64a85a0558fcddc5481935 WatchSource:0}: Error finding container 4dc9945210c7989bb5e52af61400266f2570dd5ace64a85a0558fcddc5481935: Status 404 returned error can't find the container with id 4dc9945210c7989bb5e52af61400266f2570dd5ace64a85a0558fcddc5481935 Apr 17 20:51:53.458984 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:53.458955 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/kube-auth-proxy-674746b5f4-8jnrh" event={"ID":"e9f7fbd0-dec2-4736-877f-dde68502910c","Type":"ContainerStarted","Data":"4dc9945210c7989bb5e52af61400266f2570dd5ace64a85a0558fcddc5481935"} Apr 17 20:51:55.467864 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:55.467825 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-lws-operator/lws-controller-manager-7f68665c84-2nrp7" event={"ID":"926fd6e1-ce48-4774-be2c-5280dfdc85ad","Type":"ContainerStarted","Data":"76f73ac28c0e9b89e9c30af170c6aaa5f38de7fa895b338a0af45f85527751a8"} Apr 17 20:51:55.468333 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:55.467957 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-lws-operator/lws-controller-manager-7f68665c84-2nrp7" Apr 17 20:51:55.484642 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:55.484542 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-lws-operator/lws-controller-manager-7f68665c84-2nrp7" podStartSLOduration=1.660528913 podStartE2EDuration="8.484527282s" podCreationTimestamp="2026-04-17 20:51:47 +0000 UTC" firstStartedPulling="2026-04-17 20:51:47.960603064 +0000 UTC m=+433.434381329" lastFinishedPulling="2026-04-17 20:51:54.784601419 +0000 UTC m=+440.258379698" observedRunningTime="2026-04-17 20:51:55.481760677 +0000 UTC m=+440.955538966" watchObservedRunningTime="2026-04-17 20:51:55.484527282 +0000 UTC m=+440.958305568" Apr 17 20:51:57.475992 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:57.475959 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/kube-auth-proxy-674746b5f4-8jnrh" event={"ID":"e9f7fbd0-dec2-4736-877f-dde68502910c","Type":"ContainerStarted","Data":"cd0d23d0efcaf95f886efaea287e8b300b21968f611778b1e9aeb54a654fd19b"} Apr 17 20:51:57.489919 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:51:57.489875 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/kube-auth-proxy-674746b5f4-8jnrh" podStartSLOduration=1.8939459520000002 podStartE2EDuration="5.489862109s" podCreationTimestamp="2026-04-17 20:51:52 +0000 UTC" firstStartedPulling="2026-04-17 20:51:53.433229409 +0000 UTC m=+438.907007677" lastFinishedPulling="2026-04-17 20:51:57.029145554 +0000 UTC m=+442.502923834" observedRunningTime="2026-04-17 20:51:57.488210235 +0000 UTC m=+442.961988514" watchObservedRunningTime="2026-04-17 20:51:57.489862109 +0000 UTC m=+442.963640396" Apr 17 20:52:06.473225 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:52:06.473195 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-lws-operator/lws-controller-manager-7f68665c84-2nrp7" Apr 17 20:53:36.065671 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:36.065595 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/limitador-operator-controller-manager-85c4996f8c-dd9kc"] Apr 17 20:53:36.068787 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:36.068772 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-dd9kc" Apr 17 20:53:36.070858 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:36.070836 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kuadrant-system\"/\"openshift-service-ca.crt\"" Apr 17 20:53:36.070977 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:36.070859 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kuadrant-system\"/\"kube-root-ca.crt\"" Apr 17 20:53:36.071294 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:36.071276 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kuadrant-system\"/\"limitador-operator-controller-manager-dockercfg-cm9r2\"" Apr 17 20:53:36.078046 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:36.078021 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-operator-controller-manager-85c4996f8c-dd9kc"] Apr 17 20:53:36.212383 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:36.212319 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whtqs\" (UniqueName: \"kubernetes.io/projected/8a1b8789-504b-4bce-ae1f-47d622c087ea-kube-api-access-whtqs\") pod \"limitador-operator-controller-manager-85c4996f8c-dd9kc\" (UID: \"8a1b8789-504b-4bce-ae1f-47d622c087ea\") " pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-dd9kc" Apr 17 20:53:36.313359 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:36.313333 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-whtqs\" (UniqueName: \"kubernetes.io/projected/8a1b8789-504b-4bce-ae1f-47d622c087ea-kube-api-access-whtqs\") pod \"limitador-operator-controller-manager-85c4996f8c-dd9kc\" (UID: \"8a1b8789-504b-4bce-ae1f-47d622c087ea\") " pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-dd9kc" Apr 17 20:53:36.323399 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:36.323347 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-whtqs\" (UniqueName: \"kubernetes.io/projected/8a1b8789-504b-4bce-ae1f-47d622c087ea-kube-api-access-whtqs\") pod \"limitador-operator-controller-manager-85c4996f8c-dd9kc\" (UID: \"8a1b8789-504b-4bce-ae1f-47d622c087ea\") " pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-dd9kc" Apr 17 20:53:36.379511 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:36.379489 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-dd9kc" Apr 17 20:53:36.498967 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:36.498902 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-operator-controller-manager-85c4996f8c-dd9kc"] Apr 17 20:53:36.501212 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:53:36.501186 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a1b8789_504b_4bce_ae1f_47d622c087ea.slice/crio-68ce9843551387ea5889007da6ead81b7df62297de2498e92fc03259a8e8eb86 WatchSource:0}: Error finding container 68ce9843551387ea5889007da6ead81b7df62297de2498e92fc03259a8e8eb86: Status 404 returned error can't find the container with id 68ce9843551387ea5889007da6ead81b7df62297de2498e92fc03259a8e8eb86 Apr 17 20:53:36.813504 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:36.813470 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-dd9kc" event={"ID":"8a1b8789-504b-4bce-ae1f-47d622c087ea","Type":"ContainerStarted","Data":"68ce9843551387ea5889007da6ead81b7df62297de2498e92fc03259a8e8eb86"} Apr 17 20:53:38.825519 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:38.825480 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-dd9kc" event={"ID":"8a1b8789-504b-4bce-ae1f-47d622c087ea","Type":"ContainerStarted","Data":"9c8d72df2b503aa63d7ea68a65181799ba3949baba1b2b23f98ddd9245822343"} Apr 17 20:53:38.825901 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:38.825588 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-dd9kc" Apr 17 20:53:38.844547 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:38.844500 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-dd9kc" podStartSLOduration=1.025097344 podStartE2EDuration="2.844487781s" podCreationTimestamp="2026-04-17 20:53:36 +0000 UTC" firstStartedPulling="2026-04-17 20:53:36.503035259 +0000 UTC m=+541.976813526" lastFinishedPulling="2026-04-17 20:53:38.32242569 +0000 UTC m=+543.796203963" observedRunningTime="2026-04-17 20:53:38.843235123 +0000 UTC m=+544.317013408" watchObservedRunningTime="2026-04-17 20:53:38.844487781 +0000 UTC m=+544.318266069" Apr 17 20:53:49.831895 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:49.831861 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-dd9kc" Apr 17 20:53:52.222213 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:52.222180 2577 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/limitador-operator-controller-manager-85c4996f8c-dd9kc"] Apr 17 20:53:52.222641 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:52.222386 2577 kuberuntime_container.go:864] "Killing container with a grace period" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-dd9kc" podUID="8a1b8789-504b-4bce-ae1f-47d622c087ea" containerName="manager" containerID="cri-o://9c8d72df2b503aa63d7ea68a65181799ba3949baba1b2b23f98ddd9245822343" gracePeriod=2 Apr 17 20:53:52.228647 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:52.228616 2577 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kuadrant-system/limitador-operator-controller-manager-85c4996f8c-dd9kc"] Apr 17 20:53:52.443667 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:52.443648 2577 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-dd9kc" Apr 17 20:53:52.445392 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:52.445370 2577 status_manager.go:895] "Failed to get status for pod" podUID="8a1b8789-504b-4bce-ae1f-47d622c087ea" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-dd9kc" err="pods \"limitador-operator-controller-manager-85c4996f8c-dd9kc\" is forbidden: User \"system:node:ip-10-0-137-110.ec2.internal\" cannot get resource \"pods\" in API group \"\" in the namespace \"kuadrant-system\": no relationship found between node 'ip-10-0-137-110.ec2.internal' and this object" Apr 17 20:53:52.637474 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:52.637410 2577 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whtqs\" (UniqueName: \"kubernetes.io/projected/8a1b8789-504b-4bce-ae1f-47d622c087ea-kube-api-access-whtqs\") pod \"8a1b8789-504b-4bce-ae1f-47d622c087ea\" (UID: \"8a1b8789-504b-4bce-ae1f-47d622c087ea\") " Apr 17 20:53:52.639567 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:52.639544 2577 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a1b8789-504b-4bce-ae1f-47d622c087ea-kube-api-access-whtqs" (OuterVolumeSpecName: "kube-api-access-whtqs") pod "8a1b8789-504b-4bce-ae1f-47d622c087ea" (UID: "8a1b8789-504b-4bce-ae1f-47d622c087ea"). InnerVolumeSpecName "kube-api-access-whtqs". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 17 20:53:52.738050 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:52.738027 2577 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-whtqs\" (UniqueName: \"kubernetes.io/projected/8a1b8789-504b-4bce-ae1f-47d622c087ea-kube-api-access-whtqs\") on node \"ip-10-0-137-110.ec2.internal\" DevicePath \"\"" Apr 17 20:53:52.879021 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:52.878989 2577 generic.go:358] "Generic (PLEG): container finished" podID="8a1b8789-504b-4bce-ae1f-47d622c087ea" containerID="9c8d72df2b503aa63d7ea68a65181799ba3949baba1b2b23f98ddd9245822343" exitCode=0 Apr 17 20:53:52.879166 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:52.879052 2577 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-dd9kc" Apr 17 20:53:52.879226 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:52.879061 2577 scope.go:117] "RemoveContainer" containerID="9c8d72df2b503aa63d7ea68a65181799ba3949baba1b2b23f98ddd9245822343" Apr 17 20:53:52.880880 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:52.880848 2577 status_manager.go:895] "Failed to get status for pod" podUID="8a1b8789-504b-4bce-ae1f-47d622c087ea" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-dd9kc" err="pods \"limitador-operator-controller-manager-85c4996f8c-dd9kc\" is forbidden: User \"system:node:ip-10-0-137-110.ec2.internal\" cannot get resource \"pods\" in API group \"\" in the namespace \"kuadrant-system\": no relationship found between node 'ip-10-0-137-110.ec2.internal' and this object" Apr 17 20:53:52.887696 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:52.887682 2577 scope.go:117] "RemoveContainer" containerID="9c8d72df2b503aa63d7ea68a65181799ba3949baba1b2b23f98ddd9245822343" Apr 17 20:53:52.888006 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:53:52.887988 2577 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c8d72df2b503aa63d7ea68a65181799ba3949baba1b2b23f98ddd9245822343\": container with ID starting with 9c8d72df2b503aa63d7ea68a65181799ba3949baba1b2b23f98ddd9245822343 not found: ID does not exist" containerID="9c8d72df2b503aa63d7ea68a65181799ba3949baba1b2b23f98ddd9245822343" Apr 17 20:53:52.888057 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:52.888014 2577 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c8d72df2b503aa63d7ea68a65181799ba3949baba1b2b23f98ddd9245822343"} err="failed to get container status \"9c8d72df2b503aa63d7ea68a65181799ba3949baba1b2b23f98ddd9245822343\": rpc error: code = NotFound desc = could not find container \"9c8d72df2b503aa63d7ea68a65181799ba3949baba1b2b23f98ddd9245822343\": container with ID starting with 9c8d72df2b503aa63d7ea68a65181799ba3949baba1b2b23f98ddd9245822343 not found: ID does not exist" Apr 17 20:53:52.888669 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:52.888649 2577 status_manager.go:895] "Failed to get status for pod" podUID="8a1b8789-504b-4bce-ae1f-47d622c087ea" pod="kuadrant-system/limitador-operator-controller-manager-85c4996f8c-dd9kc" err="pods \"limitador-operator-controller-manager-85c4996f8c-dd9kc\" is forbidden: User \"system:node:ip-10-0-137-110.ec2.internal\" cannot get resource \"pods\" in API group \"\" in the namespace \"kuadrant-system\": no relationship found between node 'ip-10-0-137-110.ec2.internal' and this object" Apr 17 20:53:53.120690 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:53:53.120662 2577 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a1b8789-504b-4bce-ae1f-47d622c087ea" path="/var/lib/kubelet/pods/8a1b8789-504b-4bce-ae1f-47d622c087ea/volumes" Apr 17 20:54:35.041830 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:35.041780 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-9rmnh_e562e939-99f0-4bee-952d-7af46132c860/console-operator/2.log" Apr 17 20:54:35.043765 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:35.043740 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-9rmnh_e562e939-99f0-4bee-952d-7af46132c860/console-operator/2.log" Apr 17 20:54:35.048985 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:35.048962 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dk97h_37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0/ovn-acl-logging/0.log" Apr 17 20:54:35.050410 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:35.050391 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dk97h_37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0/ovn-acl-logging/0.log" Apr 17 20:54:35.434747 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:35.434675 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/limitador-limitador-7d549b5b-rlfkl"] Apr 17 20:54:35.434989 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:35.434977 2577 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="8a1b8789-504b-4bce-ae1f-47d622c087ea" containerName="manager" Apr 17 20:54:35.435036 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:35.434991 2577 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a1b8789-504b-4bce-ae1f-47d622c087ea" containerName="manager" Apr 17 20:54:35.435075 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:35.435054 2577 memory_manager.go:356] "RemoveStaleState removing state" podUID="8a1b8789-504b-4bce-ae1f-47d622c087ea" containerName="manager" Apr 17 20:54:35.438058 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:35.438043 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/limitador-limitador-7d549b5b-rlfkl" Apr 17 20:54:35.440053 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:35.440027 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kuadrant-system\"/\"kube-root-ca.crt\"" Apr 17 20:54:35.440162 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:35.440077 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kuadrant-system\"/\"openshift-service-ca.crt\"" Apr 17 20:54:35.440162 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:35.440077 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kuadrant-system\"/\"limitador-limits-config-limitador\"" Apr 17 20:54:35.440162 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:35.440083 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kuadrant-system\"/\"default-dockercfg-d72jt\"" Apr 17 20:54:35.443419 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:35.443401 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-7d549b5b-rlfkl"] Apr 17 20:54:35.535193 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:35.535159 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-7d549b5b-rlfkl"] Apr 17 20:54:35.551736 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:35.551704 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-file\" (UniqueName: \"kubernetes.io/configmap/323d00ad-09a4-40d3-ba48-785c75ac0dc0-config-file\") pod \"limitador-limitador-7d549b5b-rlfkl\" (UID: \"323d00ad-09a4-40d3-ba48-785c75ac0dc0\") " pod="kuadrant-system/limitador-limitador-7d549b5b-rlfkl" Apr 17 20:54:35.551899 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:35.551790 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvz27\" (UniqueName: \"kubernetes.io/projected/323d00ad-09a4-40d3-ba48-785c75ac0dc0-kube-api-access-rvz27\") pod \"limitador-limitador-7d549b5b-rlfkl\" (UID: \"323d00ad-09a4-40d3-ba48-785c75ac0dc0\") " pod="kuadrant-system/limitador-limitador-7d549b5b-rlfkl" Apr 17 20:54:35.652599 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:35.652573 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-file\" (UniqueName: \"kubernetes.io/configmap/323d00ad-09a4-40d3-ba48-785c75ac0dc0-config-file\") pod \"limitador-limitador-7d549b5b-rlfkl\" (UID: \"323d00ad-09a4-40d3-ba48-785c75ac0dc0\") " pod="kuadrant-system/limitador-limitador-7d549b5b-rlfkl" Apr 17 20:54:35.652738 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:35.652626 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rvz27\" (UniqueName: \"kubernetes.io/projected/323d00ad-09a4-40d3-ba48-785c75ac0dc0-kube-api-access-rvz27\") pod \"limitador-limitador-7d549b5b-rlfkl\" (UID: \"323d00ad-09a4-40d3-ba48-785c75ac0dc0\") " pod="kuadrant-system/limitador-limitador-7d549b5b-rlfkl" Apr 17 20:54:35.653262 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:35.653240 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-file\" (UniqueName: \"kubernetes.io/configmap/323d00ad-09a4-40d3-ba48-785c75ac0dc0-config-file\") pod \"limitador-limitador-7d549b5b-rlfkl\" (UID: \"323d00ad-09a4-40d3-ba48-785c75ac0dc0\") " pod="kuadrant-system/limitador-limitador-7d549b5b-rlfkl" Apr 17 20:54:35.659858 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:35.659836 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvz27\" (UniqueName: \"kubernetes.io/projected/323d00ad-09a4-40d3-ba48-785c75ac0dc0-kube-api-access-rvz27\") pod \"limitador-limitador-7d549b5b-rlfkl\" (UID: \"323d00ad-09a4-40d3-ba48-785c75ac0dc0\") " pod="kuadrant-system/limitador-limitador-7d549b5b-rlfkl" Apr 17 20:54:35.750422 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:35.750401 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/limitador-limitador-7d549b5b-rlfkl" Apr 17 20:54:35.878911 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:35.878889 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/limitador-limitador-7d549b5b-rlfkl"] Apr 17 20:54:35.881584 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:54:35.881558 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod323d00ad_09a4_40d3_ba48_785c75ac0dc0.slice/crio-e30a17cc17b91595ba39b564c2d0f30d0f6251992cca45f97f1837b8a8b31948 WatchSource:0}: Error finding container e30a17cc17b91595ba39b564c2d0f30d0f6251992cca45f97f1837b8a8b31948: Status 404 returned error can't find the container with id e30a17cc17b91595ba39b564c2d0f30d0f6251992cca45f97f1837b8a8b31948 Apr 17 20:54:36.021639 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:36.021569 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/limitador-limitador-7d549b5b-rlfkl" event={"ID":"323d00ad-09a4-40d3-ba48-785c75ac0dc0","Type":"ContainerStarted","Data":"e30a17cc17b91595ba39b564c2d0f30d0f6251992cca45f97f1837b8a8b31948"} Apr 17 20:54:36.234735 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:36.234708 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/authorino-f99f4b5cd-sz87s"] Apr 17 20:54:36.239429 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:36.239400 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-f99f4b5cd-sz87s" Apr 17 20:54:36.241602 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:36.241483 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kuadrant-system\"/\"authorino-authorino-dockercfg-9jg8g\"" Apr 17 20:54:36.244425 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:36.244260 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-f99f4b5cd-sz87s"] Apr 17 20:54:36.358453 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:36.358382 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wmb8\" (UniqueName: \"kubernetes.io/projected/3e892578-9d83-4ee5-ae2b-8e37b27d94ee-kube-api-access-2wmb8\") pod \"authorino-f99f4b5cd-sz87s\" (UID: \"3e892578-9d83-4ee5-ae2b-8e37b27d94ee\") " pod="kuadrant-system/authorino-f99f4b5cd-sz87s" Apr 17 20:54:36.459496 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:36.459467 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-2wmb8\" (UniqueName: \"kubernetes.io/projected/3e892578-9d83-4ee5-ae2b-8e37b27d94ee-kube-api-access-2wmb8\") pod \"authorino-f99f4b5cd-sz87s\" (UID: \"3e892578-9d83-4ee5-ae2b-8e37b27d94ee\") " pod="kuadrant-system/authorino-f99f4b5cd-sz87s" Apr 17 20:54:36.467653 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:36.467618 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wmb8\" (UniqueName: \"kubernetes.io/projected/3e892578-9d83-4ee5-ae2b-8e37b27d94ee-kube-api-access-2wmb8\") pod \"authorino-f99f4b5cd-sz87s\" (UID: \"3e892578-9d83-4ee5-ae2b-8e37b27d94ee\") " pod="kuadrant-system/authorino-f99f4b5cd-sz87s" Apr 17 20:54:36.559285 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:36.559117 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-f99f4b5cd-sz87s" Apr 17 20:54:36.730349 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:36.730072 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-f99f4b5cd-sz87s"] Apr 17 20:54:37.026465 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:37.026431 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-f99f4b5cd-sz87s" event={"ID":"3e892578-9d83-4ee5-ae2b-8e37b27d94ee","Type":"ContainerStarted","Data":"82968c115ac57784238fb20ab6c8853ea077a5d1388f19fdc662591fbb5570da"} Apr 17 20:54:40.734132 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:40.734091 2577 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-f99f4b5cd-sz87s"] Apr 17 20:54:41.043293 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:41.043203 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-f99f4b5cd-sz87s" event={"ID":"3e892578-9d83-4ee5-ae2b-8e37b27d94ee","Type":"ContainerStarted","Data":"c1811d57765b0bb8bfcf501e6c08857eda8980ff47319dc7a39d2ca262aaef3d"} Apr 17 20:54:41.044501 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:41.044475 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/limitador-limitador-7d549b5b-rlfkl" event={"ID":"323d00ad-09a4-40d3-ba48-785c75ac0dc0","Type":"ContainerStarted","Data":"80cb94bf217fab48cb27bdbee309cb80f0b2ffe0743359c7984de06bde13b2da"} Apr 17 20:54:41.044646 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:41.044631 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kuadrant-system/limitador-limitador-7d549b5b-rlfkl" Apr 17 20:54:41.054003 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:41.053931 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/authorino-f99f4b5cd-sz87s" podStartSLOduration=1.359437815 podStartE2EDuration="5.053916499s" podCreationTimestamp="2026-04-17 20:54:36 +0000 UTC" firstStartedPulling="2026-04-17 20:54:36.736716445 +0000 UTC m=+602.210494712" lastFinishedPulling="2026-04-17 20:54:40.431195129 +0000 UTC m=+605.904973396" observedRunningTime="2026-04-17 20:54:41.05352339 +0000 UTC m=+606.527301677" watchObservedRunningTime="2026-04-17 20:54:41.053916499 +0000 UTC m=+606.527694783" Apr 17 20:54:41.067174 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:41.067132 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/limitador-limitador-7d549b5b-rlfkl" podStartSLOduration=1.526192316 podStartE2EDuration="6.06712078s" podCreationTimestamp="2026-04-17 20:54:35 +0000 UTC" firstStartedPulling="2026-04-17 20:54:35.883392824 +0000 UTC m=+601.357171089" lastFinishedPulling="2026-04-17 20:54:40.424321288 +0000 UTC m=+605.898099553" observedRunningTime="2026-04-17 20:54:41.0656571 +0000 UTC m=+606.539435390" watchObservedRunningTime="2026-04-17 20:54:41.06712078 +0000 UTC m=+606.540899067" Apr 17 20:54:42.047891 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:42.047824 2577 kuberuntime_container.go:864] "Killing container with a grace period" pod="kuadrant-system/authorino-f99f4b5cd-sz87s" podUID="3e892578-9d83-4ee5-ae2b-8e37b27d94ee" containerName="authorino" containerID="cri-o://c1811d57765b0bb8bfcf501e6c08857eda8980ff47319dc7a39d2ca262aaef3d" gracePeriod=30 Apr 17 20:54:42.287184 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:42.287164 2577 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-f99f4b5cd-sz87s" Apr 17 20:54:42.413030 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:42.412967 2577 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wmb8\" (UniqueName: \"kubernetes.io/projected/3e892578-9d83-4ee5-ae2b-8e37b27d94ee-kube-api-access-2wmb8\") pod \"3e892578-9d83-4ee5-ae2b-8e37b27d94ee\" (UID: \"3e892578-9d83-4ee5-ae2b-8e37b27d94ee\") " Apr 17 20:54:42.415151 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:42.415128 2577 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e892578-9d83-4ee5-ae2b-8e37b27d94ee-kube-api-access-2wmb8" (OuterVolumeSpecName: "kube-api-access-2wmb8") pod "3e892578-9d83-4ee5-ae2b-8e37b27d94ee" (UID: "3e892578-9d83-4ee5-ae2b-8e37b27d94ee"). InnerVolumeSpecName "kube-api-access-2wmb8". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 17 20:54:42.513454 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:42.513428 2577 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-2wmb8\" (UniqueName: \"kubernetes.io/projected/3e892578-9d83-4ee5-ae2b-8e37b27d94ee-kube-api-access-2wmb8\") on node \"ip-10-0-137-110.ec2.internal\" DevicePath \"\"" Apr 17 20:54:43.052128 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:43.052089 2577 generic.go:358] "Generic (PLEG): container finished" podID="3e892578-9d83-4ee5-ae2b-8e37b27d94ee" containerID="c1811d57765b0bb8bfcf501e6c08857eda8980ff47319dc7a39d2ca262aaef3d" exitCode=0 Apr 17 20:54:43.052554 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:43.052151 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-f99f4b5cd-sz87s" event={"ID":"3e892578-9d83-4ee5-ae2b-8e37b27d94ee","Type":"ContainerDied","Data":"c1811d57765b0bb8bfcf501e6c08857eda8980ff47319dc7a39d2ca262aaef3d"} Apr 17 20:54:43.052554 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:43.052156 2577 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-f99f4b5cd-sz87s" Apr 17 20:54:43.052554 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:43.052181 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-f99f4b5cd-sz87s" event={"ID":"3e892578-9d83-4ee5-ae2b-8e37b27d94ee","Type":"ContainerDied","Data":"82968c115ac57784238fb20ab6c8853ea077a5d1388f19fdc662591fbb5570da"} Apr 17 20:54:43.052554 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:43.052197 2577 scope.go:117] "RemoveContainer" containerID="c1811d57765b0bb8bfcf501e6c08857eda8980ff47319dc7a39d2ca262aaef3d" Apr 17 20:54:43.061230 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:43.061205 2577 scope.go:117] "RemoveContainer" containerID="c1811d57765b0bb8bfcf501e6c08857eda8980ff47319dc7a39d2ca262aaef3d" Apr 17 20:54:43.061487 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:54:43.061468 2577 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1811d57765b0bb8bfcf501e6c08857eda8980ff47319dc7a39d2ca262aaef3d\": container with ID starting with c1811d57765b0bb8bfcf501e6c08857eda8980ff47319dc7a39d2ca262aaef3d not found: ID does not exist" containerID="c1811d57765b0bb8bfcf501e6c08857eda8980ff47319dc7a39d2ca262aaef3d" Apr 17 20:54:43.061548 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:43.061496 2577 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1811d57765b0bb8bfcf501e6c08857eda8980ff47319dc7a39d2ca262aaef3d"} err="failed to get container status \"c1811d57765b0bb8bfcf501e6c08857eda8980ff47319dc7a39d2ca262aaef3d\": rpc error: code = NotFound desc = could not find container \"c1811d57765b0bb8bfcf501e6c08857eda8980ff47319dc7a39d2ca262aaef3d\": container with ID starting with c1811d57765b0bb8bfcf501e6c08857eda8980ff47319dc7a39d2ca262aaef3d not found: ID does not exist" Apr 17 20:54:43.069971 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:43.069944 2577 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-f99f4b5cd-sz87s"] Apr 17 20:54:43.073435 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:43.073412 2577 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kuadrant-system/authorino-f99f4b5cd-sz87s"] Apr 17 20:54:43.120461 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:43.120436 2577 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e892578-9d83-4ee5-ae2b-8e37b27d94ee" path="/var/lib/kubelet/pods/3e892578-9d83-4ee5-ae2b-8e37b27d94ee/volumes" Apr 17 20:54:50.955430 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:50.955393 2577 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/limitador-limitador-7d549b5b-rlfkl"] Apr 17 20:54:50.955865 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:50.955655 2577 kuberuntime_container.go:864] "Killing container with a grace period" pod="kuadrant-system/limitador-limitador-7d549b5b-rlfkl" podUID="323d00ad-09a4-40d3-ba48-785c75ac0dc0" containerName="limitador" containerID="cri-o://80cb94bf217fab48cb27bdbee309cb80f0b2ffe0743359c7984de06bde13b2da" gracePeriod=30 Apr 17 20:54:50.956329 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:50.956286 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kuadrant-system/limitador-limitador-7d549b5b-rlfkl" Apr 17 20:54:51.507978 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:51.507956 2577 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/limitador-limitador-7d549b5b-rlfkl" Apr 17 20:54:51.680973 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:51.680885 2577 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config-file\" (UniqueName: \"kubernetes.io/configmap/323d00ad-09a4-40d3-ba48-785c75ac0dc0-config-file\") pod \"323d00ad-09a4-40d3-ba48-785c75ac0dc0\" (UID: \"323d00ad-09a4-40d3-ba48-785c75ac0dc0\") " Apr 17 20:54:51.680973 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:51.680963 2577 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvz27\" (UniqueName: \"kubernetes.io/projected/323d00ad-09a4-40d3-ba48-785c75ac0dc0-kube-api-access-rvz27\") pod \"323d00ad-09a4-40d3-ba48-785c75ac0dc0\" (UID: \"323d00ad-09a4-40d3-ba48-785c75ac0dc0\") " Apr 17 20:54:51.681290 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:51.681261 2577 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/323d00ad-09a4-40d3-ba48-785c75ac0dc0-config-file" (OuterVolumeSpecName: "config-file") pod "323d00ad-09a4-40d3-ba48-785c75ac0dc0" (UID: "323d00ad-09a4-40d3-ba48-785c75ac0dc0"). InnerVolumeSpecName "config-file". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Apr 17 20:54:51.683091 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:51.683068 2577 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/323d00ad-09a4-40d3-ba48-785c75ac0dc0-kube-api-access-rvz27" (OuterVolumeSpecName: "kube-api-access-rvz27") pod "323d00ad-09a4-40d3-ba48-785c75ac0dc0" (UID: "323d00ad-09a4-40d3-ba48-785c75ac0dc0"). InnerVolumeSpecName "kube-api-access-rvz27". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 17 20:54:51.782117 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:51.782092 2577 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-rvz27\" (UniqueName: \"kubernetes.io/projected/323d00ad-09a4-40d3-ba48-785c75ac0dc0-kube-api-access-rvz27\") on node \"ip-10-0-137-110.ec2.internal\" DevicePath \"\"" Apr 17 20:54:51.782117 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:51.782114 2577 reconciler_common.go:299] "Volume detached for volume \"config-file\" (UniqueName: \"kubernetes.io/configmap/323d00ad-09a4-40d3-ba48-785c75ac0dc0-config-file\") on node \"ip-10-0-137-110.ec2.internal\" DevicePath \"\"" Apr 17 20:54:52.083244 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:52.083214 2577 generic.go:358] "Generic (PLEG): container finished" podID="323d00ad-09a4-40d3-ba48-785c75ac0dc0" containerID="80cb94bf217fab48cb27bdbee309cb80f0b2ffe0743359c7984de06bde13b2da" exitCode=0 Apr 17 20:54:52.083587 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:52.083282 2577 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/limitador-limitador-7d549b5b-rlfkl" Apr 17 20:54:52.083587 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:52.083281 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/limitador-limitador-7d549b5b-rlfkl" event={"ID":"323d00ad-09a4-40d3-ba48-785c75ac0dc0","Type":"ContainerDied","Data":"80cb94bf217fab48cb27bdbee309cb80f0b2ffe0743359c7984de06bde13b2da"} Apr 17 20:54:52.083587 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:52.083386 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/limitador-limitador-7d549b5b-rlfkl" event={"ID":"323d00ad-09a4-40d3-ba48-785c75ac0dc0","Type":"ContainerDied","Data":"e30a17cc17b91595ba39b564c2d0f30d0f6251992cca45f97f1837b8a8b31948"} Apr 17 20:54:52.083587 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:52.083402 2577 scope.go:117] "RemoveContainer" containerID="80cb94bf217fab48cb27bdbee309cb80f0b2ffe0743359c7984de06bde13b2da" Apr 17 20:54:52.093052 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:52.092783 2577 scope.go:117] "RemoveContainer" containerID="80cb94bf217fab48cb27bdbee309cb80f0b2ffe0743359c7984de06bde13b2da" Apr 17 20:54:52.093193 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:54:52.093166 2577 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80cb94bf217fab48cb27bdbee309cb80f0b2ffe0743359c7984de06bde13b2da\": container with ID starting with 80cb94bf217fab48cb27bdbee309cb80f0b2ffe0743359c7984de06bde13b2da not found: ID does not exist" containerID="80cb94bf217fab48cb27bdbee309cb80f0b2ffe0743359c7984de06bde13b2da" Apr 17 20:54:52.093249 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:52.093207 2577 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80cb94bf217fab48cb27bdbee309cb80f0b2ffe0743359c7984de06bde13b2da"} err="failed to get container status \"80cb94bf217fab48cb27bdbee309cb80f0b2ffe0743359c7984de06bde13b2da\": rpc error: code = NotFound desc = could not find container \"80cb94bf217fab48cb27bdbee309cb80f0b2ffe0743359c7984de06bde13b2da\": container with ID starting with 80cb94bf217fab48cb27bdbee309cb80f0b2ffe0743359c7984de06bde13b2da not found: ID does not exist" Apr 17 20:54:52.102412 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:52.102389 2577 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/limitador-limitador-7d549b5b-rlfkl"] Apr 17 20:54:52.105736 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:52.105717 2577 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kuadrant-system/limitador-limitador-7d549b5b-rlfkl"] Apr 17 20:54:53.119812 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:53.119774 2577 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="323d00ad-09a4-40d3-ba48-785c75ac0dc0" path="/var/lib/kubelet/pods/323d00ad-09a4-40d3-ba48-785c75ac0dc0/volumes" Apr 17 20:54:56.254494 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:56.254459 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["opendatahub/postgres-868db5846d-jdhnx"] Apr 17 20:54:56.254908 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:56.254793 2577 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="323d00ad-09a4-40d3-ba48-785c75ac0dc0" containerName="limitador" Apr 17 20:54:56.254908 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:56.254819 2577 state_mem.go:107] "Deleted CPUSet assignment" podUID="323d00ad-09a4-40d3-ba48-785c75ac0dc0" containerName="limitador" Apr 17 20:54:56.254908 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:56.254834 2577 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3e892578-9d83-4ee5-ae2b-8e37b27d94ee" containerName="authorino" Apr 17 20:54:56.254908 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:56.254839 2577 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e892578-9d83-4ee5-ae2b-8e37b27d94ee" containerName="authorino" Apr 17 20:54:56.254908 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:56.254890 2577 memory_manager.go:356] "RemoveStaleState removing state" podUID="323d00ad-09a4-40d3-ba48-785c75ac0dc0" containerName="limitador" Apr 17 20:54:56.254908 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:56.254897 2577 memory_manager.go:356] "RemoveStaleState removing state" podUID="3e892578-9d83-4ee5-ae2b-8e37b27d94ee" containerName="authorino" Apr 17 20:54:56.258058 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:56.258039 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/postgres-868db5846d-jdhnx" Apr 17 20:54:56.260317 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:56.260293 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"opendatahub\"/\"postgres-creds\"" Apr 17 20:54:56.260510 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:56.260296 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"opendatahub\"/\"default-dockercfg-jckc5\"" Apr 17 20:54:56.265058 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:56.265032 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/postgres-868db5846d-jdhnx"] Apr 17 20:54:56.416278 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:56.416246 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v227b\" (UniqueName: \"kubernetes.io/projected/36e1b7f5-63fe-40ed-8405-3a652226d92e-kube-api-access-v227b\") pod \"postgres-868db5846d-jdhnx\" (UID: \"36e1b7f5-63fe-40ed-8405-3a652226d92e\") " pod="opendatahub/postgres-868db5846d-jdhnx" Apr 17 20:54:56.416399 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:56.416299 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data\" (UniqueName: \"kubernetes.io/empty-dir/36e1b7f5-63fe-40ed-8405-3a652226d92e-data\") pod \"postgres-868db5846d-jdhnx\" (UID: \"36e1b7f5-63fe-40ed-8405-3a652226d92e\") " pod="opendatahub/postgres-868db5846d-jdhnx" Apr 17 20:54:56.516652 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:56.516591 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"data\" (UniqueName: \"kubernetes.io/empty-dir/36e1b7f5-63fe-40ed-8405-3a652226d92e-data\") pod \"postgres-868db5846d-jdhnx\" (UID: \"36e1b7f5-63fe-40ed-8405-3a652226d92e\") " pod="opendatahub/postgres-868db5846d-jdhnx" Apr 17 20:54:56.516652 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:56.516647 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-v227b\" (UniqueName: \"kubernetes.io/projected/36e1b7f5-63fe-40ed-8405-3a652226d92e-kube-api-access-v227b\") pod \"postgres-868db5846d-jdhnx\" (UID: \"36e1b7f5-63fe-40ed-8405-3a652226d92e\") " pod="opendatahub/postgres-868db5846d-jdhnx" Apr 17 20:54:56.517029 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:56.517008 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"data\" (UniqueName: \"kubernetes.io/empty-dir/36e1b7f5-63fe-40ed-8405-3a652226d92e-data\") pod \"postgres-868db5846d-jdhnx\" (UID: \"36e1b7f5-63fe-40ed-8405-3a652226d92e\") " pod="opendatahub/postgres-868db5846d-jdhnx" Apr 17 20:54:56.523851 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:56.523829 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-v227b\" (UniqueName: \"kubernetes.io/projected/36e1b7f5-63fe-40ed-8405-3a652226d92e-kube-api-access-v227b\") pod \"postgres-868db5846d-jdhnx\" (UID: \"36e1b7f5-63fe-40ed-8405-3a652226d92e\") " pod="opendatahub/postgres-868db5846d-jdhnx" Apr 17 20:54:56.571789 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:56.571768 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/postgres-868db5846d-jdhnx" Apr 17 20:54:56.898058 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:56.897992 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/postgres-868db5846d-jdhnx"] Apr 17 20:54:56.900938 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:54:56.900908 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod36e1b7f5_63fe_40ed_8405_3a652226d92e.slice/crio-5d70388367224caac7e8d77dcfc53ad005827c6e4edf6e96106f08e9ca9a2404 WatchSource:0}: Error finding container 5d70388367224caac7e8d77dcfc53ad005827c6e4edf6e96106f08e9ca9a2404: Status 404 returned error can't find the container with id 5d70388367224caac7e8d77dcfc53ad005827c6e4edf6e96106f08e9ca9a2404 Apr 17 20:54:57.100191 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:54:57.100151 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/postgres-868db5846d-jdhnx" event={"ID":"36e1b7f5-63fe-40ed-8405-3a652226d92e","Type":"ContainerStarted","Data":"5d70388367224caac7e8d77dcfc53ad005827c6e4edf6e96106f08e9ca9a2404"} Apr 17 20:55:02.118703 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:02.118667 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/postgres-868db5846d-jdhnx" event={"ID":"36e1b7f5-63fe-40ed-8405-3a652226d92e","Type":"ContainerStarted","Data":"c0ff27fb52c4550cbe1d0782434a5299d70f3efd01050b562934bc6ebdcbce68"} Apr 17 20:55:02.119089 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:02.118852 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="opendatahub/postgres-868db5846d-jdhnx" Apr 17 20:55:02.132239 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:02.132201 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="opendatahub/postgres-868db5846d-jdhnx" podStartSLOduration=1.434128551 podStartE2EDuration="6.132187722s" podCreationTimestamp="2026-04-17 20:54:56 +0000 UTC" firstStartedPulling="2026-04-17 20:54:56.902238801 +0000 UTC m=+622.376017066" lastFinishedPulling="2026-04-17 20:55:01.600297969 +0000 UTC m=+627.074076237" observedRunningTime="2026-04-17 20:55:02.130793424 +0000 UTC m=+627.604571711" watchObservedRunningTime="2026-04-17 20:55:02.132187722 +0000 UTC m=+627.605966009" Apr 17 20:55:08.156939 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:08.156908 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="opendatahub/postgres-868db5846d-jdhnx" Apr 17 20:55:08.675403 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:08.675366 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/authorino-8b475cf9f-bqb7z"] Apr 17 20:55:08.677568 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:08.677550 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-8b475cf9f-bqb7z" Apr 17 20:55:08.679373 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:08.679345 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kuadrant-system\"/\"kube-root-ca.crt\"" Apr 17 20:55:08.680005 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:08.679982 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kuadrant-system\"/\"openshift-service-ca.crt\"" Apr 17 20:55:08.680133 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:08.680032 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kuadrant-system\"/\"authorino-authorino-dockercfg-9jg8g\"" Apr 17 20:55:08.685153 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:08.685127 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-8b475cf9f-bqb7z"] Apr 17 20:55:08.713930 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:08.713905 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qt5s\" (UniqueName: \"kubernetes.io/projected/2b37e141-28db-4390-9278-5f1668f1649f-kube-api-access-6qt5s\") pod \"authorino-8b475cf9f-bqb7z\" (UID: \"2b37e141-28db-4390-9278-5f1668f1649f\") " pod="kuadrant-system/authorino-8b475cf9f-bqb7z" Apr 17 20:55:08.814337 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:08.814310 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6qt5s\" (UniqueName: \"kubernetes.io/projected/2b37e141-28db-4390-9278-5f1668f1649f-kube-api-access-6qt5s\") pod \"authorino-8b475cf9f-bqb7z\" (UID: \"2b37e141-28db-4390-9278-5f1668f1649f\") " pod="kuadrant-system/authorino-8b475cf9f-bqb7z" Apr 17 20:55:08.821251 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:08.821230 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qt5s\" (UniqueName: \"kubernetes.io/projected/2b37e141-28db-4390-9278-5f1668f1649f-kube-api-access-6qt5s\") pod \"authorino-8b475cf9f-bqb7z\" (UID: \"2b37e141-28db-4390-9278-5f1668f1649f\") " pod="kuadrant-system/authorino-8b475cf9f-bqb7z" Apr 17 20:55:08.899279 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:08.899251 2577 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-8b475cf9f-bqb7z"] Apr 17 20:55:08.899434 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:08.899423 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-8b475cf9f-bqb7z" Apr 17 20:55:08.923885 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:08.923860 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/authorino-648fc87456-h2tnm"] Apr 17 20:55:08.927826 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:08.927742 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-648fc87456-h2tnm" Apr 17 20:55:08.935469 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:08.935434 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-648fc87456-h2tnm"] Apr 17 20:55:09.015257 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:09.015226 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnwt8\" (UniqueName: \"kubernetes.io/projected/58e7919b-8470-4a28-8060-f6e087ebe859-kube-api-access-bnwt8\") pod \"authorino-648fc87456-h2tnm\" (UID: \"58e7919b-8470-4a28-8060-f6e087ebe859\") " pod="kuadrant-system/authorino-648fc87456-h2tnm" Apr 17 20:55:09.026233 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:09.026200 2577 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-8b475cf9f-bqb7z"] Apr 17 20:55:09.028359 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:55:09.028329 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b37e141_28db_4390_9278_5f1668f1649f.slice/crio-d6405ef98873f45f2d5e5f4ca3933d7b60f0eb29b87984aae49a54c92078744e WatchSource:0}: Error finding container d6405ef98873f45f2d5e5f4ca3933d7b60f0eb29b87984aae49a54c92078744e: Status 404 returned error can't find the container with id d6405ef98873f45f2d5e5f4ca3933d7b60f0eb29b87984aae49a54c92078744e Apr 17 20:55:09.029881 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:09.029861 2577 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 17 20:55:09.109300 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:09.109264 2577 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-648fc87456-h2tnm"] Apr 17 20:55:09.109529 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:55:09.109505 2577 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-bnwt8], unattached volumes=[], failed to process volumes=[]: context canceled" pod="kuadrant-system/authorino-648fc87456-h2tnm" podUID="58e7919b-8470-4a28-8060-f6e087ebe859" Apr 17 20:55:09.115969 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:09.115946 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-bnwt8\" (UniqueName: \"kubernetes.io/projected/58e7919b-8470-4a28-8060-f6e087ebe859-kube-api-access-bnwt8\") pod \"authorino-648fc87456-h2tnm\" (UID: \"58e7919b-8470-4a28-8060-f6e087ebe859\") " pod="kuadrant-system/authorino-648fc87456-h2tnm" Apr 17 20:55:09.123450 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:09.123427 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnwt8\" (UniqueName: \"kubernetes.io/projected/58e7919b-8470-4a28-8060-f6e087ebe859-kube-api-access-bnwt8\") pod \"authorino-648fc87456-h2tnm\" (UID: \"58e7919b-8470-4a28-8060-f6e087ebe859\") " pod="kuadrant-system/authorino-648fc87456-h2tnm" Apr 17 20:55:09.133245 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:09.133225 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kuadrant-system/authorino-56c57c5f84-gc6t4"] Apr 17 20:55:09.135411 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:09.135398 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-56c57c5f84-gc6t4" Apr 17 20:55:09.137287 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:09.137268 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kuadrant-system\"/\"authorino-server-cert\"" Apr 17 20:55:09.144649 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:09.144631 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-56c57c5f84-gc6t4"] Apr 17 20:55:09.148408 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:09.148384 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-648fc87456-h2tnm" Apr 17 20:55:09.148492 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:09.148381 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-8b475cf9f-bqb7z" event={"ID":"2b37e141-28db-4390-9278-5f1668f1649f","Type":"ContainerStarted","Data":"d6405ef98873f45f2d5e5f4ca3933d7b60f0eb29b87984aae49a54c92078744e"} Apr 17 20:55:09.152333 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:09.152311 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-648fc87456-h2tnm" Apr 17 20:55:09.216220 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:09.216195 2577 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnwt8\" (UniqueName: \"kubernetes.io/projected/58e7919b-8470-4a28-8060-f6e087ebe859-kube-api-access-bnwt8\") pod \"58e7919b-8470-4a28-8060-f6e087ebe859\" (UID: \"58e7919b-8470-4a28-8060-f6e087ebe859\") " Apr 17 20:55:09.216557 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:09.216432 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6f6f6\" (UniqueName: \"kubernetes.io/projected/6515bd08-2dbc-4177-8977-3ecdc4dcb979-kube-api-access-6f6f6\") pod \"authorino-56c57c5f84-gc6t4\" (UID: \"6515bd08-2dbc-4177-8977-3ecdc4dcb979\") " pod="kuadrant-system/authorino-56c57c5f84-gc6t4" Apr 17 20:55:09.216557 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:09.216526 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/6515bd08-2dbc-4177-8977-3ecdc4dcb979-tls-cert\") pod \"authorino-56c57c5f84-gc6t4\" (UID: \"6515bd08-2dbc-4177-8977-3ecdc4dcb979\") " pod="kuadrant-system/authorino-56c57c5f84-gc6t4" Apr 17 20:55:09.218297 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:09.218275 2577 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58e7919b-8470-4a28-8060-f6e087ebe859-kube-api-access-bnwt8" (OuterVolumeSpecName: "kube-api-access-bnwt8") pod "58e7919b-8470-4a28-8060-f6e087ebe859" (UID: "58e7919b-8470-4a28-8060-f6e087ebe859"). InnerVolumeSpecName "kube-api-access-bnwt8". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 17 20:55:09.317267 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:09.317246 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/6515bd08-2dbc-4177-8977-3ecdc4dcb979-tls-cert\") pod \"authorino-56c57c5f84-gc6t4\" (UID: \"6515bd08-2dbc-4177-8977-3ecdc4dcb979\") " pod="kuadrant-system/authorino-56c57c5f84-gc6t4" Apr 17 20:55:09.317378 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:09.317308 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6f6f6\" (UniqueName: \"kubernetes.io/projected/6515bd08-2dbc-4177-8977-3ecdc4dcb979-kube-api-access-6f6f6\") pod \"authorino-56c57c5f84-gc6t4\" (UID: \"6515bd08-2dbc-4177-8977-3ecdc4dcb979\") " pod="kuadrant-system/authorino-56c57c5f84-gc6t4" Apr 17 20:55:09.317378 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:09.317340 2577 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-bnwt8\" (UniqueName: \"kubernetes.io/projected/58e7919b-8470-4a28-8060-f6e087ebe859-kube-api-access-bnwt8\") on node \"ip-10-0-137-110.ec2.internal\" DevicePath \"\"" Apr 17 20:55:09.319473 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:09.319454 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-cert\" (UniqueName: \"kubernetes.io/secret/6515bd08-2dbc-4177-8977-3ecdc4dcb979-tls-cert\") pod \"authorino-56c57c5f84-gc6t4\" (UID: \"6515bd08-2dbc-4177-8977-3ecdc4dcb979\") " pod="kuadrant-system/authorino-56c57c5f84-gc6t4" Apr 17 20:55:09.323924 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:09.323901 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6f6f6\" (UniqueName: \"kubernetes.io/projected/6515bd08-2dbc-4177-8977-3ecdc4dcb979-kube-api-access-6f6f6\") pod \"authorino-56c57c5f84-gc6t4\" (UID: \"6515bd08-2dbc-4177-8977-3ecdc4dcb979\") " pod="kuadrant-system/authorino-56c57c5f84-gc6t4" Apr 17 20:55:09.445366 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:09.445344 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-56c57c5f84-gc6t4" Apr 17 20:55:09.570482 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:09.570438 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kuadrant-system/authorino-56c57c5f84-gc6t4"] Apr 17 20:55:09.572474 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:55:09.572443 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6515bd08_2dbc_4177_8977_3ecdc4dcb979.slice/crio-c0178fd2ec5a44619877b38ffe55e44ccf8fe0604bf80d3e359b596a51880772 WatchSource:0}: Error finding container c0178fd2ec5a44619877b38ffe55e44ccf8fe0604bf80d3e359b596a51880772: Status 404 returned error can't find the container with id c0178fd2ec5a44619877b38ffe55e44ccf8fe0604bf80d3e359b596a51880772 Apr 17 20:55:10.153288 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:10.153196 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-56c57c5f84-gc6t4" event={"ID":"6515bd08-2dbc-4177-8977-3ecdc4dcb979","Type":"ContainerStarted","Data":"345451fb0ccf00dcbb342a588ba7e5cb1158406368952395fbc03b5c57138d59"} Apr 17 20:55:10.153288 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:10.153240 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-56c57c5f84-gc6t4" event={"ID":"6515bd08-2dbc-4177-8977-3ecdc4dcb979","Type":"ContainerStarted","Data":"c0178fd2ec5a44619877b38ffe55e44ccf8fe0604bf80d3e359b596a51880772"} Apr 17 20:55:10.154490 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:10.154461 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-8b475cf9f-bqb7z" event={"ID":"2b37e141-28db-4390-9278-5f1668f1649f","Type":"ContainerStarted","Data":"64e16a9efd385bc54b38cffb656a3a11d01114800d917d0d2e53df5178550083"} Apr 17 20:55:10.154490 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:10.154478 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-648fc87456-h2tnm" Apr 17 20:55:10.154645 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:10.154558 2577 kuberuntime_container.go:864] "Killing container with a grace period" pod="kuadrant-system/authorino-8b475cf9f-bqb7z" podUID="2b37e141-28db-4390-9278-5f1668f1649f" containerName="authorino" containerID="cri-o://64e16a9efd385bc54b38cffb656a3a11d01114800d917d0d2e53df5178550083" gracePeriod=30 Apr 17 20:55:10.168358 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:10.168311 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/authorino-56c57c5f84-gc6t4" podStartSLOduration=0.834938883 podStartE2EDuration="1.168295015s" podCreationTimestamp="2026-04-17 20:55:09 +0000 UTC" firstStartedPulling="2026-04-17 20:55:09.573918125 +0000 UTC m=+635.047696390" lastFinishedPulling="2026-04-17 20:55:09.907274257 +0000 UTC m=+635.381052522" observedRunningTime="2026-04-17 20:55:10.165931051 +0000 UTC m=+635.639709338" watchObservedRunningTime="2026-04-17 20:55:10.168295015 +0000 UTC m=+635.642073303" Apr 17 20:55:10.184674 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:10.184634 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kuadrant-system/authorino-8b475cf9f-bqb7z" podStartSLOduration=1.8336062260000001 podStartE2EDuration="2.184620793s" podCreationTimestamp="2026-04-17 20:55:08 +0000 UTC" firstStartedPulling="2026-04-17 20:55:09.030042623 +0000 UTC m=+634.503820902" lastFinishedPulling="2026-04-17 20:55:09.381057201 +0000 UTC m=+634.854835469" observedRunningTime="2026-04-17 20:55:10.182909132 +0000 UTC m=+635.656687419" watchObservedRunningTime="2026-04-17 20:55:10.184620793 +0000 UTC m=+635.658399110" Apr 17 20:55:10.229499 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:10.229474 2577 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-648fc87456-h2tnm"] Apr 17 20:55:10.232961 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:10.232928 2577 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kuadrant-system/authorino-648fc87456-h2tnm"] Apr 17 20:55:10.383457 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:10.383429 2577 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-8b475cf9f-bqb7z" Apr 17 20:55:10.426770 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:10.426715 2577 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qt5s\" (UniqueName: \"kubernetes.io/projected/2b37e141-28db-4390-9278-5f1668f1649f-kube-api-access-6qt5s\") pod \"2b37e141-28db-4390-9278-5f1668f1649f\" (UID: \"2b37e141-28db-4390-9278-5f1668f1649f\") " Apr 17 20:55:10.428740 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:10.428717 2577 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b37e141-28db-4390-9278-5f1668f1649f-kube-api-access-6qt5s" (OuterVolumeSpecName: "kube-api-access-6qt5s") pod "2b37e141-28db-4390-9278-5f1668f1649f" (UID: "2b37e141-28db-4390-9278-5f1668f1649f"). InnerVolumeSpecName "kube-api-access-6qt5s". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 17 20:55:10.527778 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:10.527755 2577 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-6qt5s\" (UniqueName: \"kubernetes.io/projected/2b37e141-28db-4390-9278-5f1668f1649f-kube-api-access-6qt5s\") on node \"ip-10-0-137-110.ec2.internal\" DevicePath \"\"" Apr 17 20:55:10.933466 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:10.933433 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["opendatahub/maas-controller-6d4c8f55f9-fd4pp"] Apr 17 20:55:10.933766 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:10.933754 2577 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="2b37e141-28db-4390-9278-5f1668f1649f" containerName="authorino" Apr 17 20:55:10.933836 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:10.933768 2577 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b37e141-28db-4390-9278-5f1668f1649f" containerName="authorino" Apr 17 20:55:10.933876 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:10.933854 2577 memory_manager.go:356] "RemoveStaleState removing state" podUID="2b37e141-28db-4390-9278-5f1668f1649f" containerName="authorino" Apr 17 20:55:10.939104 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:10.939086 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-6d4c8f55f9-fd4pp" Apr 17 20:55:10.941033 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:10.941008 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"opendatahub\"/\"maas-controller-dockercfg-d94fj\"" Apr 17 20:55:10.942691 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:10.942668 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/maas-controller-6d4c8f55f9-fd4pp"] Apr 17 20:55:11.032655 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.032612 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmdbt\" (UniqueName: \"kubernetes.io/projected/4fccb7b6-35a5-4d1b-9d56-caeaf6ed9cc8-kube-api-access-xmdbt\") pod \"maas-controller-6d4c8f55f9-fd4pp\" (UID: \"4fccb7b6-35a5-4d1b-9d56-caeaf6ed9cc8\") " pod="opendatahub/maas-controller-6d4c8f55f9-fd4pp" Apr 17 20:55:11.066929 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.066894 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["opendatahub/maas-controller-69646d8cdf-trqwh"] Apr 17 20:55:11.072651 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.072632 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-69646d8cdf-trqwh" Apr 17 20:55:11.079263 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.079203 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/maas-controller-69646d8cdf-trqwh"] Apr 17 20:55:11.120688 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.120662 2577 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58e7919b-8470-4a28-8060-f6e087ebe859" path="/var/lib/kubelet/pods/58e7919b-8470-4a28-8060-f6e087ebe859/volumes" Apr 17 20:55:11.133609 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.133589 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-xmdbt\" (UniqueName: \"kubernetes.io/projected/4fccb7b6-35a5-4d1b-9d56-caeaf6ed9cc8-kube-api-access-xmdbt\") pod \"maas-controller-6d4c8f55f9-fd4pp\" (UID: \"4fccb7b6-35a5-4d1b-9d56-caeaf6ed9cc8\") " pod="opendatahub/maas-controller-6d4c8f55f9-fd4pp" Apr 17 20:55:11.133691 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.133653 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58gln\" (UniqueName: \"kubernetes.io/projected/4792930c-ec98-4b80-9284-cdbf5c515012-kube-api-access-58gln\") pod \"maas-controller-69646d8cdf-trqwh\" (UID: \"4792930c-ec98-4b80-9284-cdbf5c515012\") " pod="opendatahub/maas-controller-69646d8cdf-trqwh" Apr 17 20:55:11.142235 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.142211 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmdbt\" (UniqueName: \"kubernetes.io/projected/4fccb7b6-35a5-4d1b-9d56-caeaf6ed9cc8-kube-api-access-xmdbt\") pod \"maas-controller-6d4c8f55f9-fd4pp\" (UID: \"4fccb7b6-35a5-4d1b-9d56-caeaf6ed9cc8\") " pod="opendatahub/maas-controller-6d4c8f55f9-fd4pp" Apr 17 20:55:11.158189 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.158169 2577 generic.go:358] "Generic (PLEG): container finished" podID="2b37e141-28db-4390-9278-5f1668f1649f" containerID="64e16a9efd385bc54b38cffb656a3a11d01114800d917d0d2e53df5178550083" exitCode=0 Apr 17 20:55:11.158286 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.158226 2577 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kuadrant-system/authorino-8b475cf9f-bqb7z" Apr 17 20:55:11.158286 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.158255 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-8b475cf9f-bqb7z" event={"ID":"2b37e141-28db-4390-9278-5f1668f1649f","Type":"ContainerDied","Data":"64e16a9efd385bc54b38cffb656a3a11d01114800d917d0d2e53df5178550083"} Apr 17 20:55:11.158398 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.158299 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kuadrant-system/authorino-8b475cf9f-bqb7z" event={"ID":"2b37e141-28db-4390-9278-5f1668f1649f","Type":"ContainerDied","Data":"d6405ef98873f45f2d5e5f4ca3933d7b60f0eb29b87984aae49a54c92078744e"} Apr 17 20:55:11.158398 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.158322 2577 scope.go:117] "RemoveContainer" containerID="64e16a9efd385bc54b38cffb656a3a11d01114800d917d0d2e53df5178550083" Apr 17 20:55:11.165976 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.165951 2577 scope.go:117] "RemoveContainer" containerID="64e16a9efd385bc54b38cffb656a3a11d01114800d917d0d2e53df5178550083" Apr 17 20:55:11.166257 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:55:11.166238 2577 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64e16a9efd385bc54b38cffb656a3a11d01114800d917d0d2e53df5178550083\": container with ID starting with 64e16a9efd385bc54b38cffb656a3a11d01114800d917d0d2e53df5178550083 not found: ID does not exist" containerID="64e16a9efd385bc54b38cffb656a3a11d01114800d917d0d2e53df5178550083" Apr 17 20:55:11.166335 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.166263 2577 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64e16a9efd385bc54b38cffb656a3a11d01114800d917d0d2e53df5178550083"} err="failed to get container status \"64e16a9efd385bc54b38cffb656a3a11d01114800d917d0d2e53df5178550083\": rpc error: code = NotFound desc = could not find container \"64e16a9efd385bc54b38cffb656a3a11d01114800d917d0d2e53df5178550083\": container with ID starting with 64e16a9efd385bc54b38cffb656a3a11d01114800d917d0d2e53df5178550083 not found: ID does not exist" Apr 17 20:55:11.176922 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.176894 2577 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["kuadrant-system/authorino-8b475cf9f-bqb7z"] Apr 17 20:55:11.180987 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.180968 2577 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["opendatahub/maas-controller-6d4c8f55f9-fd4pp"] Apr 17 20:55:11.181155 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.181145 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-6d4c8f55f9-fd4pp" Apr 17 20:55:11.186557 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.186536 2577 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["kuadrant-system/authorino-8b475cf9f-bqb7z"] Apr 17 20:55:11.205452 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.205428 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["opendatahub/maas-controller-58cbf4b6f4-vkqzh"] Apr 17 20:55:11.210962 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.210942 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-58cbf4b6f4-vkqzh" Apr 17 20:55:11.215511 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.215489 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/maas-controller-58cbf4b6f4-vkqzh"] Apr 17 20:55:11.236823 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.236593 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-58gln\" (UniqueName: \"kubernetes.io/projected/4792930c-ec98-4b80-9284-cdbf5c515012-kube-api-access-58gln\") pod \"maas-controller-69646d8cdf-trqwh\" (UID: \"4792930c-ec98-4b80-9284-cdbf5c515012\") " pod="opendatahub/maas-controller-69646d8cdf-trqwh" Apr 17 20:55:11.237233 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.236886 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cdz4\" (UniqueName: \"kubernetes.io/projected/e804a8aa-e5fa-4219-808c-d537195b6319-kube-api-access-8cdz4\") pod \"maas-controller-58cbf4b6f4-vkqzh\" (UID: \"e804a8aa-e5fa-4219-808c-d537195b6319\") " pod="opendatahub/maas-controller-58cbf4b6f4-vkqzh" Apr 17 20:55:11.243578 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.243521 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-58gln\" (UniqueName: \"kubernetes.io/projected/4792930c-ec98-4b80-9284-cdbf5c515012-kube-api-access-58gln\") pod \"maas-controller-69646d8cdf-trqwh\" (UID: \"4792930c-ec98-4b80-9284-cdbf5c515012\") " pod="opendatahub/maas-controller-69646d8cdf-trqwh" Apr 17 20:55:11.312317 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.312156 2577 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["opendatahub/maas-controller-6d4c8f55f9-fd4pp"] Apr 17 20:55:11.315049 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:55:11.315023 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4fccb7b6_35a5_4d1b_9d56_caeaf6ed9cc8.slice/crio-97af318f565ae74941eb196ae0f289968be386e9af04d69fd3bf624c4a199ccf WatchSource:0}: Error finding container 97af318f565ae74941eb196ae0f289968be386e9af04d69fd3bf624c4a199ccf: Status 404 returned error can't find the container with id 97af318f565ae74941eb196ae0f289968be386e9af04d69fd3bf624c4a199ccf Apr 17 20:55:11.338094 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.338039 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8cdz4\" (UniqueName: \"kubernetes.io/projected/e804a8aa-e5fa-4219-808c-d537195b6319-kube-api-access-8cdz4\") pod \"maas-controller-58cbf4b6f4-vkqzh\" (UID: \"e804a8aa-e5fa-4219-808c-d537195b6319\") " pod="opendatahub/maas-controller-58cbf4b6f4-vkqzh" Apr 17 20:55:11.345047 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.345026 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cdz4\" (UniqueName: \"kubernetes.io/projected/e804a8aa-e5fa-4219-808c-d537195b6319-kube-api-access-8cdz4\") pod \"maas-controller-58cbf4b6f4-vkqzh\" (UID: \"e804a8aa-e5fa-4219-808c-d537195b6319\") " pod="opendatahub/maas-controller-58cbf4b6f4-vkqzh" Apr 17 20:55:11.384046 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.384019 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-69646d8cdf-trqwh" Apr 17 20:55:11.523939 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.523907 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-58cbf4b6f4-vkqzh" Apr 17 20:55:11.640092 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.640069 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/maas-controller-58cbf4b6f4-vkqzh"] Apr 17 20:55:11.642642 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:55:11.642618 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode804a8aa_e5fa_4219_808c_d537195b6319.slice/crio-c04d756fba587d4618e370d001f335db6823c65eb6ec782d5f10df8740106a44 WatchSource:0}: Error finding container c04d756fba587d4618e370d001f335db6823c65eb6ec782d5f10df8740106a44: Status 404 returned error can't find the container with id c04d756fba587d4618e370d001f335db6823c65eb6ec782d5f10df8740106a44 Apr 17 20:55:11.705911 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:11.705889 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["opendatahub/maas-controller-69646d8cdf-trqwh"] Apr 17 20:55:11.707506 ip-10-0-137-110 kubenswrapper[2577]: W0417 20:55:11.707482 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4792930c_ec98_4b80_9284_cdbf5c515012.slice/crio-5db7d58e70aa344f2dce5e30b3e8c3b2e6c6d169bd6363ffbdff468236fbe34b WatchSource:0}: Error finding container 5db7d58e70aa344f2dce5e30b3e8c3b2e6c6d169bd6363ffbdff468236fbe34b: Status 404 returned error can't find the container with id 5db7d58e70aa344f2dce5e30b3e8c3b2e6c6d169bd6363ffbdff468236fbe34b Apr 17 20:55:12.163759 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:12.163702 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-69646d8cdf-trqwh" event={"ID":"4792930c-ec98-4b80-9284-cdbf5c515012","Type":"ContainerStarted","Data":"5db7d58e70aa344f2dce5e30b3e8c3b2e6c6d169bd6363ffbdff468236fbe34b"} Apr 17 20:55:12.165050 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:12.164992 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-6d4c8f55f9-fd4pp" event={"ID":"4fccb7b6-35a5-4d1b-9d56-caeaf6ed9cc8","Type":"ContainerStarted","Data":"97af318f565ae74941eb196ae0f289968be386e9af04d69fd3bf624c4a199ccf"} Apr 17 20:55:12.167386 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:12.167316 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-58cbf4b6f4-vkqzh" event={"ID":"e804a8aa-e5fa-4219-808c-d537195b6319","Type":"ContainerStarted","Data":"c04d756fba587d4618e370d001f335db6823c65eb6ec782d5f10df8740106a44"} Apr 17 20:55:13.123936 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:13.123904 2577 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b37e141-28db-4390-9278-5f1668f1649f" path="/var/lib/kubelet/pods/2b37e141-28db-4390-9278-5f1668f1649f/volumes" Apr 17 20:55:15.180547 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:15.180505 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-58cbf4b6f4-vkqzh" event={"ID":"e804a8aa-e5fa-4219-808c-d537195b6319","Type":"ContainerStarted","Data":"2594f64fcb27b97b38ec9905f51a91d067b807dfdfb37b047d4947e176c09e51"} Apr 17 20:55:15.181550 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:15.180670 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="opendatahub/maas-controller-58cbf4b6f4-vkqzh" Apr 17 20:55:15.182320 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:15.182288 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-69646d8cdf-trqwh" event={"ID":"4792930c-ec98-4b80-9284-cdbf5c515012","Type":"ContainerStarted","Data":"afeec21397d333eafe2b4938cb9a614e56f3d9301294edbcd313040d876d9b33"} Apr 17 20:55:15.182437 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:15.182416 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="opendatahub/maas-controller-69646d8cdf-trqwh" Apr 17 20:55:15.183856 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:15.183823 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-6d4c8f55f9-fd4pp" event={"ID":"4fccb7b6-35a5-4d1b-9d56-caeaf6ed9cc8","Type":"ContainerStarted","Data":"caaaaf224da07fa827ec6c56d7600afb9ed193ab9fa5a2466ffb54db00d0b44c"} Apr 17 20:55:15.183951 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:15.183907 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="opendatahub/maas-controller-6d4c8f55f9-fd4pp" Apr 17 20:55:15.183951 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:15.183910 2577 kuberuntime_container.go:864] "Killing container with a grace period" pod="opendatahub/maas-controller-6d4c8f55f9-fd4pp" podUID="4fccb7b6-35a5-4d1b-9d56-caeaf6ed9cc8" containerName="manager" containerID="cri-o://caaaaf224da07fa827ec6c56d7600afb9ed193ab9fa5a2466ffb54db00d0b44c" gracePeriod=10 Apr 17 20:55:15.199566 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:15.199526 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="opendatahub/maas-controller-58cbf4b6f4-vkqzh" podStartSLOduration=0.881864297 podStartE2EDuration="4.19951068s" podCreationTimestamp="2026-04-17 20:55:11 +0000 UTC" firstStartedPulling="2026-04-17 20:55:11.644478147 +0000 UTC m=+637.118256412" lastFinishedPulling="2026-04-17 20:55:14.962124527 +0000 UTC m=+640.435902795" observedRunningTime="2026-04-17 20:55:15.197939337 +0000 UTC m=+640.671717624" watchObservedRunningTime="2026-04-17 20:55:15.19951068 +0000 UTC m=+640.673288967" Apr 17 20:55:15.211647 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:15.211571 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="opendatahub/maas-controller-6d4c8f55f9-fd4pp" podStartSLOduration=1.573779972 podStartE2EDuration="5.211557297s" podCreationTimestamp="2026-04-17 20:55:10 +0000 UTC" firstStartedPulling="2026-04-17 20:55:11.316358938 +0000 UTC m=+636.790137202" lastFinishedPulling="2026-04-17 20:55:14.954136262 +0000 UTC m=+640.427914527" observedRunningTime="2026-04-17 20:55:15.211120316 +0000 UTC m=+640.684898604" watchObservedRunningTime="2026-04-17 20:55:15.211557297 +0000 UTC m=+640.685335584" Apr 17 20:55:15.224089 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:15.224048 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="opendatahub/maas-controller-69646d8cdf-trqwh" podStartSLOduration=0.978557111 podStartE2EDuration="4.224038792s" podCreationTimestamp="2026-04-17 20:55:11 +0000 UTC" firstStartedPulling="2026-04-17 20:55:11.708709254 +0000 UTC m=+637.182487519" lastFinishedPulling="2026-04-17 20:55:14.954190933 +0000 UTC m=+640.427969200" observedRunningTime="2026-04-17 20:55:15.222537997 +0000 UTC m=+640.696316283" watchObservedRunningTime="2026-04-17 20:55:15.224038792 +0000 UTC m=+640.697817080" Apr 17 20:55:15.529268 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:15.529247 2577 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-6d4c8f55f9-fd4pp" Apr 17 20:55:15.577885 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:15.577852 2577 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmdbt\" (UniqueName: \"kubernetes.io/projected/4fccb7b6-35a5-4d1b-9d56-caeaf6ed9cc8-kube-api-access-xmdbt\") pod \"4fccb7b6-35a5-4d1b-9d56-caeaf6ed9cc8\" (UID: \"4fccb7b6-35a5-4d1b-9d56-caeaf6ed9cc8\") " Apr 17 20:55:15.579992 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:15.579971 2577 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fccb7b6-35a5-4d1b-9d56-caeaf6ed9cc8-kube-api-access-xmdbt" (OuterVolumeSpecName: "kube-api-access-xmdbt") pod "4fccb7b6-35a5-4d1b-9d56-caeaf6ed9cc8" (UID: "4fccb7b6-35a5-4d1b-9d56-caeaf6ed9cc8"). InnerVolumeSpecName "kube-api-access-xmdbt". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 17 20:55:15.679052 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:15.679026 2577 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-xmdbt\" (UniqueName: \"kubernetes.io/projected/4fccb7b6-35a5-4d1b-9d56-caeaf6ed9cc8-kube-api-access-xmdbt\") on node \"ip-10-0-137-110.ec2.internal\" DevicePath \"\"" Apr 17 20:55:16.188626 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:16.188589 2577 generic.go:358] "Generic (PLEG): container finished" podID="4fccb7b6-35a5-4d1b-9d56-caeaf6ed9cc8" containerID="caaaaf224da07fa827ec6c56d7600afb9ed193ab9fa5a2466ffb54db00d0b44c" exitCode=0 Apr 17 20:55:16.189018 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:16.188649 2577 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-6d4c8f55f9-fd4pp" Apr 17 20:55:16.189018 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:16.188659 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-6d4c8f55f9-fd4pp" event={"ID":"4fccb7b6-35a5-4d1b-9d56-caeaf6ed9cc8","Type":"ContainerDied","Data":"caaaaf224da07fa827ec6c56d7600afb9ed193ab9fa5a2466ffb54db00d0b44c"} Apr 17 20:55:16.189018 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:16.188696 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-6d4c8f55f9-fd4pp" event={"ID":"4fccb7b6-35a5-4d1b-9d56-caeaf6ed9cc8","Type":"ContainerDied","Data":"97af318f565ae74941eb196ae0f289968be386e9af04d69fd3bf624c4a199ccf"} Apr 17 20:55:16.189018 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:16.188712 2577 scope.go:117] "RemoveContainer" containerID="caaaaf224da07fa827ec6c56d7600afb9ed193ab9fa5a2466ffb54db00d0b44c" Apr 17 20:55:16.201928 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:16.201905 2577 scope.go:117] "RemoveContainer" containerID="caaaaf224da07fa827ec6c56d7600afb9ed193ab9fa5a2466ffb54db00d0b44c" Apr 17 20:55:16.202198 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:55:16.202175 2577 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"caaaaf224da07fa827ec6c56d7600afb9ed193ab9fa5a2466ffb54db00d0b44c\": container with ID starting with caaaaf224da07fa827ec6c56d7600afb9ed193ab9fa5a2466ffb54db00d0b44c not found: ID does not exist" containerID="caaaaf224da07fa827ec6c56d7600afb9ed193ab9fa5a2466ffb54db00d0b44c" Apr 17 20:55:16.202268 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:16.202210 2577 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"caaaaf224da07fa827ec6c56d7600afb9ed193ab9fa5a2466ffb54db00d0b44c"} err="failed to get container status \"caaaaf224da07fa827ec6c56d7600afb9ed193ab9fa5a2466ffb54db00d0b44c\": rpc error: code = NotFound desc = could not find container \"caaaaf224da07fa827ec6c56d7600afb9ed193ab9fa5a2466ffb54db00d0b44c\": container with ID starting with caaaaf224da07fa827ec6c56d7600afb9ed193ab9fa5a2466ffb54db00d0b44c not found: ID does not exist" Apr 17 20:55:16.210761 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:16.210739 2577 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["opendatahub/maas-controller-6d4c8f55f9-fd4pp"] Apr 17 20:55:16.212258 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:16.212236 2577 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["opendatahub/maas-controller-6d4c8f55f9-fd4pp"] Apr 17 20:55:17.121302 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:17.121268 2577 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fccb7b6-35a5-4d1b-9d56-caeaf6ed9cc8" path="/var/lib/kubelet/pods/4fccb7b6-35a5-4d1b-9d56-caeaf6ed9cc8/volumes" Apr 17 20:55:26.193991 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:26.193961 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="opendatahub/maas-controller-58cbf4b6f4-vkqzh" Apr 17 20:55:26.194398 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:26.194314 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="opendatahub/maas-controller-69646d8cdf-trqwh" Apr 17 20:55:26.242014 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:26.241988 2577 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["opendatahub/maas-controller-69646d8cdf-trqwh"] Apr 17 20:55:26.242164 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:26.242141 2577 kuberuntime_container.go:864] "Killing container with a grace period" pod="opendatahub/maas-controller-69646d8cdf-trqwh" podUID="4792930c-ec98-4b80-9284-cdbf5c515012" containerName="manager" containerID="cri-o://afeec21397d333eafe2b4938cb9a614e56f3d9301294edbcd313040d876d9b33" gracePeriod=10 Apr 17 20:55:26.502395 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:26.502372 2577 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-69646d8cdf-trqwh" Apr 17 20:55:26.560531 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:26.560503 2577 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58gln\" (UniqueName: \"kubernetes.io/projected/4792930c-ec98-4b80-9284-cdbf5c515012-kube-api-access-58gln\") pod \"4792930c-ec98-4b80-9284-cdbf5c515012\" (UID: \"4792930c-ec98-4b80-9284-cdbf5c515012\") " Apr 17 20:55:26.562599 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:26.562578 2577 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4792930c-ec98-4b80-9284-cdbf5c515012-kube-api-access-58gln" (OuterVolumeSpecName: "kube-api-access-58gln") pod "4792930c-ec98-4b80-9284-cdbf5c515012" (UID: "4792930c-ec98-4b80-9284-cdbf5c515012"). InnerVolumeSpecName "kube-api-access-58gln". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 17 20:55:26.662046 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:26.662015 2577 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-58gln\" (UniqueName: \"kubernetes.io/projected/4792930c-ec98-4b80-9284-cdbf5c515012-kube-api-access-58gln\") on node \"ip-10-0-137-110.ec2.internal\" DevicePath \"\"" Apr 17 20:55:27.223435 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:27.223406 2577 generic.go:358] "Generic (PLEG): container finished" podID="4792930c-ec98-4b80-9284-cdbf5c515012" containerID="afeec21397d333eafe2b4938cb9a614e56f3d9301294edbcd313040d876d9b33" exitCode=0 Apr 17 20:55:27.223824 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:27.223457 2577 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-69646d8cdf-trqwh" Apr 17 20:55:27.223824 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:27.223459 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-69646d8cdf-trqwh" event={"ID":"4792930c-ec98-4b80-9284-cdbf5c515012","Type":"ContainerDied","Data":"afeec21397d333eafe2b4938cb9a614e56f3d9301294edbcd313040d876d9b33"} Apr 17 20:55:27.223824 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:27.223572 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-69646d8cdf-trqwh" event={"ID":"4792930c-ec98-4b80-9284-cdbf5c515012","Type":"ContainerDied","Data":"5db7d58e70aa344f2dce5e30b3e8c3b2e6c6d169bd6363ffbdff468236fbe34b"} Apr 17 20:55:27.223824 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:27.223600 2577 scope.go:117] "RemoveContainer" containerID="afeec21397d333eafe2b4938cb9a614e56f3d9301294edbcd313040d876d9b33" Apr 17 20:55:27.231663 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:27.231647 2577 scope.go:117] "RemoveContainer" containerID="afeec21397d333eafe2b4938cb9a614e56f3d9301294edbcd313040d876d9b33" Apr 17 20:55:27.231929 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:55:27.231912 2577 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afeec21397d333eafe2b4938cb9a614e56f3d9301294edbcd313040d876d9b33\": container with ID starting with afeec21397d333eafe2b4938cb9a614e56f3d9301294edbcd313040d876d9b33 not found: ID does not exist" containerID="afeec21397d333eafe2b4938cb9a614e56f3d9301294edbcd313040d876d9b33" Apr 17 20:55:27.231993 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:27.231936 2577 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afeec21397d333eafe2b4938cb9a614e56f3d9301294edbcd313040d876d9b33"} err="failed to get container status \"afeec21397d333eafe2b4938cb9a614e56f3d9301294edbcd313040d876d9b33\": rpc error: code = NotFound desc = could not find container \"afeec21397d333eafe2b4938cb9a614e56f3d9301294edbcd313040d876d9b33\": container with ID starting with afeec21397d333eafe2b4938cb9a614e56f3d9301294edbcd313040d876d9b33 not found: ID does not exist" Apr 17 20:55:27.236729 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:27.236708 2577 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["opendatahub/maas-controller-69646d8cdf-trqwh"] Apr 17 20:55:27.239780 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:27.239760 2577 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["opendatahub/maas-controller-69646d8cdf-trqwh"] Apr 17 20:55:29.121004 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:29.120967 2577 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4792930c-ec98-4b80-9284-cdbf5c515012" path="/var/lib/kubelet/pods/4792930c-ec98-4b80-9284-cdbf5c515012/volumes" Apr 17 20:55:40.916548 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:40.916513 2577 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["opendatahub/maas-controller-58cbf4b6f4-vkqzh"] Apr 17 20:55:40.917059 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:40.916848 2577 kuberuntime_container.go:864] "Killing container with a grace period" pod="opendatahub/maas-controller-58cbf4b6f4-vkqzh" podUID="e804a8aa-e5fa-4219-808c-d537195b6319" containerName="manager" containerID="cri-o://2594f64fcb27b97b38ec9905f51a91d067b807dfdfb37b047d4947e176c09e51" gracePeriod=10 Apr 17 20:55:41.153646 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:41.153625 2577 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-58cbf4b6f4-vkqzh" Apr 17 20:55:41.265311 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:41.265288 2577 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cdz4\" (UniqueName: \"kubernetes.io/projected/e804a8aa-e5fa-4219-808c-d537195b6319-kube-api-access-8cdz4\") pod \"e804a8aa-e5fa-4219-808c-d537195b6319\" (UID: \"e804a8aa-e5fa-4219-808c-d537195b6319\") " Apr 17 20:55:41.267292 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:41.267265 2577 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e804a8aa-e5fa-4219-808c-d537195b6319-kube-api-access-8cdz4" (OuterVolumeSpecName: "kube-api-access-8cdz4") pod "e804a8aa-e5fa-4219-808c-d537195b6319" (UID: "e804a8aa-e5fa-4219-808c-d537195b6319"). InnerVolumeSpecName "kube-api-access-8cdz4". PluginName "kubernetes.io/projected", VolumeGIDValue "" Apr 17 20:55:41.268704 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:41.268680 2577 generic.go:358] "Generic (PLEG): container finished" podID="e804a8aa-e5fa-4219-808c-d537195b6319" containerID="2594f64fcb27b97b38ec9905f51a91d067b807dfdfb37b047d4947e176c09e51" exitCode=0 Apr 17 20:55:41.268833 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:41.268742 2577 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="opendatahub/maas-controller-58cbf4b6f4-vkqzh" Apr 17 20:55:41.268833 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:41.268759 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-58cbf4b6f4-vkqzh" event={"ID":"e804a8aa-e5fa-4219-808c-d537195b6319","Type":"ContainerDied","Data":"2594f64fcb27b97b38ec9905f51a91d067b807dfdfb37b047d4947e176c09e51"} Apr 17 20:55:41.268833 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:41.268787 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="opendatahub/maas-controller-58cbf4b6f4-vkqzh" event={"ID":"e804a8aa-e5fa-4219-808c-d537195b6319","Type":"ContainerDied","Data":"c04d756fba587d4618e370d001f335db6823c65eb6ec782d5f10df8740106a44"} Apr 17 20:55:41.268833 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:41.268817 2577 scope.go:117] "RemoveContainer" containerID="2594f64fcb27b97b38ec9905f51a91d067b807dfdfb37b047d4947e176c09e51" Apr 17 20:55:41.277729 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:41.277712 2577 scope.go:117] "RemoveContainer" containerID="2594f64fcb27b97b38ec9905f51a91d067b807dfdfb37b047d4947e176c09e51" Apr 17 20:55:41.277978 ip-10-0-137-110 kubenswrapper[2577]: E0417 20:55:41.277961 2577 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2594f64fcb27b97b38ec9905f51a91d067b807dfdfb37b047d4947e176c09e51\": container with ID starting with 2594f64fcb27b97b38ec9905f51a91d067b807dfdfb37b047d4947e176c09e51 not found: ID does not exist" containerID="2594f64fcb27b97b38ec9905f51a91d067b807dfdfb37b047d4947e176c09e51" Apr 17 20:55:41.278030 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:41.277984 2577 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2594f64fcb27b97b38ec9905f51a91d067b807dfdfb37b047d4947e176c09e51"} err="failed to get container status \"2594f64fcb27b97b38ec9905f51a91d067b807dfdfb37b047d4947e176c09e51\": rpc error: code = NotFound desc = could not find container \"2594f64fcb27b97b38ec9905f51a91d067b807dfdfb37b047d4947e176c09e51\": container with ID starting with 2594f64fcb27b97b38ec9905f51a91d067b807dfdfb37b047d4947e176c09e51 not found: ID does not exist" Apr 17 20:55:41.287929 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:41.287910 2577 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["opendatahub/maas-controller-58cbf4b6f4-vkqzh"] Apr 17 20:55:41.291719 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:41.291696 2577 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["opendatahub/maas-controller-58cbf4b6f4-vkqzh"] Apr 17 20:55:41.366426 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:41.366403 2577 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-8cdz4\" (UniqueName: \"kubernetes.io/projected/e804a8aa-e5fa-4219-808c-d537195b6319-kube-api-access-8cdz4\") on node \"ip-10-0-137-110.ec2.internal\" DevicePath \"\"" Apr 17 20:55:43.120121 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:55:43.120085 2577 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e804a8aa-e5fa-4219-808c-d537195b6319" path="/var/lib/kubelet/pods/e804a8aa-e5fa-4219-808c-d537195b6319/volumes" Apr 17 20:59:35.069939 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:59:35.069859 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-9rmnh_e562e939-99f0-4bee-952d-7af46132c860/console-operator/2.log" Apr 17 20:59:35.072123 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:59:35.072096 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-9rmnh_e562e939-99f0-4bee-952d-7af46132c860/console-operator/2.log" Apr 17 20:59:35.075761 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:59:35.075738 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dk97h_37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0/ovn-acl-logging/0.log" Apr 17 20:59:35.077765 ip-10-0-137-110 kubenswrapper[2577]: I0417 20:59:35.077744 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dk97h_37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0/ovn-acl-logging/0.log" Apr 17 21:04:35.093669 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:04:35.093634 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-9rmnh_e562e939-99f0-4bee-952d-7af46132c860/console-operator/2.log" Apr 17 21:04:35.096556 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:04:35.096535 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-9rmnh_e562e939-99f0-4bee-952d-7af46132c860/console-operator/2.log" Apr 17 21:04:35.099249 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:04:35.099228 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dk97h_37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0/ovn-acl-logging/0.log" Apr 17 21:04:35.101870 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:04:35.101850 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dk97h_37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0/ovn-acl-logging/0.log" Apr 17 21:05:48.620984 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:48.620943 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_authorino-56c57c5f84-gc6t4_6515bd08-2dbc-4177-8977-3ecdc4dcb979/authorino/0.log" Apr 17 21:05:51.942614 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:51.942576 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/opendatahub_opendatahub-operator-controller-manager-6dc4849f89-szktp_c5901a00-633e-4ca1-8035-4599be6e70d8/manager/0.log" Apr 17 21:05:52.046537 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:52.046508 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/opendatahub_postgres-868db5846d-jdhnx_36e1b7f5-63fe-40ed-8405-3a652226d92e/postgres/0.log" Apr 17 21:05:53.280473 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:53.280442 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_authorino-56c57c5f84-gc6t4_6515bd08-2dbc-4177-8977-3ecdc4dcb979/authorino/0.log" Apr 17 21:05:54.615648 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:54.615611 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_kube-auth-proxy-674746b5f4-8jnrh_e9f7fbd0-dec2-4736-877f-dde68502910c/kube-auth-proxy/0.log" Apr 17 21:05:58.547674 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:58.547645 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-must-gather-znlkx/must-gather-7xc8t"] Apr 17 21:05:58.548080 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:58.547991 2577 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="4792930c-ec98-4b80-9284-cdbf5c515012" containerName="manager" Apr 17 21:05:58.548080 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:58.548003 2577 state_mem.go:107] "Deleted CPUSet assignment" podUID="4792930c-ec98-4b80-9284-cdbf5c515012" containerName="manager" Apr 17 21:05:58.548080 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:58.548027 2577 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="4fccb7b6-35a5-4d1b-9d56-caeaf6ed9cc8" containerName="manager" Apr 17 21:05:58.548080 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:58.548033 2577 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fccb7b6-35a5-4d1b-9d56-caeaf6ed9cc8" containerName="manager" Apr 17 21:05:58.548080 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:58.548041 2577 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="e804a8aa-e5fa-4219-808c-d537195b6319" containerName="manager" Apr 17 21:05:58.548080 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:58.548047 2577 state_mem.go:107] "Deleted CPUSet assignment" podUID="e804a8aa-e5fa-4219-808c-d537195b6319" containerName="manager" Apr 17 21:05:58.548309 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:58.548107 2577 memory_manager.go:356] "RemoveStaleState removing state" podUID="e804a8aa-e5fa-4219-808c-d537195b6319" containerName="manager" Apr 17 21:05:58.548309 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:58.548119 2577 memory_manager.go:356] "RemoveStaleState removing state" podUID="4fccb7b6-35a5-4d1b-9d56-caeaf6ed9cc8" containerName="manager" Apr 17 21:05:58.548309 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:58.548125 2577 memory_manager.go:356] "RemoveStaleState removing state" podUID="4792930c-ec98-4b80-9284-cdbf5c515012" containerName="manager" Apr 17 21:05:58.551738 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:58.551710 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-znlkx/must-gather-7xc8t" Apr 17 21:05:58.554297 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:58.554266 2577 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-must-gather-znlkx\"/\"default-dockercfg-9snlf\"" Apr 17 21:05:58.554765 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:58.554739 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-znlkx\"/\"openshift-service-ca.crt\"" Apr 17 21:05:58.554950 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:58.554925 2577 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-znlkx\"/\"kube-root-ca.crt\"" Apr 17 21:05:58.565195 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:58.565172 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-znlkx/must-gather-7xc8t"] Apr 17 21:05:58.627170 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:58.627144 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/10e458ca-c89b-458d-b401-b98a5e35c33f-must-gather-output\") pod \"must-gather-7xc8t\" (UID: \"10e458ca-c89b-458d-b401-b98a5e35c33f\") " pod="openshift-must-gather-znlkx/must-gather-7xc8t" Apr 17 21:05:58.627294 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:58.627191 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2q5vv\" (UniqueName: \"kubernetes.io/projected/10e458ca-c89b-458d-b401-b98a5e35c33f-kube-api-access-2q5vv\") pod \"must-gather-7xc8t\" (UID: \"10e458ca-c89b-458d-b401-b98a5e35c33f\") " pod="openshift-must-gather-znlkx/must-gather-7xc8t" Apr 17 21:05:58.728046 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:58.728017 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-2q5vv\" (UniqueName: \"kubernetes.io/projected/10e458ca-c89b-458d-b401-b98a5e35c33f-kube-api-access-2q5vv\") pod \"must-gather-7xc8t\" (UID: \"10e458ca-c89b-458d-b401-b98a5e35c33f\") " pod="openshift-must-gather-znlkx/must-gather-7xc8t" Apr 17 21:05:58.728153 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:58.728108 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/10e458ca-c89b-458d-b401-b98a5e35c33f-must-gather-output\") pod \"must-gather-7xc8t\" (UID: \"10e458ca-c89b-458d-b401-b98a5e35c33f\") " pod="openshift-must-gather-znlkx/must-gather-7xc8t" Apr 17 21:05:58.728394 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:58.728376 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/10e458ca-c89b-458d-b401-b98a5e35c33f-must-gather-output\") pod \"must-gather-7xc8t\" (UID: \"10e458ca-c89b-458d-b401-b98a5e35c33f\") " pod="openshift-must-gather-znlkx/must-gather-7xc8t" Apr 17 21:05:58.735010 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:58.734990 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-2q5vv\" (UniqueName: \"kubernetes.io/projected/10e458ca-c89b-458d-b401-b98a5e35c33f-kube-api-access-2q5vv\") pod \"must-gather-7xc8t\" (UID: \"10e458ca-c89b-458d-b401-b98a5e35c33f\") " pod="openshift-must-gather-znlkx/must-gather-7xc8t" Apr 17 21:05:58.863159 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:58.863096 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-znlkx/must-gather-7xc8t" Apr 17 21:05:58.981086 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:58.981050 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-znlkx/must-gather-7xc8t"] Apr 17 21:05:58.983575 ip-10-0-137-110 kubenswrapper[2577]: W0417 21:05:58.983549 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10e458ca_c89b_458d_b401_b98a5e35c33f.slice/crio-732055bb1c274064060ab2588c4e598c5feab4b401331170c9475a60882b6cef WatchSource:0}: Error finding container 732055bb1c274064060ab2588c4e598c5feab4b401331170c9475a60882b6cef: Status 404 returned error can't find the container with id 732055bb1c274064060ab2588c4e598c5feab4b401331170c9475a60882b6cef Apr 17 21:05:58.985328 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:58.985312 2577 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Apr 17 21:05:59.309948 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:05:59.309919 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-znlkx/must-gather-7xc8t" event={"ID":"10e458ca-c89b-458d-b401-b98a5e35c33f","Type":"ContainerStarted","Data":"732055bb1c274064060ab2588c4e598c5feab4b401331170c9475a60882b6cef"} Apr 17 21:06:00.315471 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:00.315381 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-znlkx/must-gather-7xc8t" event={"ID":"10e458ca-c89b-458d-b401-b98a5e35c33f","Type":"ContainerStarted","Data":"06aff8ede42340f48a597e08333219de680a79bae8d5d5b2a3f8aa805473d038"} Apr 17 21:06:00.315471 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:00.315423 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-znlkx/must-gather-7xc8t" event={"ID":"10e458ca-c89b-458d-b401-b98a5e35c33f","Type":"ContainerStarted","Data":"0c1fad8b99f147cc4106452374d5b1c7c1c364024e7afe2b90894d7b93bfddb8"} Apr 17 21:06:00.328242 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:00.328191 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-znlkx/must-gather-7xc8t" podStartSLOduration=1.510735959 podStartE2EDuration="2.328173462s" podCreationTimestamp="2026-04-17 21:05:58 +0000 UTC" firstStartedPulling="2026-04-17 21:05:58.985465804 +0000 UTC m=+1284.459244069" lastFinishedPulling="2026-04-17 21:05:59.802903304 +0000 UTC m=+1285.276681572" observedRunningTime="2026-04-17 21:06:00.328039236 +0000 UTC m=+1285.801817525" watchObservedRunningTime="2026-04-17 21:06:00.328173462 +0000 UTC m=+1285.801951751" Apr 17 21:06:01.343244 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:01.343209 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_global-pull-secret-syncer-flcph_103da256-764a-4e19-8698-330c81744ce0/global-pull-secret-syncer/0.log" Apr 17 21:06:01.444893 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:01.444860 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_konnectivity-agent-cxc79_906a412c-f8af-4800-9e64-23e1b3be650e/konnectivity-agent/0.log" Apr 17 21:06:01.511044 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:01.511013 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_kube-apiserver-proxy-ip-10-0-137-110.ec2.internal_3114a7bb68dc8d0f555ef7f4a1089352/haproxy/0.log" Apr 17 21:06:05.426206 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:05.426170 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/kuadrant-system_authorino-56c57c5f84-gc6t4_6515bd08-2dbc-4177-8977-3ecdc4dcb979/authorino/0.log" Apr 17 21:06:07.090784 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:07.090706 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_cluster-monitoring-operator-75587bd455-7cqn6_5506563e-216d-4b61-97dc-8cea5be3f474/cluster-monitoring-operator/0.log" Apr 17 21:06:07.202328 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:07.202274 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_metrics-server-5c998b9cc6-kqcc5_595b0fbe-52fe-4bf3-be9d-8249e0ca08aa/metrics-server/0.log" Apr 17 21:06:07.336261 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:07.336237 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-lpkzb_a044b44f-06c8-425f-9f9f-24aaeddea464/node-exporter/0.log" Apr 17 21:06:07.362779 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:07.362709 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-lpkzb_a044b44f-06c8-425f-9f9f-24aaeddea464/kube-rbac-proxy/0.log" Apr 17 21:06:07.382664 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:07.382530 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-lpkzb_a044b44f-06c8-425f-9f9f-24aaeddea464/init-textfile/0.log" Apr 17 21:06:07.576669 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:07.576642 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_356df3f8-dff9-4963-a883-fe17c0c75d95/prometheus/0.log" Apr 17 21:06:07.599497 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:07.599466 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_356df3f8-dff9-4963-a883-fe17c0c75d95/config-reloader/0.log" Apr 17 21:06:07.617394 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:07.617306 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_356df3f8-dff9-4963-a883-fe17c0c75d95/thanos-sidecar/0.log" Apr 17 21:06:07.637613 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:07.637583 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_356df3f8-dff9-4963-a883-fe17c0c75d95/kube-rbac-proxy-web/0.log" Apr 17 21:06:07.659566 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:07.659516 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_356df3f8-dff9-4963-a883-fe17c0c75d95/kube-rbac-proxy/0.log" Apr 17 21:06:07.680534 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:07.680497 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_356df3f8-dff9-4963-a883-fe17c0c75d95/kube-rbac-proxy-thanos/0.log" Apr 17 21:06:07.701854 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:07.701823 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_356df3f8-dff9-4963-a883-fe17c0c75d95/init-config-reloader/0.log" Apr 17 21:06:09.699379 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:09.699346 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-9rmnh_e562e939-99f0-4bee-952d-7af46132c860/console-operator/2.log" Apr 17 21:06:09.708136 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:09.708101 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-9d4b6777b-9rmnh_e562e939-99f0-4bee-952d-7af46132c860/console-operator/3.log" Apr 17 21:06:09.905905 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:09.905860 2577 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-must-gather-znlkx/perf-node-gather-daemonset-cnlpb"] Apr 17 21:06:09.909598 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:09.909570 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-znlkx/perf-node-gather-daemonset-cnlpb" Apr 17 21:06:09.920181 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:09.920154 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-znlkx/perf-node-gather-daemonset-cnlpb"] Apr 17 21:06:09.946831 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:09.946777 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmhx5\" (UniqueName: \"kubernetes.io/projected/bb435c37-77ef-49d2-838d-7b1172c67fba-kube-api-access-rmhx5\") pod \"perf-node-gather-daemonset-cnlpb\" (UID: \"bb435c37-77ef-49d2-838d-7b1172c67fba\") " pod="openshift-must-gather-znlkx/perf-node-gather-daemonset-cnlpb" Apr 17 21:06:09.947087 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:09.947066 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/bb435c37-77ef-49d2-838d-7b1172c67fba-podres\") pod \"perf-node-gather-daemonset-cnlpb\" (UID: \"bb435c37-77ef-49d2-838d-7b1172c67fba\") " pod="openshift-must-gather-znlkx/perf-node-gather-daemonset-cnlpb" Apr 17 21:06:09.947242 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:09.947225 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bb435c37-77ef-49d2-838d-7b1172c67fba-lib-modules\") pod \"perf-node-gather-daemonset-cnlpb\" (UID: \"bb435c37-77ef-49d2-838d-7b1172c67fba\") " pod="openshift-must-gather-znlkx/perf-node-gather-daemonset-cnlpb" Apr 17 21:06:09.947384 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:09.947371 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bb435c37-77ef-49d2-838d-7b1172c67fba-sys\") pod \"perf-node-gather-daemonset-cnlpb\" (UID: \"bb435c37-77ef-49d2-838d-7b1172c67fba\") " pod="openshift-must-gather-znlkx/perf-node-gather-daemonset-cnlpb" Apr 17 21:06:09.947486 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:09.947471 2577 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/bb435c37-77ef-49d2-838d-7b1172c67fba-proc\") pod \"perf-node-gather-daemonset-cnlpb\" (UID: \"bb435c37-77ef-49d2-838d-7b1172c67fba\") " pod="openshift-must-gather-znlkx/perf-node-gather-daemonset-cnlpb" Apr 17 21:06:10.048727 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:10.048651 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bb435c37-77ef-49d2-838d-7b1172c67fba-sys\") pod \"perf-node-gather-daemonset-cnlpb\" (UID: \"bb435c37-77ef-49d2-838d-7b1172c67fba\") " pod="openshift-must-gather-znlkx/perf-node-gather-daemonset-cnlpb" Apr 17 21:06:10.048727 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:10.048686 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/bb435c37-77ef-49d2-838d-7b1172c67fba-proc\") pod \"perf-node-gather-daemonset-cnlpb\" (UID: \"bb435c37-77ef-49d2-838d-7b1172c67fba\") " pod="openshift-must-gather-znlkx/perf-node-gather-daemonset-cnlpb" Apr 17 21:06:10.048961 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:10.048744 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rmhx5\" (UniqueName: \"kubernetes.io/projected/bb435c37-77ef-49d2-838d-7b1172c67fba-kube-api-access-rmhx5\") pod \"perf-node-gather-daemonset-cnlpb\" (UID: \"bb435c37-77ef-49d2-838d-7b1172c67fba\") " pod="openshift-must-gather-znlkx/perf-node-gather-daemonset-cnlpb" Apr 17 21:06:10.048961 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:10.048770 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bb435c37-77ef-49d2-838d-7b1172c67fba-sys\") pod \"perf-node-gather-daemonset-cnlpb\" (UID: \"bb435c37-77ef-49d2-838d-7b1172c67fba\") " pod="openshift-must-gather-znlkx/perf-node-gather-daemonset-cnlpb" Apr 17 21:06:10.048961 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:10.048831 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/bb435c37-77ef-49d2-838d-7b1172c67fba-proc\") pod \"perf-node-gather-daemonset-cnlpb\" (UID: \"bb435c37-77ef-49d2-838d-7b1172c67fba\") " pod="openshift-must-gather-znlkx/perf-node-gather-daemonset-cnlpb" Apr 17 21:06:10.048961 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:10.048857 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/bb435c37-77ef-49d2-838d-7b1172c67fba-podres\") pod \"perf-node-gather-daemonset-cnlpb\" (UID: \"bb435c37-77ef-49d2-838d-7b1172c67fba\") " pod="openshift-must-gather-znlkx/perf-node-gather-daemonset-cnlpb" Apr 17 21:06:10.048961 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:10.048916 2577 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bb435c37-77ef-49d2-838d-7b1172c67fba-lib-modules\") pod \"perf-node-gather-daemonset-cnlpb\" (UID: \"bb435c37-77ef-49d2-838d-7b1172c67fba\") " pod="openshift-must-gather-znlkx/perf-node-gather-daemonset-cnlpb" Apr 17 21:06:10.049238 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:10.049029 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/bb435c37-77ef-49d2-838d-7b1172c67fba-podres\") pod \"perf-node-gather-daemonset-cnlpb\" (UID: \"bb435c37-77ef-49d2-838d-7b1172c67fba\") " pod="openshift-must-gather-znlkx/perf-node-gather-daemonset-cnlpb" Apr 17 21:06:10.049238 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:10.049073 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bb435c37-77ef-49d2-838d-7b1172c67fba-lib-modules\") pod \"perf-node-gather-daemonset-cnlpb\" (UID: \"bb435c37-77ef-49d2-838d-7b1172c67fba\") " pod="openshift-must-gather-znlkx/perf-node-gather-daemonset-cnlpb" Apr 17 21:06:10.057323 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:10.057300 2577 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmhx5\" (UniqueName: \"kubernetes.io/projected/bb435c37-77ef-49d2-838d-7b1172c67fba-kube-api-access-rmhx5\") pod \"perf-node-gather-daemonset-cnlpb\" (UID: \"bb435c37-77ef-49d2-838d-7b1172c67fba\") " pod="openshift-must-gather-znlkx/perf-node-gather-daemonset-cnlpb" Apr 17 21:06:10.223991 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:10.223953 2577 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-znlkx/perf-node-gather-daemonset-cnlpb" Apr 17 21:06:10.403079 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:10.403033 2577 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-znlkx/perf-node-gather-daemonset-cnlpb"] Apr 17 21:06:10.404569 ip-10-0-137-110 kubenswrapper[2577]: W0417 21:06:10.404537 2577 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podbb435c37_77ef_49d2_838d_7b1172c67fba.slice/crio-137dbae5c63b1345f62b6fbf0bc9e44ed12bc2927919c8f2966a59f7779f77b1 WatchSource:0}: Error finding container 137dbae5c63b1345f62b6fbf0bc9e44ed12bc2927919c8f2966a59f7779f77b1: Status 404 returned error can't find the container with id 137dbae5c63b1345f62b6fbf0bc9e44ed12bc2927919c8f2966a59f7779f77b1 Apr 17 21:06:10.753595 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:10.753566 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-storage-operator_volume-data-source-validator-7c6cbb6c87-dhmj2_2cfc1b29-dfac-4fb3-878c-0961d357f28e/volume-data-source-validator/0.log" Apr 17 21:06:11.366501 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:11.366470 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-znlkx/perf-node-gather-daemonset-cnlpb" event={"ID":"bb435c37-77ef-49d2-838d-7b1172c67fba","Type":"ContainerStarted","Data":"5f599113ccca7c739c7176ce842dbb5afa8b75f27fae387e9bb3e0d61d18cd94"} Apr 17 21:06:11.366501 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:11.366504 2577 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-znlkx/perf-node-gather-daemonset-cnlpb" event={"ID":"bb435c37-77ef-49d2-838d-7b1172c67fba","Type":"ContainerStarted","Data":"137dbae5c63b1345f62b6fbf0bc9e44ed12bc2927919c8f2966a59f7779f77b1"} Apr 17 21:06:11.366750 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:11.366558 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-must-gather-znlkx/perf-node-gather-daemonset-cnlpb" Apr 17 21:06:11.383496 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:11.383437 2577 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-znlkx/perf-node-gather-daemonset-cnlpb" podStartSLOduration=2.383419573 podStartE2EDuration="2.383419573s" podCreationTimestamp="2026-04-17 21:06:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-04-17 21:06:11.381164278 +0000 UTC m=+1296.854942567" watchObservedRunningTime="2026-04-17 21:06:11.383419573 +0000 UTC m=+1296.857197862" Apr 17 21:06:11.514864 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:11.514834 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_dns-default-bc6wf_821ba3d4-448f-45b9-bcec-701682592a0f/dns/0.log" Apr 17 21:06:11.532394 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:11.532371 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_dns-default-bc6wf_821ba3d4-448f-45b9-bcec-701682592a0f/kube-rbac-proxy/0.log" Apr 17 21:06:11.657091 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:11.657022 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_node-resolver-cwp9d_c9f9af16-3b9f-4ebb-8025-9b9162c7eaa1/dns-node-resolver/0.log" Apr 17 21:06:12.098742 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:12.098710 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_image-registry-57fdff8cb7-8qxnn_d0f3d88e-5398-4272-bf50-7675902545aa/registry/0.log" Apr 17 21:06:12.136016 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:12.135990 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_node-ca-v8fbw_5fa81115-694f-4f2f-8797-8962d102b7e7/node-ca/0.log" Apr 17 21:06:12.957713 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:12.957681 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_kube-auth-proxy-674746b5f4-8jnrh_e9f7fbd0-dec2-4736-877f-dde68502910c/kube-auth-proxy/0.log" Apr 17 21:06:13.545583 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:13.545546 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-canary_ingress-canary-q59g7_79b05e51-04fa-4473-b612-c762aa0e3856/serve-healthcheck-canary/0.log" Apr 17 21:06:13.956204 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:13.956165 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-operator-585dfdc468-5jzh4_18ff00e6-c4f7-4852-b689-94ec992eb400/insights-operator/0.log" Apr 17 21:06:13.956351 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:13.956224 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-operator-585dfdc468-5jzh4_18ff00e6-c4f7-4852-b689-94ec992eb400/insights-operator/1.log" Apr 17 21:06:14.112186 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:14.112156 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-rnnj4_4e8e0c9a-eac8-4af9-bd4c-927f0b406f02/kube-rbac-proxy/0.log" Apr 17 21:06:14.130071 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:14.130048 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-rnnj4_4e8e0c9a-eac8-4af9-bd4c-927f0b406f02/exporter/0.log" Apr 17 21:06:14.148854 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:14.148833 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-rnnj4_4e8e0c9a-eac8-4af9-bd4c-927f0b406f02/extractor/0.log" Apr 17 21:06:16.045994 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:16.045963 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/opendatahub_opendatahub-operator-controller-manager-6dc4849f89-szktp_c5901a00-633e-4ca1-8035-4599be6e70d8/manager/0.log" Apr 17 21:06:16.063919 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:16.063883 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/opendatahub_postgres-868db5846d-jdhnx_36e1b7f5-63fe-40ed-8405-3a652226d92e/postgres/0.log" Apr 17 21:06:17.071474 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:17.071446 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-lws-operator_lws-controller-manager-7f68665c84-2nrp7_926fd6e1-ce48-4774-be2c-5280dfdc85ad/manager/0.log" Apr 17 21:06:17.381285 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:17.381219 2577 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-must-gather-znlkx/perf-node-gather-daemonset-cnlpb" Apr 17 21:06:21.375651 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:21.375618 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator_migrator-74bb7799d9-wjmrh_944799f6-55a3-4bc4-a6e0-3a72a00ae117/migrator/0.log" Apr 17 21:06:21.392199 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:21.392176 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator_migrator-74bb7799d9-wjmrh_944799f6-55a3-4bc4-a6e0-3a72a00ae117/graceful-termination/0.log" Apr 17 21:06:21.726161 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:21.726132 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator-operator_kube-storage-version-migrator-operator-6769c5d45-fs4ts_05c5418a-43f4-476b-8e92-b7b56513d5af/kube-storage-version-migrator-operator/1.log" Apr 17 21:06:21.728315 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:21.728284 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator-operator_kube-storage-version-migrator-operator-6769c5d45-fs4ts_05c5418a-43f4-476b-8e92-b7b56513d5af/kube-storage-version-migrator-operator/0.log" Apr 17 21:06:22.650374 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:22.650341 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2lsfw_4b68ff3e-8a24-45e1-a70f-83f0539a9623/kube-multus/0.log" Apr 17 21:06:22.971789 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:22.971767 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-zds6c_06332f89-1179-4cde-9a2b-4335714933a1/kube-multus-additional-cni-plugins/0.log" Apr 17 21:06:22.988665 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:22.988643 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-zds6c_06332f89-1179-4cde-9a2b-4335714933a1/egress-router-binary-copy/0.log" Apr 17 21:06:23.007969 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:23.007935 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-zds6c_06332f89-1179-4cde-9a2b-4335714933a1/cni-plugins/0.log" Apr 17 21:06:23.024748 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:23.024726 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-zds6c_06332f89-1179-4cde-9a2b-4335714933a1/bond-cni-plugin/0.log" Apr 17 21:06:23.043380 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:23.043363 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-zds6c_06332f89-1179-4cde-9a2b-4335714933a1/routeoverride-cni/0.log" Apr 17 21:06:23.060218 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:23.060194 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-zds6c_06332f89-1179-4cde-9a2b-4335714933a1/whereabouts-cni-bincopy/0.log" Apr 17 21:06:23.077775 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:23.077757 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-zds6c_06332f89-1179-4cde-9a2b-4335714933a1/whereabouts-cni/0.log" Apr 17 21:06:23.127221 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:23.127205 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-7ddt5_1a6721ad-2d96-4d01-a9db-b8654256e62f/network-metrics-daemon/0.log" Apr 17 21:06:23.154973 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:23.154954 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-7ddt5_1a6721ad-2d96-4d01-a9db-b8654256e62f/kube-rbac-proxy/0.log" Apr 17 21:06:24.017826 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:24.017757 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dk97h_37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0/ovn-controller/0.log" Apr 17 21:06:24.031976 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:24.031952 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dk97h_37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0/ovn-acl-logging/0.log" Apr 17 21:06:24.044871 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:24.044842 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dk97h_37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0/ovn-acl-logging/1.log" Apr 17 21:06:24.063633 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:24.063616 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dk97h_37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0/kube-rbac-proxy-node/0.log" Apr 17 21:06:24.081997 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:24.081974 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dk97h_37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0/kube-rbac-proxy-ovn-metrics/0.log" Apr 17 21:06:24.096466 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:24.096448 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dk97h_37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0/northd/0.log" Apr 17 21:06:24.114327 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:24.114312 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dk97h_37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0/nbdb/0.log" Apr 17 21:06:24.131641 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:24.131619 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dk97h_37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0/sbdb/0.log" Apr 17 21:06:24.303698 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:24.303632 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dk97h_37eb787b-5bf1-4d2b-b68e-5b7e766fe8e0/ovnkube-controller/0.log" Apr 17 21:06:25.826698 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:25.826663 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-8894fc9bd-z74x2_16649afe-84e4-4bed-b4c3-a24055c6af4d/check-endpoints/0.log" Apr 17 21:06:25.870033 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:25.869998 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-target-cm2lg_ac379788-302d-4181-8970-56f47ae312c1/network-check-target-container/0.log" Apr 17 21:06:26.869188 ip-10-0-137-110 kubenswrapper[2577]: I0417 21:06:26.869155 2577 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-operator_iptables-alerter-mkzlr_2e78f338-82b0-4062-8438-2b45fae0f4cf/iptables-alerter/0.log"