Feb 17 12:46:33.543402 ip-10-0-131-118 systemd[1]: Starting Kubernetes Kubelet... Feb 17 12:46:33.948574 ip-10-0-131-118 kubenswrapper[2572]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 17 12:46:33.948574 ip-10-0-131-118 kubenswrapper[2572]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Feb 17 12:46:33.948574 ip-10-0-131-118 kubenswrapper[2572]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 17 12:46:33.948574 ip-10-0-131-118 kubenswrapper[2572]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Feb 17 12:46:33.948574 ip-10-0-131-118 kubenswrapper[2572]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 17 12:46:33.950981 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.950895 2572 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 17 12:46:33.955270 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955249 2572 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 17 12:46:33.955270 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955267 2572 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Feb 17 12:46:33.955270 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955271 2572 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Feb 17 12:46:33.955270 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955275 2572 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Feb 17 12:46:33.955270 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955278 2572 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Feb 17 12:46:33.955463 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955285 2572 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Feb 17 12:46:33.955463 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955290 2572 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Feb 17 12:46:33.955463 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955295 2572 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Feb 17 12:46:33.955463 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955299 2572 feature_gate.go:328] unrecognized feature gate: SignatureStores Feb 17 12:46:33.955463 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955302 2572 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Feb 17 12:46:33.955463 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955306 2572 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Feb 17 12:46:33.955463 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955309 2572 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Feb 17 12:46:33.955463 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955312 2572 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Feb 17 12:46:33.955463 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955315 2572 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Feb 17 12:46:33.955463 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955318 2572 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Feb 17 12:46:33.955463 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955321 2572 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Feb 17 12:46:33.955463 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955323 2572 feature_gate.go:328] unrecognized feature gate: GatewayAPI Feb 17 12:46:33.955463 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955326 2572 feature_gate.go:328] unrecognized feature gate: Example Feb 17 12:46:33.955463 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955332 2572 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Feb 17 12:46:33.955463 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955334 2572 feature_gate.go:328] unrecognized feature gate: DualReplica Feb 17 12:46:33.955463 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955337 2572 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Feb 17 12:46:33.955463 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955340 2572 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Feb 17 12:46:33.955463 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955342 2572 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Feb 17 12:46:33.955463 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955345 2572 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Feb 17 12:46:33.955463 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955348 2572 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Feb 17 12:46:33.955950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955350 2572 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Feb 17 12:46:33.955950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955353 2572 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Feb 17 12:46:33.955950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955356 2572 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Feb 17 12:46:33.955950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955359 2572 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Feb 17 12:46:33.955950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955362 2572 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Feb 17 12:46:33.955950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955368 2572 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Feb 17 12:46:33.955950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955371 2572 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Feb 17 12:46:33.955950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955374 2572 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Feb 17 12:46:33.955950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955376 2572 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Feb 17 12:46:33.955950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955379 2572 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Feb 17 12:46:33.955950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955382 2572 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Feb 17 12:46:33.955950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955384 2572 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 17 12:46:33.955950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955387 2572 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Feb 17 12:46:33.955950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955391 2572 feature_gate.go:328] unrecognized feature gate: InsightsConfig Feb 17 12:46:33.955950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955394 2572 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Feb 17 12:46:33.955950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955398 2572 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Feb 17 12:46:33.955950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955400 2572 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Feb 17 12:46:33.955950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955406 2572 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Feb 17 12:46:33.955950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955408 2572 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 17 12:46:33.956461 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955411 2572 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 17 12:46:33.956461 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955414 2572 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Feb 17 12:46:33.956461 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955417 2572 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Feb 17 12:46:33.956461 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955420 2572 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Feb 17 12:46:33.956461 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955422 2572 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Feb 17 12:46:33.956461 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955425 2572 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Feb 17 12:46:33.956461 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955428 2572 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Feb 17 12:46:33.956461 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955430 2572 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Feb 17 12:46:33.956461 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955433 2572 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Feb 17 12:46:33.956461 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955436 2572 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Feb 17 12:46:33.956461 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955438 2572 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Feb 17 12:46:33.956461 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955443 2572 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Feb 17 12:46:33.956461 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955446 2572 feature_gate.go:328] unrecognized feature gate: PinnedImages Feb 17 12:46:33.956461 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955449 2572 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Feb 17 12:46:33.956461 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955451 2572 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Feb 17 12:46:33.956461 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955454 2572 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Feb 17 12:46:33.956461 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955457 2572 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Feb 17 12:46:33.956461 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955460 2572 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 17 12:46:33.956461 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955463 2572 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Feb 17 12:46:33.956951 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955465 2572 feature_gate.go:328] unrecognized feature gate: OVNObservability Feb 17 12:46:33.956951 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955468 2572 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Feb 17 12:46:33.956951 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955471 2572 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Feb 17 12:46:33.956951 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955474 2572 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Feb 17 12:46:33.956951 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955477 2572 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Feb 17 12:46:33.956951 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955482 2572 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Feb 17 12:46:33.956951 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955485 2572 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Feb 17 12:46:33.956951 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955489 2572 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Feb 17 12:46:33.956951 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955492 2572 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Feb 17 12:46:33.956951 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955495 2572 feature_gate.go:328] unrecognized feature gate: Example2 Feb 17 12:46:33.956951 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955498 2572 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Feb 17 12:46:33.956951 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955500 2572 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Feb 17 12:46:33.956951 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955503 2572 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 17 12:46:33.956951 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955506 2572 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Feb 17 12:46:33.956951 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955508 2572 feature_gate.go:328] unrecognized feature gate: NewOLM Feb 17 12:46:33.956951 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955511 2572 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Feb 17 12:46:33.956951 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955514 2572 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Feb 17 12:46:33.956951 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955520 2572 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Feb 17 12:46:33.956951 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955523 2572 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Feb 17 12:46:33.956951 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955526 2572 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Feb 17 12:46:33.957581 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955530 2572 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 17 12:46:33.957581 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955534 2572 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Feb 17 12:46:33.957581 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.955537 2572 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Feb 17 12:46:33.957581 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956168 2572 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Feb 17 12:46:33.957581 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956174 2572 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 17 12:46:33.957581 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956178 2572 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Feb 17 12:46:33.957581 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956181 2572 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 17 12:46:33.957581 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956184 2572 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Feb 17 12:46:33.957581 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956187 2572 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Feb 17 12:46:33.957581 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956190 2572 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Feb 17 12:46:33.957581 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956194 2572 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Feb 17 12:46:33.957581 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956199 2572 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Feb 17 12:46:33.957581 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956202 2572 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Feb 17 12:46:33.957581 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956206 2572 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Feb 17 12:46:33.957581 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956208 2572 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Feb 17 12:46:33.957581 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956211 2572 feature_gate.go:328] unrecognized feature gate: Example Feb 17 12:46:33.957581 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956214 2572 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 17 12:46:33.957581 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956217 2572 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Feb 17 12:46:33.957581 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956219 2572 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Feb 17 12:46:33.958048 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956223 2572 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Feb 17 12:46:33.958048 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956226 2572 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 17 12:46:33.958048 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956229 2572 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Feb 17 12:46:33.958048 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956232 2572 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Feb 17 12:46:33.958048 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956235 2572 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Feb 17 12:46:33.958048 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956240 2572 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Feb 17 12:46:33.958048 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956243 2572 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Feb 17 12:46:33.958048 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956247 2572 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Feb 17 12:46:33.958048 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956252 2572 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Feb 17 12:46:33.958048 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956255 2572 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Feb 17 12:46:33.958048 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956259 2572 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Feb 17 12:46:33.958048 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956262 2572 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Feb 17 12:46:33.958048 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956265 2572 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Feb 17 12:46:33.958048 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956268 2572 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Feb 17 12:46:33.958048 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956270 2572 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Feb 17 12:46:33.958048 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956273 2572 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Feb 17 12:46:33.958048 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956276 2572 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Feb 17 12:46:33.958048 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956278 2572 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Feb 17 12:46:33.958048 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956283 2572 feature_gate.go:328] unrecognized feature gate: DualReplica Feb 17 12:46:33.958532 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956287 2572 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Feb 17 12:46:33.958532 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956290 2572 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Feb 17 12:46:33.958532 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956293 2572 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Feb 17 12:46:33.958532 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956295 2572 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Feb 17 12:46:33.958532 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956298 2572 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Feb 17 12:46:33.958532 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956302 2572 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Feb 17 12:46:33.958532 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956305 2572 feature_gate.go:328] unrecognized feature gate: InsightsConfig Feb 17 12:46:33.958532 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956308 2572 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Feb 17 12:46:33.958532 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956310 2572 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Feb 17 12:46:33.958532 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956313 2572 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Feb 17 12:46:33.958532 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956318 2572 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 17 12:46:33.958532 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956321 2572 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Feb 17 12:46:33.958532 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956324 2572 feature_gate.go:328] unrecognized feature gate: NewOLM Feb 17 12:46:33.958532 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956326 2572 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Feb 17 12:46:33.958532 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956329 2572 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Feb 17 12:46:33.958532 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956332 2572 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Feb 17 12:46:33.958532 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956335 2572 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Feb 17 12:46:33.958532 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956338 2572 feature_gate.go:328] unrecognized feature gate: PinnedImages Feb 17 12:46:33.958532 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956340 2572 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Feb 17 12:46:33.958532 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956343 2572 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Feb 17 12:46:33.959030 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956346 2572 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Feb 17 12:46:33.959030 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956348 2572 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 17 12:46:33.959030 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956351 2572 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Feb 17 12:46:33.959030 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956356 2572 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Feb 17 12:46:33.959030 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956359 2572 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 17 12:46:33.959030 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956361 2572 feature_gate.go:328] unrecognized feature gate: GatewayAPI Feb 17 12:46:33.959030 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956365 2572 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Feb 17 12:46:33.959030 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956368 2572 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Feb 17 12:46:33.959030 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956370 2572 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Feb 17 12:46:33.959030 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956373 2572 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Feb 17 12:46:33.959030 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956376 2572 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Feb 17 12:46:33.959030 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956379 2572 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Feb 17 12:46:33.959030 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956381 2572 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Feb 17 12:46:33.959030 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956384 2572 feature_gate.go:328] unrecognized feature gate: Example2 Feb 17 12:46:33.959030 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956387 2572 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Feb 17 12:46:33.959030 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956392 2572 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Feb 17 12:46:33.959030 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956395 2572 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Feb 17 12:46:33.959030 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956398 2572 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Feb 17 12:46:33.959030 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956401 2572 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Feb 17 12:46:33.959030 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956404 2572 feature_gate.go:328] unrecognized feature gate: SignatureStores Feb 17 12:46:33.959529 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956406 2572 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Feb 17 12:46:33.959529 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956409 2572 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Feb 17 12:46:33.959529 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956412 2572 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Feb 17 12:46:33.959529 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956414 2572 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Feb 17 12:46:33.959529 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956417 2572 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Feb 17 12:46:33.959529 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956420 2572 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Feb 17 12:46:33.959529 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956422 2572 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Feb 17 12:46:33.959529 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956425 2572 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Feb 17 12:46:33.959529 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956430 2572 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Feb 17 12:46:33.959529 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956786 2572 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Feb 17 12:46:33.959529 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.956790 2572 feature_gate.go:328] unrecognized feature gate: OVNObservability Feb 17 12:46:33.959529 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957831 2572 flags.go:64] FLAG: --address="0.0.0.0" Feb 17 12:46:33.959529 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957850 2572 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Feb 17 12:46:33.959529 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957858 2572 flags.go:64] FLAG: --anonymous-auth="true" Feb 17 12:46:33.959529 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957864 2572 flags.go:64] FLAG: --application-metrics-count-limit="100" Feb 17 12:46:33.959529 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957869 2572 flags.go:64] FLAG: --authentication-token-webhook="false" Feb 17 12:46:33.959529 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957872 2572 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Feb 17 12:46:33.959529 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957877 2572 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Feb 17 12:46:33.959529 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957882 2572 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Feb 17 12:46:33.959529 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957885 2572 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Feb 17 12:46:33.959529 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957889 2572 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Feb 17 12:46:33.960058 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957892 2572 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Feb 17 12:46:33.960058 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957895 2572 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Feb 17 12:46:33.960058 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957898 2572 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Feb 17 12:46:33.960058 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957902 2572 flags.go:64] FLAG: --cgroup-root="" Feb 17 12:46:33.960058 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957905 2572 flags.go:64] FLAG: --cgroups-per-qos="true" Feb 17 12:46:33.960058 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957908 2572 flags.go:64] FLAG: --client-ca-file="" Feb 17 12:46:33.960058 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957911 2572 flags.go:64] FLAG: --cloud-config="" Feb 17 12:46:33.960058 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957913 2572 flags.go:64] FLAG: --cloud-provider="external" Feb 17 12:46:33.960058 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957916 2572 flags.go:64] FLAG: --cluster-dns="[]" Feb 17 12:46:33.960058 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957921 2572 flags.go:64] FLAG: --cluster-domain="" Feb 17 12:46:33.960058 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957924 2572 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Feb 17 12:46:33.960058 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957927 2572 flags.go:64] FLAG: --config-dir="" Feb 17 12:46:33.960058 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957930 2572 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Feb 17 12:46:33.960058 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957933 2572 flags.go:64] FLAG: --container-log-max-files="5" Feb 17 12:46:33.960058 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957937 2572 flags.go:64] FLAG: --container-log-max-size="10Mi" Feb 17 12:46:33.960058 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957941 2572 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Feb 17 12:46:33.960058 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957944 2572 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Feb 17 12:46:33.960058 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957948 2572 flags.go:64] FLAG: --containerd-namespace="k8s.io" Feb 17 12:46:33.960058 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957951 2572 flags.go:64] FLAG: --contention-profiling="false" Feb 17 12:46:33.960058 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957954 2572 flags.go:64] FLAG: --cpu-cfs-quota="true" Feb 17 12:46:33.960058 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957957 2572 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Feb 17 12:46:33.960058 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957961 2572 flags.go:64] FLAG: --cpu-manager-policy="none" Feb 17 12:46:33.960058 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957964 2572 flags.go:64] FLAG: --cpu-manager-policy-options="" Feb 17 12:46:33.960058 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957969 2572 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Feb 17 12:46:33.960058 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957972 2572 flags.go:64] FLAG: --enable-controller-attach-detach="true" Feb 17 12:46:33.960667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957975 2572 flags.go:64] FLAG: --enable-debugging-handlers="true" Feb 17 12:46:33.960667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957978 2572 flags.go:64] FLAG: --enable-load-reader="false" Feb 17 12:46:33.960667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957981 2572 flags.go:64] FLAG: --enable-server="true" Feb 17 12:46:33.960667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957984 2572 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Feb 17 12:46:33.960667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957990 2572 flags.go:64] FLAG: --event-burst="100" Feb 17 12:46:33.960667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957993 2572 flags.go:64] FLAG: --event-qps="50" Feb 17 12:46:33.960667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957996 2572 flags.go:64] FLAG: --event-storage-age-limit="default=0" Feb 17 12:46:33.960667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.957999 2572 flags.go:64] FLAG: --event-storage-event-limit="default=0" Feb 17 12:46:33.960667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958003 2572 flags.go:64] FLAG: --eviction-hard="" Feb 17 12:46:33.960667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958006 2572 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Feb 17 12:46:33.960667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958010 2572 flags.go:64] FLAG: --eviction-minimum-reclaim="" Feb 17 12:46:33.960667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958012 2572 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Feb 17 12:46:33.960667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958016 2572 flags.go:64] FLAG: --eviction-soft="" Feb 17 12:46:33.960667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958018 2572 flags.go:64] FLAG: --eviction-soft-grace-period="" Feb 17 12:46:33.960667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958021 2572 flags.go:64] FLAG: --exit-on-lock-contention="false" Feb 17 12:46:33.960667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958024 2572 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Feb 17 12:46:33.960667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958027 2572 flags.go:64] FLAG: --experimental-mounter-path="" Feb 17 12:46:33.960667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958030 2572 flags.go:64] FLAG: --fail-cgroupv1="false" Feb 17 12:46:33.960667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958033 2572 flags.go:64] FLAG: --fail-swap-on="true" Feb 17 12:46:33.960667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958036 2572 flags.go:64] FLAG: --feature-gates="" Feb 17 12:46:33.960667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958040 2572 flags.go:64] FLAG: --file-check-frequency="20s" Feb 17 12:46:33.960667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958043 2572 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Feb 17 12:46:33.960667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958047 2572 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Feb 17 12:46:33.960667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958050 2572 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Feb 17 12:46:33.960667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958054 2572 flags.go:64] FLAG: --healthz-port="10248" Feb 17 12:46:33.960667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958057 2572 flags.go:64] FLAG: --help="false" Feb 17 12:46:33.961324 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958061 2572 flags.go:64] FLAG: --hostname-override="ip-10-0-131-118.ec2.internal" Feb 17 12:46:33.961324 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958064 2572 flags.go:64] FLAG: --housekeeping-interval="10s" Feb 17 12:46:33.961324 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958068 2572 flags.go:64] FLAG: --http-check-frequency="20s" Feb 17 12:46:33.961324 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958071 2572 flags.go:64] FLAG: --image-credential-provider-bin-dir="/usr/libexec/kubelet-image-credential-provider-plugins" Feb 17 12:46:33.961324 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958076 2572 flags.go:64] FLAG: --image-credential-provider-config="/etc/kubernetes/credential-providers/ecr-credential-provider.yaml" Feb 17 12:46:33.961324 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958080 2572 flags.go:64] FLAG: --image-gc-high-threshold="85" Feb 17 12:46:33.961324 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958083 2572 flags.go:64] FLAG: --image-gc-low-threshold="80" Feb 17 12:46:33.961324 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958086 2572 flags.go:64] FLAG: --image-service-endpoint="" Feb 17 12:46:33.961324 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958089 2572 flags.go:64] FLAG: --kernel-memcg-notification="false" Feb 17 12:46:33.961324 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958092 2572 flags.go:64] FLAG: --kube-api-burst="100" Feb 17 12:46:33.961324 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958096 2572 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Feb 17 12:46:33.961324 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958099 2572 flags.go:64] FLAG: --kube-api-qps="50" Feb 17 12:46:33.961324 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958117 2572 flags.go:64] FLAG: --kube-reserved="" Feb 17 12:46:33.961324 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958121 2572 flags.go:64] FLAG: --kube-reserved-cgroup="" Feb 17 12:46:33.961324 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958123 2572 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Feb 17 12:46:33.961324 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958127 2572 flags.go:64] FLAG: --kubelet-cgroups="" Feb 17 12:46:33.961324 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958130 2572 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Feb 17 12:46:33.961324 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958133 2572 flags.go:64] FLAG: --lock-file="" Feb 17 12:46:33.961324 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958136 2572 flags.go:64] FLAG: --log-cadvisor-usage="false" Feb 17 12:46:33.961324 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958139 2572 flags.go:64] FLAG: --log-flush-frequency="5s" Feb 17 12:46:33.961324 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958143 2572 flags.go:64] FLAG: --log-json-info-buffer-size="0" Feb 17 12:46:33.961324 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958154 2572 flags.go:64] FLAG: --log-json-split-stream="false" Feb 17 12:46:33.961324 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958157 2572 flags.go:64] FLAG: --log-text-info-buffer-size="0" Feb 17 12:46:33.961917 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958161 2572 flags.go:64] FLAG: --log-text-split-stream="false" Feb 17 12:46:33.961917 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958164 2572 flags.go:64] FLAG: --logging-format="text" Feb 17 12:46:33.961917 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958167 2572 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Feb 17 12:46:33.961917 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958171 2572 flags.go:64] FLAG: --make-iptables-util-chains="true" Feb 17 12:46:33.961917 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958174 2572 flags.go:64] FLAG: --manifest-url="" Feb 17 12:46:33.961917 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958177 2572 flags.go:64] FLAG: --manifest-url-header="" Feb 17 12:46:33.961917 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958184 2572 flags.go:64] FLAG: --max-housekeeping-interval="15s" Feb 17 12:46:33.961917 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958188 2572 flags.go:64] FLAG: --max-open-files="1000000" Feb 17 12:46:33.961917 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958192 2572 flags.go:64] FLAG: --max-pods="110" Feb 17 12:46:33.961917 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958195 2572 flags.go:64] FLAG: --maximum-dead-containers="-1" Feb 17 12:46:33.961917 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958198 2572 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Feb 17 12:46:33.961917 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958201 2572 flags.go:64] FLAG: --memory-manager-policy="None" Feb 17 12:46:33.961917 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958204 2572 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Feb 17 12:46:33.961917 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958207 2572 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Feb 17 12:46:33.961917 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958211 2572 flags.go:64] FLAG: --node-ip="0.0.0.0" Feb 17 12:46:33.961917 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958214 2572 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/worker=,node.openshift.io/os_id=rhel" Feb 17 12:46:33.961917 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958224 2572 flags.go:64] FLAG: --node-status-max-images="50" Feb 17 12:46:33.961917 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958227 2572 flags.go:64] FLAG: --node-status-update-frequency="10s" Feb 17 12:46:33.961917 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958230 2572 flags.go:64] FLAG: --oom-score-adj="-999" Feb 17 12:46:33.961917 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958234 2572 flags.go:64] FLAG: --pod-cidr="" Feb 17 12:46:33.961917 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958237 2572 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ca3deca44439f185f4632d34b1d894f5fa75cccf603cfd634a130c5928811e73" Feb 17 12:46:33.961917 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958242 2572 flags.go:64] FLAG: --pod-manifest-path="" Feb 17 12:46:33.961917 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958245 2572 flags.go:64] FLAG: --pod-max-pids="-1" Feb 17 12:46:33.961917 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958248 2572 flags.go:64] FLAG: --pods-per-core="0" Feb 17 12:46:33.962491 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958251 2572 flags.go:64] FLAG: --port="10250" Feb 17 12:46:33.962491 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958254 2572 flags.go:64] FLAG: --protect-kernel-defaults="false" Feb 17 12:46:33.962491 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958257 2572 flags.go:64] FLAG: --provider-id="aws:///us-east-1a/i-0adaaa3445b657532" Feb 17 12:46:33.962491 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958260 2572 flags.go:64] FLAG: --qos-reserved="" Feb 17 12:46:33.962491 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958263 2572 flags.go:64] FLAG: --read-only-port="10255" Feb 17 12:46:33.962491 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958266 2572 flags.go:64] FLAG: --register-node="true" Feb 17 12:46:33.962491 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958269 2572 flags.go:64] FLAG: --register-schedulable="true" Feb 17 12:46:33.962491 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958273 2572 flags.go:64] FLAG: --register-with-taints="" Feb 17 12:46:33.962491 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958277 2572 flags.go:64] FLAG: --registry-burst="10" Feb 17 12:46:33.962491 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958280 2572 flags.go:64] FLAG: --registry-qps="5" Feb 17 12:46:33.962491 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958282 2572 flags.go:64] FLAG: --reserved-cpus="" Feb 17 12:46:33.962491 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958285 2572 flags.go:64] FLAG: --reserved-memory="" Feb 17 12:46:33.962491 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958292 2572 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Feb 17 12:46:33.962491 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958295 2572 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Feb 17 12:46:33.962491 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958298 2572 flags.go:64] FLAG: --rotate-certificates="false" Feb 17 12:46:33.962491 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958301 2572 flags.go:64] FLAG: --rotate-server-certificates="false" Feb 17 12:46:33.962491 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958304 2572 flags.go:64] FLAG: --runonce="false" Feb 17 12:46:33.962491 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958306 2572 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Feb 17 12:46:33.962491 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958309 2572 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Feb 17 12:46:33.962491 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958313 2572 flags.go:64] FLAG: --seccomp-default="false" Feb 17 12:46:33.962491 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958316 2572 flags.go:64] FLAG: --serialize-image-pulls="true" Feb 17 12:46:33.962491 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958319 2572 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Feb 17 12:46:33.962491 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958322 2572 flags.go:64] FLAG: --storage-driver-db="cadvisor" Feb 17 12:46:33.962491 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958325 2572 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Feb 17 12:46:33.962491 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958328 2572 flags.go:64] FLAG: --storage-driver-password="root" Feb 17 12:46:33.962491 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958331 2572 flags.go:64] FLAG: --storage-driver-secure="false" Feb 17 12:46:33.963187 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958334 2572 flags.go:64] FLAG: --storage-driver-table="stats" Feb 17 12:46:33.963187 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958337 2572 flags.go:64] FLAG: --storage-driver-user="root" Feb 17 12:46:33.963187 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958340 2572 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Feb 17 12:46:33.963187 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958344 2572 flags.go:64] FLAG: --sync-frequency="1m0s" Feb 17 12:46:33.963187 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958346 2572 flags.go:64] FLAG: --system-cgroups="" Feb 17 12:46:33.963187 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958349 2572 flags.go:64] FLAG: --system-reserved="cpu=500m,ephemeral-storage=1Gi,memory=1Gi" Feb 17 12:46:33.963187 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958356 2572 flags.go:64] FLAG: --system-reserved-cgroup="" Feb 17 12:46:33.963187 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958359 2572 flags.go:64] FLAG: --tls-cert-file="" Feb 17 12:46:33.963187 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958362 2572 flags.go:64] FLAG: --tls-cipher-suites="[]" Feb 17 12:46:33.963187 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958366 2572 flags.go:64] FLAG: --tls-min-version="" Feb 17 12:46:33.963187 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958369 2572 flags.go:64] FLAG: --tls-private-key-file="" Feb 17 12:46:33.963187 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958372 2572 flags.go:64] FLAG: --topology-manager-policy="none" Feb 17 12:46:33.963187 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958375 2572 flags.go:64] FLAG: --topology-manager-policy-options="" Feb 17 12:46:33.963187 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958378 2572 flags.go:64] FLAG: --topology-manager-scope="container" Feb 17 12:46:33.963187 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958381 2572 flags.go:64] FLAG: --v="2" Feb 17 12:46:33.963187 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958386 2572 flags.go:64] FLAG: --version="false" Feb 17 12:46:33.963187 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958390 2572 flags.go:64] FLAG: --vmodule="" Feb 17 12:46:33.963187 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958394 2572 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Feb 17 12:46:33.963187 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958397 2572 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Feb 17 12:46:33.963187 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958501 2572 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Feb 17 12:46:33.963187 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958505 2572 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Feb 17 12:46:33.963187 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958508 2572 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Feb 17 12:46:33.963187 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958511 2572 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Feb 17 12:46:33.963187 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958514 2572 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Feb 17 12:46:33.963801 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958517 2572 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Feb 17 12:46:33.963801 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958520 2572 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Feb 17 12:46:33.963801 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958524 2572 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Feb 17 12:46:33.963801 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958526 2572 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Feb 17 12:46:33.963801 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958530 2572 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Feb 17 12:46:33.963801 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958533 2572 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Feb 17 12:46:33.963801 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958536 2572 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Feb 17 12:46:33.963801 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958539 2572 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Feb 17 12:46:33.963801 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958542 2572 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Feb 17 12:46:33.963801 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958546 2572 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 17 12:46:33.963801 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958550 2572 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Feb 17 12:46:33.963801 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958553 2572 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Feb 17 12:46:33.963801 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958556 2572 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Feb 17 12:46:33.963801 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958559 2572 feature_gate.go:328] unrecognized feature gate: InsightsConfig Feb 17 12:46:33.963801 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958561 2572 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Feb 17 12:46:33.963801 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958564 2572 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Feb 17 12:46:33.963801 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958567 2572 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Feb 17 12:46:33.963801 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958570 2572 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Feb 17 12:46:33.963801 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958572 2572 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Feb 17 12:46:33.963801 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958574 2572 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Feb 17 12:46:33.964338 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958577 2572 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 17 12:46:33.964338 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958580 2572 feature_gate.go:328] unrecognized feature gate: Example Feb 17 12:46:33.964338 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958583 2572 feature_gate.go:328] unrecognized feature gate: GatewayAPI Feb 17 12:46:33.964338 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958586 2572 feature_gate.go:328] unrecognized feature gate: Example2 Feb 17 12:46:33.964338 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958588 2572 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Feb 17 12:46:33.964338 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958591 2572 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Feb 17 12:46:33.964338 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958593 2572 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Feb 17 12:46:33.964338 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958596 2572 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Feb 17 12:46:33.964338 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958598 2572 feature_gate.go:328] unrecognized feature gate: NewOLM Feb 17 12:46:33.964338 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958601 2572 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Feb 17 12:46:33.964338 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958604 2572 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Feb 17 12:46:33.964338 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958606 2572 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Feb 17 12:46:33.964338 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958609 2572 feature_gate.go:328] unrecognized feature gate: OVNObservability Feb 17 12:46:33.964338 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958612 2572 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Feb 17 12:46:33.964338 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958631 2572 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Feb 17 12:46:33.964338 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958635 2572 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Feb 17 12:46:33.964338 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958638 2572 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 17 12:46:33.964338 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958641 2572 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 17 12:46:33.964338 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958644 2572 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Feb 17 12:46:33.964338 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958647 2572 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Feb 17 12:46:33.964834 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958650 2572 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Feb 17 12:46:33.964834 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958653 2572 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Feb 17 12:46:33.964834 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958657 2572 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Feb 17 12:46:33.964834 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958661 2572 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Feb 17 12:46:33.964834 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958664 2572 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Feb 17 12:46:33.964834 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958668 2572 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Feb 17 12:46:33.964834 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958671 2572 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Feb 17 12:46:33.964834 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958674 2572 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 17 12:46:33.964834 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958679 2572 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Feb 17 12:46:33.964834 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958682 2572 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Feb 17 12:46:33.964834 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958685 2572 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 17 12:46:33.964834 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958688 2572 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Feb 17 12:46:33.964834 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958690 2572 feature_gate.go:328] unrecognized feature gate: SignatureStores Feb 17 12:46:33.964834 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958693 2572 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Feb 17 12:46:33.964834 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958696 2572 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Feb 17 12:46:33.964834 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958699 2572 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Feb 17 12:46:33.964834 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958701 2572 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Feb 17 12:46:33.964834 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958704 2572 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Feb 17 12:46:33.964834 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958707 2572 feature_gate.go:328] unrecognized feature gate: PinnedImages Feb 17 12:46:33.965325 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958709 2572 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Feb 17 12:46:33.965325 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958712 2572 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Feb 17 12:46:33.965325 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958715 2572 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Feb 17 12:46:33.965325 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958717 2572 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Feb 17 12:46:33.965325 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958720 2572 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Feb 17 12:46:33.965325 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958722 2572 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Feb 17 12:46:33.965325 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958728 2572 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Feb 17 12:46:33.965325 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958731 2572 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Feb 17 12:46:33.965325 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958733 2572 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Feb 17 12:46:33.965325 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958736 2572 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Feb 17 12:46:33.965325 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958739 2572 feature_gate.go:328] unrecognized feature gate: DualReplica Feb 17 12:46:33.965325 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958741 2572 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Feb 17 12:46:33.965325 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958744 2572 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Feb 17 12:46:33.965325 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958747 2572 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Feb 17 12:46:33.965325 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958749 2572 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Feb 17 12:46:33.965325 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958752 2572 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 17 12:46:33.965325 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958755 2572 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Feb 17 12:46:33.965325 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958757 2572 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Feb 17 12:46:33.965325 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958760 2572 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Feb 17 12:46:33.965325 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958763 2572 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Feb 17 12:46:33.965814 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958765 2572 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Feb 17 12:46:33.965814 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.958770 2572 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Feb 17 12:46:33.965814 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.958779 2572 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Feb 17 12:46:33.965814 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.965064 2572 server.go:530] "Kubelet version" kubeletVersion="v1.33.6" Feb 17 12:46:33.965814 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.965083 2572 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 17 12:46:33.965814 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965153 2572 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Feb 17 12:46:33.965814 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965160 2572 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Feb 17 12:46:33.965814 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965164 2572 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Feb 17 12:46:33.965814 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965167 2572 feature_gate.go:328] unrecognized feature gate: Example2 Feb 17 12:46:33.965814 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965170 2572 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Feb 17 12:46:33.965814 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965173 2572 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Feb 17 12:46:33.965814 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965176 2572 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Feb 17 12:46:33.965814 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965180 2572 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Feb 17 12:46:33.965814 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965182 2572 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Feb 17 12:46:33.965814 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965185 2572 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Feb 17 12:46:33.965814 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965188 2572 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Feb 17 12:46:33.966224 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965191 2572 feature_gate.go:328] unrecognized feature gate: OVNObservability Feb 17 12:46:33.966224 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965195 2572 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Feb 17 12:46:33.966224 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965198 2572 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Feb 17 12:46:33.966224 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965200 2572 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Feb 17 12:46:33.966224 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965204 2572 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Feb 17 12:46:33.966224 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965206 2572 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Feb 17 12:46:33.966224 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965210 2572 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Feb 17 12:46:33.966224 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965212 2572 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Feb 17 12:46:33.966224 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965215 2572 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Feb 17 12:46:33.966224 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965218 2572 feature_gate.go:328] unrecognized feature gate: Example Feb 17 12:46:33.966224 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965220 2572 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Feb 17 12:46:33.966224 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965223 2572 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Feb 17 12:46:33.966224 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965226 2572 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Feb 17 12:46:33.966224 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965228 2572 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Feb 17 12:46:33.966224 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965233 2572 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Feb 17 12:46:33.966224 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965235 2572 feature_gate.go:328] unrecognized feature gate: NewOLM Feb 17 12:46:33.966224 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965238 2572 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Feb 17 12:46:33.966224 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965241 2572 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Feb 17 12:46:33.966224 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965244 2572 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Feb 17 12:46:33.966224 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965246 2572 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Feb 17 12:46:33.966747 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965251 2572 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Feb 17 12:46:33.966747 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965257 2572 feature_gate.go:328] unrecognized feature gate: DualReplica Feb 17 12:46:33.966747 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965260 2572 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Feb 17 12:46:33.966747 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965263 2572 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Feb 17 12:46:33.966747 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965266 2572 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 17 12:46:33.966747 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965269 2572 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Feb 17 12:46:33.966747 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965272 2572 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Feb 17 12:46:33.966747 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965275 2572 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 17 12:46:33.966747 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965278 2572 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Feb 17 12:46:33.966747 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965281 2572 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Feb 17 12:46:33.966747 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965284 2572 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Feb 17 12:46:33.966747 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965288 2572 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 17 12:46:33.966747 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965292 2572 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Feb 17 12:46:33.966747 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965296 2572 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Feb 17 12:46:33.966747 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965299 2572 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Feb 17 12:46:33.966747 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965302 2572 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Feb 17 12:46:33.966747 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965305 2572 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Feb 17 12:46:33.966747 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965308 2572 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Feb 17 12:46:33.966747 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965311 2572 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Feb 17 12:46:33.967241 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965313 2572 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Feb 17 12:46:33.967241 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965316 2572 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Feb 17 12:46:33.967241 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965319 2572 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Feb 17 12:46:33.967241 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965321 2572 feature_gate.go:328] unrecognized feature gate: PinnedImages Feb 17 12:46:33.967241 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965324 2572 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Feb 17 12:46:33.967241 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965327 2572 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Feb 17 12:46:33.967241 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965329 2572 feature_gate.go:328] unrecognized feature gate: SignatureStores Feb 17 12:46:33.967241 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965332 2572 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 17 12:46:33.967241 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965335 2572 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 17 12:46:33.967241 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965339 2572 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Feb 17 12:46:33.967241 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965342 2572 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Feb 17 12:46:33.967241 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965344 2572 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Feb 17 12:46:33.967241 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965347 2572 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Feb 17 12:46:33.967241 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965351 2572 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Feb 17 12:46:33.967241 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965354 2572 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 17 12:46:33.967241 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965356 2572 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Feb 17 12:46:33.967241 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965359 2572 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Feb 17 12:46:33.967241 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965362 2572 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Feb 17 12:46:33.967241 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965365 2572 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Feb 17 12:46:33.967241 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965367 2572 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Feb 17 12:46:33.967727 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965370 2572 feature_gate.go:328] unrecognized feature gate: GatewayAPI Feb 17 12:46:33.967727 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965373 2572 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Feb 17 12:46:33.967727 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965375 2572 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Feb 17 12:46:33.967727 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965378 2572 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Feb 17 12:46:33.967727 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965381 2572 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 17 12:46:33.967727 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965384 2572 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Feb 17 12:46:33.967727 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965386 2572 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Feb 17 12:46:33.967727 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965389 2572 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Feb 17 12:46:33.967727 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965392 2572 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Feb 17 12:46:33.967727 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965395 2572 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Feb 17 12:46:33.967727 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965397 2572 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Feb 17 12:46:33.967727 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965400 2572 feature_gate.go:328] unrecognized feature gate: InsightsConfig Feb 17 12:46:33.967727 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965403 2572 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Feb 17 12:46:33.967727 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965405 2572 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Feb 17 12:46:33.967727 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965408 2572 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Feb 17 12:46:33.967727 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965411 2572 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Feb 17 12:46:33.968113 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.965416 2572 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Feb 17 12:46:33.968113 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965512 2572 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Feb 17 12:46:33.968113 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965516 2572 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Feb 17 12:46:33.968113 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965519 2572 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Feb 17 12:46:33.968113 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965522 2572 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Feb 17 12:46:33.968113 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965526 2572 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Feb 17 12:46:33.968113 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965529 2572 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Feb 17 12:46:33.968113 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965532 2572 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Feb 17 12:46:33.968113 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965535 2572 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Feb 17 12:46:33.968113 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965538 2572 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Feb 17 12:46:33.968113 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965541 2572 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Feb 17 12:46:33.968113 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965543 2572 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Feb 17 12:46:33.968113 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965546 2572 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Feb 17 12:46:33.968113 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965549 2572 feature_gate.go:328] unrecognized feature gate: SignatureStores Feb 17 12:46:33.968113 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965552 2572 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Feb 17 12:46:33.968494 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965555 2572 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Feb 17 12:46:33.968494 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965557 2572 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 17 12:46:33.968494 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965560 2572 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Feb 17 12:46:33.968494 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965563 2572 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Feb 17 12:46:33.968494 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965566 2572 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Feb 17 12:46:33.968494 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965568 2572 feature_gate.go:328] unrecognized feature gate: InsightsConfig Feb 17 12:46:33.968494 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965571 2572 feature_gate.go:328] unrecognized feature gate: OVNObservability Feb 17 12:46:33.968494 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965574 2572 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Feb 17 12:46:33.968494 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965576 2572 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Feb 17 12:46:33.968494 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965579 2572 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Feb 17 12:46:33.968494 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965582 2572 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Feb 17 12:46:33.968494 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965585 2572 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Feb 17 12:46:33.968494 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965588 2572 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Feb 17 12:46:33.968494 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965590 2572 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Feb 17 12:46:33.968494 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965593 2572 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Feb 17 12:46:33.968494 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965596 2572 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 17 12:46:33.968494 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965599 2572 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Feb 17 12:46:33.968494 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965602 2572 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Feb 17 12:46:33.968494 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965604 2572 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 17 12:46:33.968950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965608 2572 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 17 12:46:33.968950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965611 2572 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Feb 17 12:46:33.968950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965613 2572 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Feb 17 12:46:33.968950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965616 2572 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Feb 17 12:46:33.968950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965624 2572 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Feb 17 12:46:33.968950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965627 2572 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Feb 17 12:46:33.968950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965630 2572 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Feb 17 12:46:33.968950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965633 2572 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Feb 17 12:46:33.968950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965636 2572 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Feb 17 12:46:33.968950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965638 2572 feature_gate.go:328] unrecognized feature gate: DualReplica Feb 17 12:46:33.968950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965641 2572 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Feb 17 12:46:33.968950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965644 2572 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 17 12:46:33.968950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965646 2572 feature_gate.go:328] unrecognized feature gate: GatewayAPI Feb 17 12:46:33.968950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965649 2572 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Feb 17 12:46:33.968950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965651 2572 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Feb 17 12:46:33.968950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965654 2572 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Feb 17 12:46:33.968950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965657 2572 feature_gate.go:328] unrecognized feature gate: NewOLM Feb 17 12:46:33.968950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965660 2572 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Feb 17 12:46:33.968950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965662 2572 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Feb 17 12:46:33.968950 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965665 2572 feature_gate.go:328] unrecognized feature gate: PinnedImages Feb 17 12:46:33.969509 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965667 2572 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Feb 17 12:46:33.969509 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965670 2572 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 17 12:46:33.969509 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965673 2572 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Feb 17 12:46:33.969509 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965675 2572 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Feb 17 12:46:33.969509 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965678 2572 feature_gate.go:328] unrecognized feature gate: Example Feb 17 12:46:33.969509 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965680 2572 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Feb 17 12:46:33.969509 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965683 2572 feature_gate.go:328] unrecognized feature gate: Example2 Feb 17 12:46:33.969509 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965685 2572 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Feb 17 12:46:33.969509 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965688 2572 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Feb 17 12:46:33.969509 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965690 2572 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Feb 17 12:46:33.969509 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965693 2572 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Feb 17 12:46:33.969509 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965696 2572 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Feb 17 12:46:33.969509 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965699 2572 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 17 12:46:33.969509 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965702 2572 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Feb 17 12:46:33.969509 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965705 2572 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Feb 17 12:46:33.969509 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965709 2572 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Feb 17 12:46:33.969509 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965713 2572 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Feb 17 12:46:33.969509 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965716 2572 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Feb 17 12:46:33.969509 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965719 2572 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Feb 17 12:46:33.969956 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965722 2572 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Feb 17 12:46:33.969956 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965725 2572 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Feb 17 12:46:33.969956 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965728 2572 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Feb 17 12:46:33.969956 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965731 2572 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Feb 17 12:46:33.969956 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965734 2572 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Feb 17 12:46:33.969956 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965737 2572 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Feb 17 12:46:33.969956 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965739 2572 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Feb 17 12:46:33.969956 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965742 2572 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Feb 17 12:46:33.969956 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965745 2572 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Feb 17 12:46:33.969956 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965748 2572 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Feb 17 12:46:33.969956 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965750 2572 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Feb 17 12:46:33.969956 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965753 2572 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Feb 17 12:46:33.969956 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965756 2572 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Feb 17 12:46:33.969956 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:33.965758 2572 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Feb 17 12:46:33.969956 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.965763 2572 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Feb 17 12:46:33.970365 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.966554 2572 server.go:962] "Client rotation is on, will bootstrap in background" Feb 17 12:46:33.970365 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.969996 2572 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Feb 17 12:46:33.970832 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.970813 2572 server.go:1019] "Starting client certificate rotation" Feb 17 12:46:33.970929 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.970913 2572 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kube-apiserver-client-kubelet" Feb 17 12:46:33.970965 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.970956 2572 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kube-apiserver-client-kubelet" Feb 17 12:46:33.993133 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.993095 2572 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 17 12:46:33.994866 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:33.994845 2572 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 17 12:46:34.009585 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.009560 2572 log.go:25] "Validated CRI v1 runtime API" Feb 17 12:46:34.017892 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.017869 2572 log.go:25] "Validated CRI v1 image API" Feb 17 12:46:34.019510 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.019494 2572 server.go:1452] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Feb 17 12:46:34.024652 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.024636 2572 reflector.go:430] "Caches populated" logger="kubernetes.io/kube-apiserver-client-kubelet" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Feb 17 12:46:34.029309 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.029285 2572 fs.go:135] Filesystem UUIDs: map[0f211c87-5edc-4d9b-a0e5-50bf489c8241:/dev/nvme0n1p4 7B77-95E7:/dev/nvme0n1p2 bef63d95-57d2-438e-9596-7034dceb0e28:/dev/nvme0n1p3] Feb 17 12:46:34.029369 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.029309 2572 fs.go:136] Filesystem partitions: map[/dev/nvme0n1p3:{mountpoint:/boot major:259 minor:3 fsType:ext4 blockSize:0} /dev/nvme0n1p4:{mountpoint:/var major:259 minor:4 fsType:xfs blockSize:0} /dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:31 fsType:tmpfs blockSize:0} composefs_0-33:{mountpoint:/ major:0 minor:33 fsType:overlay blockSize:0}] Feb 17 12:46:34.034205 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.034083 2572 manager.go:217] Machine: {Timestamp:2026-02-17 12:46:34.03300218 +0000 UTC m=+0.383412982 CPUVendorID:GenuineIntel NumCores:8 NumPhysicalCores:4 NumSockets:1 CpuFrequency:3099522 MemoryCapacity:33164492800 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:ec2de014f823d76bb20a1232a5f0324e SystemUUID:ec2de014-f823-d76b-b20a-1232a5f0324e BootID:f64c1eb2-929a-40ce-bf79-64d5200ad6c9 Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6632898560 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/nvme0n1p4 DeviceMajor:259 DeviceMinor:4 Capacity:128243970048 Type:vfs Inodes:62651840 HasInodes:true} {Device:composefs_0-33 DeviceMajor:0 DeviceMinor:33 Capacity:6090752 Type:vfs Inodes:18446744073709551615 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:31 Capacity:16582246400 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/nvme0n1p3 DeviceMajor:259 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16582246400 Type:vfs Inodes:4048400 HasInodes:true}] DiskMap:map[259:0:{Name:nvme0n1 Major:259 Minor:0 Size:128849018880 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:02:e6:59:9c:84:ed Speed:0 Mtu:9001} {Name:ens5 MacAddress:02:e6:59:9c:84:ed Speed:0 Mtu:9001} {Name:ovs-system MacAddress:76:a6:83:98:b4:d5 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33164492800 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0 4] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:1 Threads:[1 5] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:2 Threads:[2 6] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:} {Id:3 Threads:[3 7] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0 BookID: DrawerID:}] Caches:[{Id:0 Size:37486592 Type:Unified Level:3}] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Feb 17 12:46:34.034205 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.034201 2572 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Feb 17 12:46:34.034330 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.034318 2572 manager.go:233] Version: {KernelVersion:5.14.0-570.86.1.el9_6.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 9.6.20260204-0 (Plow) DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Feb 17 12:46:34.036085 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.036056 2572 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 17 12:46:34.036242 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.036088 2572 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ip-10-0-131-118.ec2.internal","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"500m","ephemeral-storage":"1Gi","memory":"1Gi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 17 12:46:34.036288 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.036252 2572 topology_manager.go:138] "Creating topology manager with none policy" Feb 17 12:46:34.036288 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.036260 2572 container_manager_linux.go:306] "Creating device plugin manager" Feb 17 12:46:34.036288 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.036268 2572 manager.go:141] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 17 12:46:34.036967 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.036957 2572 server.go:72] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 17 12:46:34.038411 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.038401 2572 state_mem.go:36] "Initialized new in-memory state store" Feb 17 12:46:34.038513 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.038504 2572 server.go:1267] "Using root directory" path="/var/lib/kubelet" Feb 17 12:46:34.040563 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.040553 2572 kubelet.go:491] "Attempting to sync node with API server" Feb 17 12:46:34.040596 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.040566 2572 kubelet.go:386] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 17 12:46:34.040596 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.040579 2572 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Feb 17 12:46:34.040596 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.040588 2572 kubelet.go:397] "Adding apiserver pod source" Feb 17 12:46:34.040596 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.040596 2572 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 17 12:46:34.042183 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.042171 2572 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Feb 17 12:46:34.042227 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.042189 2572 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Feb 17 12:46:34.043660 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.043644 2572 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-vw26v" Feb 17 12:46:34.045847 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.045829 2572 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="cri-o" version="1.33.9-2.rhaos4.20.gitb9ac835.el9" apiVersion="v1" Feb 17 12:46:34.047902 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.047886 2572 kubelet.go:953] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Feb 17 12:46:34.049742 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.049729 2572 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Feb 17 12:46:34.049811 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.049746 2572 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Feb 17 12:46:34.049811 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.049755 2572 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Feb 17 12:46:34.049811 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.049762 2572 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Feb 17 12:46:34.049811 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.049772 2572 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Feb 17 12:46:34.049811 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.049779 2572 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/secret" Feb 17 12:46:34.049811 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.049785 2572 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Feb 17 12:46:34.049811 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.049791 2572 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Feb 17 12:46:34.049811 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.049798 2572 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/fc" Feb 17 12:46:34.049811 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.049804 2572 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Feb 17 12:46:34.050043 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.049818 2572 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/projected" Feb 17 12:46:34.050199 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.050189 2572 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Feb 17 12:46:34.050904 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.050891 2572 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-vw26v" Feb 17 12:46:34.051006 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.050993 2572 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/csi" Feb 17 12:46:34.051044 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.051010 2572 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/image" Feb 17 12:46:34.054633 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.054619 2572 watchdog_linux.go:99] "Systemd watchdog is not enabled" Feb 17 12:46:34.054704 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.054668 2572 server.go:1295] "Started kubelet" Feb 17 12:46:34.054776 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.054720 2572 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Feb 17 12:46:34.054843 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.054763 2572 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 17 12:46:34.054904 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.054872 2572 server_v1.go:47] "podresources" method="list" useActivePods=true Feb 17 12:46:34.055181 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.055164 2572 reflector.go:430] "Caches populated" type="*v1.Node" reflector="k8s.io/client-go/informers/factory.go:160" Feb 17 12:46:34.055448 ip-10-0-131-118 systemd[1]: Started Kubernetes Kubelet. Feb 17 12:46:34.056541 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.056526 2572 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 17 12:46:34.059704 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.059683 2572 nodeinfomanager.go:417] Failed to publish CSINode: nodes "ip-10-0-131-118.ec2.internal" not found Feb 17 12:46:34.060613 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.060583 2572 server.go:317] "Adding debug handlers to kubelet server" Feb 17 12:46:34.060912 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.060889 2572 reflector.go:430] "Caches populated" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" Feb 17 12:46:34.063244 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.063114 2572 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kubelet-serving" Feb 17 12:46:34.063589 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.063575 2572 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 17 12:46:34.063710 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.063695 2572 volume_manager.go:295] "The desired_state_of_world populator starts" Feb 17 12:46:34.063710 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.063710 2572 volume_manager.go:297] "Starting Kubelet Volume Manager" Feb 17 12:46:34.063829 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.063815 2572 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Feb 17 12:46:34.063934 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.063918 2572 reconstruct.go:97] "Volume reconstruction finished" Feb 17 12:46:34.063934 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.063930 2572 reconciler.go:26] "Reconciler: start to sync state" Feb 17 12:46:34.064573 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:34.064551 2572 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"ip-10-0-131-118.ec2.internal\" not found" Feb 17 12:46:34.066480 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.066458 2572 reflector.go:430] "Caches populated" type="*v1.CSIDriver" reflector="k8s.io/client-go/informers/factory.go:160" Feb 17 12:46:34.066912 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.066891 2572 factory.go:55] Registering systemd factory Feb 17 12:46:34.066912 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.066914 2572 factory.go:223] Registration of the systemd container factory successfully Feb 17 12:46:34.067250 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.067232 2572 factory.go:153] Registering CRI-O factory Feb 17 12:46:34.067325 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.067252 2572 factory.go:223] Registration of the crio container factory successfully Feb 17 12:46:34.067325 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:34.067281 2572 kubelet.go:1618] "Image garbage collection failed once. Stats initialization may not have completed yet" err="failed to get imageFs info: unable to find data in memory cache" Feb 17 12:46:34.067423 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.067338 2572 factory.go:221] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Feb 17 12:46:34.067423 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.067364 2572 factory.go:103] Registering Raw factory Feb 17 12:46:34.067423 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.067376 2572 manager.go:1196] Started watching for new ooms in manager Feb 17 12:46:34.067732 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.067718 2572 manager.go:319] Starting recovery of all containers Feb 17 12:46:34.068421 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:34.068396 2572 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ip-10-0-131-118.ec2.internal\" not found" node="ip-10-0-131-118.ec2.internal" Feb 17 12:46:34.074638 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.074618 2572 nodeinfomanager.go:417] Failed to publish CSINode: nodes "ip-10-0-131-118.ec2.internal" not found Feb 17 12:46:34.076815 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.076779 2572 manager.go:324] Recovery completed Feb 17 12:46:34.080539 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.080526 2572 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Feb 17 12:46:34.082332 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.082318 2572 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-131-118.ec2.internal" event="NodeHasSufficientMemory" Feb 17 12:46:34.082382 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.082345 2572 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-131-118.ec2.internal" event="NodeHasNoDiskPressure" Feb 17 12:46:34.082382 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.082356 2572 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-131-118.ec2.internal" event="NodeHasSufficientPID" Feb 17 12:46:34.082775 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.082758 2572 cpu_manager.go:222] "Starting CPU manager" policy="none" Feb 17 12:46:34.082775 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.082772 2572 cpu_manager.go:223] "Reconciling" reconcilePeriod="10s" Feb 17 12:46:34.082890 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.082792 2572 state_mem.go:36] "Initialized new in-memory state store" Feb 17 12:46:34.085117 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.085092 2572 policy_none.go:49] "None policy: Start" Feb 17 12:46:34.085167 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.085121 2572 memory_manager.go:186] "Starting memorymanager" policy="None" Feb 17 12:46:34.085167 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.085133 2572 state_mem.go:35] "Initializing new in-memory state store" Feb 17 12:46:34.123794 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.123759 2572 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Feb 17 12:46:34.127885 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.124660 2572 manager.go:341] "Starting Device Plugin manager" Feb 17 12:46:34.127885 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:34.124724 2572 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Feb 17 12:46:34.127885 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.124734 2572 server.go:85] "Starting device plugin registration server" Feb 17 12:46:34.127885 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.124939 2572 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 17 12:46:34.127885 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.124950 2572 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 17 12:46:34.127885 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.125141 2572 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Feb 17 12:46:34.127885 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.125197 2572 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Feb 17 12:46:34.127885 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.125205 2572 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 17 12:46:34.127885 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.125238 2572 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Feb 17 12:46:34.127885 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.125285 2572 status_manager.go:230] "Starting to sync pod status with apiserver" Feb 17 12:46:34.127885 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.125309 2572 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Feb 17 12:46:34.127885 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.125318 2572 kubelet.go:2451] "Starting kubelet main sync loop" Feb 17 12:46:34.127885 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:34.125483 2572 kubelet.go:2475] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Feb 17 12:46:34.127885 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:34.125855 2572 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="non-existent label \"crio-containers\"" Feb 17 12:46:34.127885 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:34.125894 2572 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ip-10-0-131-118.ec2.internal\" not found" Feb 17 12:46:34.128382 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.128157 2572 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Feb 17 12:46:34.134058 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.134043 2572 nodeinfomanager.go:417] Failed to publish CSINode: nodes "ip-10-0-131-118.ec2.internal" not found Feb 17 12:46:34.225480 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.225393 2572 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Feb 17 12:46:34.225631 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.225546 2572 kubelet.go:2537] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-131-118.ec2.internal","kube-system/kube-apiserver-proxy-ip-10-0-131-118.ec2.internal"] Feb 17 12:46:34.225631 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.225606 2572 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Feb 17 12:46:34.226439 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.226422 2572 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-131-118.ec2.internal" event="NodeHasSufficientMemory" Feb 17 12:46:34.226526 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.226422 2572 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-131-118.ec2.internal" event="NodeHasSufficientMemory" Feb 17 12:46:34.226526 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.226458 2572 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-131-118.ec2.internal" event="NodeHasNoDiskPressure" Feb 17 12:46:34.226526 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.226474 2572 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-131-118.ec2.internal" event="NodeHasSufficientPID" Feb 17 12:46:34.226526 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.226478 2572 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-131-118.ec2.internal" event="NodeHasNoDiskPressure" Feb 17 12:46:34.226526 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.226490 2572 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-131-118.ec2.internal" event="NodeHasSufficientPID" Feb 17 12:46:34.226526 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.226504 2572 kubelet_node_status.go:78] "Attempting to register node" node="ip-10-0-131-118.ec2.internal" Feb 17 12:46:34.227572 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.227557 2572 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Feb 17 12:46:34.227694 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.227679 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-131-118.ec2.internal" Feb 17 12:46:34.227750 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.227713 2572 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Feb 17 12:46:34.229892 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.229870 2572 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-131-118.ec2.internal" event="NodeHasSufficientMemory" Feb 17 12:46:34.229980 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.229902 2572 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-131-118.ec2.internal" event="NodeHasNoDiskPressure" Feb 17 12:46:34.229980 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.229917 2572 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-131-118.ec2.internal" event="NodeHasSufficientPID" Feb 17 12:46:34.229980 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.229919 2572 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-131-118.ec2.internal" event="NodeHasSufficientMemory" Feb 17 12:46:34.229980 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.229947 2572 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-131-118.ec2.internal" event="NodeHasNoDiskPressure" Feb 17 12:46:34.229980 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.229960 2572 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-131-118.ec2.internal" event="NodeHasSufficientPID" Feb 17 12:46:34.231026 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.231009 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/kube-apiserver-proxy-ip-10-0-131-118.ec2.internal" Feb 17 12:46:34.231089 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.231035 2572 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Feb 17 12:46:34.231763 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.231738 2572 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-131-118.ec2.internal" event="NodeHasSufficientMemory" Feb 17 12:46:34.231816 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.231775 2572 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-131-118.ec2.internal" event="NodeHasNoDiskPressure" Feb 17 12:46:34.231816 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.231786 2572 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-131-118.ec2.internal" event="NodeHasSufficientPID" Feb 17 12:46:34.235512 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.235496 2572 kubelet_node_status.go:81] "Successfully registered node" node="ip-10-0-131-118.ec2.internal" Feb 17 12:46:34.235604 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:34.235521 2572 kubelet_node_status.go:597] "Error updating node status, will retry" err="error getting node \"ip-10-0-131-118.ec2.internal\": node \"ip-10-0-131-118.ec2.internal\" not found" Feb 17 12:46:34.245628 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.245613 2572 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-131-118.ec2.internal" Feb 17 12:46:34.249849 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.249835 2572 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-131-118.ec2.internal" Feb 17 12:46:34.258489 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.258467 2572 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Feb 17 12:46:34.261505 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.261490 2572 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Feb 17 12:46:34.264618 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.264604 2572 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-131-118.ec2.internal" Feb 17 12:46:34.281466 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.281444 2572 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Feb 17 12:46:34.281569 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:34.281484 2572 kubelet.go:3342] "Failed creating a mirror pod" err="pods \"kube-rbac-proxy-crio-ip-10-0-131-118.ec2.internal\" already exists" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-131-118.ec2.internal" Feb 17 12:46:34.281569 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.281497 2572 kubelet.go:3340] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-131-118.ec2.internal" Feb 17 12:46:34.295950 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.295929 2572 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Feb 17 12:46:34.296034 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:34.295968 2572 kubelet.go:3342] "Failed creating a mirror pod" err="pods \"kube-apiserver-proxy-ip-10-0-131-118.ec2.internal\" already exists" pod="kube-system/kube-apiserver-proxy-ip-10-0-131-118.ec2.internal" Feb 17 12:46:34.365597 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.365563 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/9b1938025d97d9e77a432753c9d25dde-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-131-118.ec2.internal\" (UID: \"9b1938025d97d9e77a432753c9d25dde\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-131-118.ec2.internal" Feb 17 12:46:34.365597 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.365598 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9b1938025d97d9e77a432753c9d25dde-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-131-118.ec2.internal\" (UID: \"9b1938025d97d9e77a432753c9d25dde\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-131-118.ec2.internal" Feb 17 12:46:34.365797 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.365624 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/cde915e910705d98ba55e7806902bce3-config\") pod \"kube-apiserver-proxy-ip-10-0-131-118.ec2.internal\" (UID: \"cde915e910705d98ba55e7806902bce3\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-131-118.ec2.internal" Feb 17 12:46:34.466357 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.466328 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/9b1938025d97d9e77a432753c9d25dde-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-131-118.ec2.internal\" (UID: \"9b1938025d97d9e77a432753c9d25dde\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-131-118.ec2.internal" Feb 17 12:46:34.466506 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.466367 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9b1938025d97d9e77a432753c9d25dde-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-131-118.ec2.internal\" (UID: \"9b1938025d97d9e77a432753c9d25dde\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-131-118.ec2.internal" Feb 17 12:46:34.466506 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.466385 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/cde915e910705d98ba55e7806902bce3-config\") pod \"kube-apiserver-proxy-ip-10-0-131-118.ec2.internal\" (UID: \"cde915e910705d98ba55e7806902bce3\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-131-118.ec2.internal" Feb 17 12:46:34.466893 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.466878 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/host-path/cde915e910705d98ba55e7806902bce3-config\") pod \"kube-apiserver-proxy-ip-10-0-131-118.ec2.internal\" (UID: \"cde915e910705d98ba55e7806902bce3\") " pod="kube-system/kube-apiserver-proxy-ip-10-0-131-118.ec2.internal" Feb 17 12:46:34.466947 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.466885 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9b1938025d97d9e77a432753c9d25dde-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-ip-10-0-131-118.ec2.internal\" (UID: \"9b1938025d97d9e77a432753c9d25dde\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-131-118.ec2.internal" Feb 17 12:46:34.466947 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.466886 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/9b1938025d97d9e77a432753c9d25dde-etc-kube\") pod \"kube-rbac-proxy-crio-ip-10-0-131-118.ec2.internal\" (UID: \"9b1938025d97d9e77a432753c9d25dde\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-131-118.ec2.internal" Feb 17 12:46:34.560476 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.560376 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-131-118.ec2.internal" Feb 17 12:46:34.564293 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.564272 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/kube-apiserver-proxy-ip-10-0-131-118.ec2.internal" Feb 17 12:46:34.970811 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.970783 2572 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 17 12:46:34.971618 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.970931 2572 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" err="very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received" Feb 17 12:46:34.971618 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.970962 2572 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" err="very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received" Feb 17 12:46:34.971618 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:34.970962 2572 reflector.go:556] "Warning: watch ended with error" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" err="very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received" Feb 17 12:46:35.041618 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.041591 2572 apiserver.go:52] "Watching apiserver" Feb 17 12:46:35.043619 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:35.043583 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcde915e910705d98ba55e7806902bce3.slice/crio-0314b54a89b423bfd4bcf164627445db57712ff661aa0160b447d8246e97327b WatchSource:0}: Error finding container 0314b54a89b423bfd4bcf164627445db57712ff661aa0160b447d8246e97327b: Status 404 returned error can't find the container with id 0314b54a89b423bfd4bcf164627445db57712ff661aa0160b447d8246e97327b Feb 17 12:46:35.044296 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:35.044260 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b1938025d97d9e77a432753c9d25dde.slice/crio-b4e60fdd9faf79f78250c6ac3d28caf5b0161a5723f4b150444a7dc020e4d581 WatchSource:0}: Error finding container b4e60fdd9faf79f78250c6ac3d28caf5b0161a5723f4b150444a7dc020e4d581: Status 404 returned error can't find the container with id b4e60fdd9faf79f78250c6ac3d28caf5b0161a5723f4b150444a7dc020e4d581 Feb 17 12:46:35.046475 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.045608 2572 reflector.go:430] "Caches populated" type="*v1.Pod" reflector="pkg/kubelet/config/apiserver.go:66" Feb 17 12:46:35.046475 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.046134 2572 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-f2tlm","openshift-dns/node-resolver-jqfj5","openshift-image-registry/node-ca-qjw6p","openshift-multus/network-metrics-daemon-vmpwv","openshift-network-operator/iptables-alerter-fm5zs","openshift-ovn-kubernetes/ovnkube-node-lb8m9","kube-system/konnectivity-agent-kmsrd","kube-system/kube-apiserver-proxy-ip-10-0-131-118.ec2.internal","openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv","openshift-cluster-node-tuning-operator/tuned-48j6f","openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-131-118.ec2.internal","openshift-multus/multus-additional-cni-plugins-wfszv","openshift-multus/multus-tsb7s"] Feb 17 12:46:35.048974 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.048960 2572 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 17 12:46:35.049698 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.049683 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:46:35.049792 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:35.049770 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-f2tlm" podUID="f1e10a0f-13dd-44a0-9d79-e6375e12ac24" Feb 17 12:46:35.051829 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.051813 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-jqfj5" Feb 17 12:46:35.052842 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.052812 2572 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kube-apiserver-client-kubelet" expiration="2028-02-17 12:41:34 +0000 UTC" deadline="2027-09-03 13:09:24.962168125 +0000 UTC" Feb 17 12:46:35.052908 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.052841 2572 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kube-apiserver-client-kubelet" sleep="13512h22m49.909330565s" Feb 17 12:46:35.053369 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.053347 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"kube-root-ca.crt\"" Feb 17 12:46:35.053467 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.053379 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"openshift-service-ca.crt\"" Feb 17 12:46:35.053467 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.053400 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"node-resolver-dockercfg-rh78h\"" Feb 17 12:46:35.055853 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.055837 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-qjw6p" Feb 17 12:46:35.057186 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.057170 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"kube-root-ca.crt\"" Feb 17 12:46:35.058950 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.057522 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"openshift-service-ca.crt\"" Feb 17 12:46:35.058950 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.057529 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"image-registry-certificates\"" Feb 17 12:46:35.058950 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.057789 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"node-ca-dockercfg-nj8cg\"" Feb 17 12:46:35.058950 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.058931 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:46:35.059403 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:35.059020 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vmpwv" podUID="685523da-d834-463a-8a1a-ef70069c19c5" Feb 17 12:46:35.059403 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.059327 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-fm5zs" Feb 17 12:46:35.060816 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.060795 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-dockercfg-2qnhh\"" Feb 17 12:46:35.060976 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.060962 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"kube-root-ca.crt\"" Feb 17 12:46:35.061052 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.061021 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"openshift-service-ca.crt\"" Feb 17 12:46:35.061126 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.061055 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-script\"" Feb 17 12:46:35.061610 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.061594 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.063148 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.063131 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-node-dockercfg-bpl4q\"" Feb 17 12:46:35.063332 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.063316 2572 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kubelet-serving" Feb 17 12:46:35.063428 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.063335 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-node-metrics-cert\"" Feb 17 12:46:35.064054 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.063810 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/konnectivity-agent-kmsrd" Feb 17 12:46:35.064054 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.063858 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"kube-root-ca.crt\"" Feb 17 12:46:35.064054 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.063876 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-config\"" Feb 17 12:46:35.064054 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.063931 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"openshift-service-ca.crt\"" Feb 17 12:46:35.064054 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.063876 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"env-overrides\"" Feb 17 12:46:35.064054 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.064041 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-script-lib\"" Feb 17 12:46:35.065458 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.065439 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"konnectivity-agent\"" Feb 17 12:46:35.065613 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.065597 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"default-dockercfg-9k8j8\"" Feb 17 12:46:35.065685 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.065622 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"kube-system\"/\"konnectivity-ca-bundle\"" Feb 17 12:46:35.066363 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.066337 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" Feb 17 12:46:35.067922 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.067906 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-csi-drivers\"/\"aws-ebs-csi-driver-node-metrics-serving-cert\"" Feb 17 12:46:35.068023 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.067943 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-csi-drivers\"/\"openshift-service-ca.crt\"" Feb 17 12:46:35.068023 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.067908 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-csi-drivers\"/\"aws-ebs-csi-driver-node-sa-dockercfg-dhnfh\"" Feb 17 12:46:35.068023 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.068008 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-csi-drivers\"/\"kube-root-ca.crt\"" Feb 17 12:46:35.068459 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.068446 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.068738 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.068721 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs\") pod \"network-metrics-daemon-vmpwv\" (UID: \"685523da-d834-463a-8a1a-ef70069c19c5\") " pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:46:35.068787 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.068746 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ba1edd82-a8f5-4e73-ac84-987cbbec0939-host-slash\") pod \"iptables-alerter-fm5zs\" (UID: \"ba1edd82-a8f5-4e73-ac84-987cbbec0939\") " pod="openshift-network-operator/iptables-alerter-fm5zs" Feb 17 12:46:35.068787 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.068769 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-run-ovn\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.068864 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.068786 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m6g7\" (UniqueName: \"kubernetes.io/projected/01eb7d37-a76a-4933-87df-44cf8a8d3152-kube-api-access-2m6g7\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.068864 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.068827 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/22f3475b-a5d2-4d66-818f-5010e57aac8e-hosts-file\") pod \"node-resolver-jqfj5\" (UID: \"22f3475b-a5d2-4d66-818f-5010e57aac8e\") " pod="openshift-dns/node-resolver-jqfj5" Feb 17 12:46:35.068864 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.068856 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-log-socket\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.068976 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.068894 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-host-run-ovn-kubernetes\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.068976 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.068912 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/01eb7d37-a76a-4933-87df-44cf8a8d3152-ovnkube-config\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.068976 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.068938 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/3d7ab4ec-e6ac-454b-8724-36a22d414dae-socket-dir\") pod \"aws-ebs-csi-driver-node-ghtvv\" (UID: \"3d7ab4ec-e6ac-454b-8724-36a22d414dae\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" Feb 17 12:46:35.068976 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.068955 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/3d7ab4ec-e6ac-454b-8724-36a22d414dae-registration-dir\") pod \"aws-ebs-csi-driver-node-ghtvv\" (UID: \"3d7ab4ec-e6ac-454b-8724-36a22d414dae\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" Feb 17 12:46:35.069139 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.068970 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/3d7ab4ec-e6ac-454b-8724-36a22d414dae-device-dir\") pod \"aws-ebs-csi-driver-node-ghtvv\" (UID: \"3d7ab4ec-e6ac-454b-8724-36a22d414dae\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" Feb 17 12:46:35.069139 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069026 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fcrg\" (UniqueName: \"kubernetes.io/projected/22f3475b-a5d2-4d66-818f-5010e57aac8e-kube-api-access-2fcrg\") pod \"node-resolver-jqfj5\" (UID: \"22f3475b-a5d2-4d66-818f-5010e57aac8e\") " pod="openshift-dns/node-resolver-jqfj5" Feb 17 12:46:35.069139 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069046 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7jz4\" (UniqueName: \"kubernetes.io/projected/685523da-d834-463a-8a1a-ef70069c19c5-kube-api-access-x7jz4\") pod \"network-metrics-daemon-vmpwv\" (UID: \"685523da-d834-463a-8a1a-ef70069c19c5\") " pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:46:35.069139 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069061 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-etc-openvswitch\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.069139 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069084 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-node-log\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.069139 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069117 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.069139 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069135 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3d7ab4ec-e6ac-454b-8724-36a22d414dae-kubelet-dir\") pod \"aws-ebs-csi-driver-node-ghtvv\" (UID: \"3d7ab4ec-e6ac-454b-8724-36a22d414dae\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" Feb 17 12:46:35.069385 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069155 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/3d7ab4ec-e6ac-454b-8724-36a22d414dae-etc-selinux\") pod \"aws-ebs-csi-driver-node-ghtvv\" (UID: \"3d7ab4ec-e6ac-454b-8724-36a22d414dae\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" Feb 17 12:46:35.069385 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069178 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-var-lib-openvswitch\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.069385 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069200 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/01eb7d37-a76a-4933-87df-44cf8a8d3152-env-overrides\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.069385 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069224 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/01eb7d37-a76a-4933-87df-44cf8a8d3152-ovnkube-script-lib\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.069385 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069255 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qksk7\" (UniqueName: \"kubernetes.io/projected/3d7ab4ec-e6ac-454b-8724-36a22d414dae-kube-api-access-qksk7\") pod \"aws-ebs-csi-driver-node-ghtvv\" (UID: \"3d7ab4ec-e6ac-454b-8724-36a22d414dae\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" Feb 17 12:46:35.069385 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069278 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79v2q\" (UniqueName: \"kubernetes.io/projected/ba1edd82-a8f5-4e73-ac84-987cbbec0939-kube-api-access-79v2q\") pod \"iptables-alerter-fm5zs\" (UID: \"ba1edd82-a8f5-4e73-ac84-987cbbec0939\") " pod="openshift-network-operator/iptables-alerter-fm5zs" Feb 17 12:46:35.069385 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069298 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-run-systemd\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.069385 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069322 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-run-openvswitch\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.069385 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069352 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/30392cb9-7527-44af-833b-567cd508520e-agent-certs\") pod \"konnectivity-agent-kmsrd\" (UID: \"30392cb9-7527-44af-833b-567cd508520e\") " pod="kube-system/konnectivity-agent-kmsrd" Feb 17 12:46:35.069385 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069378 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/30392cb9-7527-44af-833b-567cd508520e-konnectivity-ca\") pod \"konnectivity-agent-kmsrd\" (UID: \"30392cb9-7527-44af-833b-567cd508520e\") " pod="kube-system/konnectivity-agent-kmsrd" Feb 17 12:46:35.069819 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069404 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-host-kubelet\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.069819 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069425 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-host-slash\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.069819 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069459 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8738b60c-920d-4bc5-933f-01d0fb8d68d3-serviceca\") pod \"node-ca-qjw6p\" (UID: \"8738b60c-920d-4bc5-933f-01d0fb8d68d3\") " pod="openshift-image-registry/node-ca-qjw6p" Feb 17 12:46:35.069819 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069481 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cz42\" (UniqueName: \"kubernetes.io/projected/8738b60c-920d-4bc5-933f-01d0fb8d68d3-kube-api-access-2cz42\") pod \"node-ca-qjw6p\" (UID: \"8738b60c-920d-4bc5-933f-01d0fb8d68d3\") " pod="openshift-image-registry/node-ca-qjw6p" Feb 17 12:46:35.069819 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069497 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/ba1edd82-a8f5-4e73-ac84-987cbbec0939-iptables-alerter-script\") pod \"iptables-alerter-fm5zs\" (UID: \"ba1edd82-a8f5-4e73-ac84-987cbbec0939\") " pod="openshift-network-operator/iptables-alerter-fm5zs" Feb 17 12:46:35.069819 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069519 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-systemd-units\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.069819 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069542 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-host-cni-bin\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.069819 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069557 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/01eb7d37-a76a-4933-87df-44cf8a8d3152-ovn-node-metrics-cert\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.069819 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069571 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/22f3475b-a5d2-4d66-818f-5010e57aac8e-tmp-dir\") pod \"node-resolver-jqfj5\" (UID: \"22f3475b-a5d2-4d66-818f-5010e57aac8e\") " pod="openshift-dns/node-resolver-jqfj5" Feb 17 12:46:35.069819 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069585 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-host-run-netns\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.069819 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069605 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-host-cni-netd\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.069819 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069643 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8738b60c-920d-4bc5-933f-01d0fb8d68d3-host\") pod \"node-ca-qjw6p\" (UID: \"8738b60c-920d-4bc5-933f-01d0fb8d68d3\") " pod="openshift-image-registry/node-ca-qjw6p" Feb 17 12:46:35.069819 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069676 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/3d7ab4ec-e6ac-454b-8724-36a22d414dae-sys-fs\") pod \"aws-ebs-csi-driver-node-ghtvv\" (UID: \"3d7ab4ec-e6ac-454b-8724-36a22d414dae\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" Feb 17 12:46:35.069819 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069704 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdsf2\" (UniqueName: \"kubernetes.io/projected/f1e10a0f-13dd-44a0-9d79-e6375e12ac24-kube-api-access-vdsf2\") pod \"network-check-target-f2tlm\" (UID: \"f1e10a0f-13dd-44a0-9d79-e6375e12ac24\") " pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:46:35.070262 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069907 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"tuned-dockercfg-rs2lg\"" Feb 17 12:46:35.070262 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069923 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"openshift-service-ca.crt\"" Feb 17 12:46:35.070262 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.069947 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-node-tuning-operator\"/\"kube-root-ca.crt\"" Feb 17 12:46:35.070752 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.070733 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-wfszv" Feb 17 12:46:35.070837 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.070778 2572 reflector.go:430] "Caches populated" logger="kubernetes.io/kubelet-serving" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Feb 17 12:46:35.072231 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.072160 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"whereabouts-flatfile-config\"" Feb 17 12:46:35.072414 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.072396 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"default-cni-sysctl-allowlist\"" Feb 17 12:46:35.072486 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.072460 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-qdrgb\"" Feb 17 12:46:35.072543 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.072511 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"cni-copy-resources\"" Feb 17 12:46:35.072543 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.072463 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"kube-root-ca.crt\"" Feb 17 12:46:35.072642 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.072404 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"openshift-service-ca.crt\"" Feb 17 12:46:35.073084 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.073070 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.074514 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.074497 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"multus-daemon-config\"" Feb 17 12:46:35.074940 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.074927 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"default-dockercfg-nqdzj\"" Feb 17 12:46:35.100235 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.100212 2572 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kubelet-serving" csr="csr-qzd4p" Feb 17 12:46:35.111349 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.111329 2572 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kubelet-serving" csr="csr-qzd4p" Feb 17 12:46:35.128747 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.128692 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-131-118.ec2.internal" event={"ID":"cde915e910705d98ba55e7806902bce3","Type":"ContainerStarted","Data":"0314b54a89b423bfd4bcf164627445db57712ff661aa0160b447d8246e97327b"} Feb 17 12:46:35.129549 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.129528 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-131-118.ec2.internal" event={"ID":"9b1938025d97d9e77a432753c9d25dde","Type":"ContainerStarted","Data":"b4e60fdd9faf79f78250c6ac3d28caf5b0161a5723f4b150444a7dc020e4d581"} Feb 17 12:46:35.165367 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.165342 2572 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Feb 17 12:46:35.170071 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170051 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs\") pod \"network-metrics-daemon-vmpwv\" (UID: \"685523da-d834-463a-8a1a-ef70069c19c5\") " pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:46:35.170153 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170079 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ba1edd82-a8f5-4e73-ac84-987cbbec0939-host-slash\") pod \"iptables-alerter-fm5zs\" (UID: \"ba1edd82-a8f5-4e73-ac84-987cbbec0939\") " pod="openshift-network-operator/iptables-alerter-fm5zs" Feb 17 12:46:35.170153 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170096 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-run-ovn\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.170153 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170131 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-2m6g7\" (UniqueName: \"kubernetes.io/projected/01eb7d37-a76a-4933-87df-44cf8a8d3152-kube-api-access-2m6g7\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.170294 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170153 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-host-var-lib-cni-bin\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.170294 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170193 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-run-ovn\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.170294 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170203 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ba1edd82-a8f5-4e73-ac84-987cbbec0939-host-slash\") pod \"iptables-alerter-fm5zs\" (UID: \"ba1edd82-a8f5-4e73-ac84-987cbbec0939\") " pod="openshift-network-operator/iptables-alerter-fm5zs" Feb 17 12:46:35.170294 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:35.170214 2572 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 12:46:35.170294 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170196 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-host-var-lib-cni-multus\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.170526 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170290 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/3d7ab4ec-e6ac-454b-8724-36a22d414dae-socket-dir\") pod \"aws-ebs-csi-driver-node-ghtvv\" (UID: \"3d7ab4ec-e6ac-454b-8724-36a22d414dae\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" Feb 17 12:46:35.170526 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:35.170338 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs podName:685523da-d834-463a-8a1a-ef70069c19c5 nodeName:}" failed. No retries permitted until 2026-02-17 12:46:35.670309309 +0000 UTC m=+2.020720128 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs") pod "network-metrics-daemon-vmpwv" (UID: "685523da-d834-463a-8a1a-ef70069c19c5") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 12:46:35.170526 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170372 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/3d7ab4ec-e6ac-454b-8724-36a22d414dae-registration-dir\") pod \"aws-ebs-csi-driver-node-ghtvv\" (UID: \"3d7ab4ec-e6ac-454b-8724-36a22d414dae\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" Feb 17 12:46:35.170526 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170408 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/9751ddaa-e412-4e4e-b97a-f4e5a7a050a9-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-wfszv\" (UID: \"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9\") " pod="openshift-multus/multus-additional-cni-plugins-wfszv" Feb 17 12:46:35.170526 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170431 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/3d7ab4ec-e6ac-454b-8724-36a22d414dae-socket-dir\") pod \"aws-ebs-csi-driver-node-ghtvv\" (UID: \"3d7ab4ec-e6ac-454b-8724-36a22d414dae\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" Feb 17 12:46:35.170526 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170439 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-os-release\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.170526 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170468 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/3d7ab4ec-e6ac-454b-8724-36a22d414dae-registration-dir\") pod \"aws-ebs-csi-driver-node-ghtvv\" (UID: \"3d7ab4ec-e6ac-454b-8724-36a22d414dae\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" Feb 17 12:46:35.170526 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170469 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-host-var-lib-kubelet\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.170526 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170514 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-2fcrg\" (UniqueName: \"kubernetes.io/projected/22f3475b-a5d2-4d66-818f-5010e57aac8e-kube-api-access-2fcrg\") pod \"node-resolver-jqfj5\" (UID: \"22f3475b-a5d2-4d66-818f-5010e57aac8e\") " pod="openshift-dns/node-resolver-jqfj5" Feb 17 12:46:35.170961 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170576 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-etc-openvswitch\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.170961 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170614 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-etc-openvswitch\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.170961 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170646 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.170961 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170677 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3d7ab4ec-e6ac-454b-8724-36a22d414dae-kubelet-dir\") pod \"aws-ebs-csi-driver-node-ghtvv\" (UID: \"3d7ab4ec-e6ac-454b-8724-36a22d414dae\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" Feb 17 12:46:35.170961 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170698 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/3d7ab4ec-e6ac-454b-8724-36a22d414dae-etc-selinux\") pod \"aws-ebs-csi-driver-node-ghtvv\" (UID: \"3d7ab4ec-e6ac-454b-8724-36a22d414dae\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" Feb 17 12:46:35.170961 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170729 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-multus-socket-dir-parent\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.170961 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170740 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-selinux\" (UniqueName: \"kubernetes.io/host-path/3d7ab4ec-e6ac-454b-8724-36a22d414dae-etc-selinux\") pod \"aws-ebs-csi-driver-node-ghtvv\" (UID: \"3d7ab4ec-e6ac-454b-8724-36a22d414dae\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" Feb 17 12:46:35.170961 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170750 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-host-run-k8s-cni-cncf-io\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.170961 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170746 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.170961 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170782 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3d7ab4ec-e6ac-454b-8724-36a22d414dae-kubelet-dir\") pod \"aws-ebs-csi-driver-node-ghtvv\" (UID: \"3d7ab4ec-e6ac-454b-8724-36a22d414dae\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" Feb 17 12:46:35.170961 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170802 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-host-run-multus-certs\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.170961 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170834 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-var-lib-openvswitch\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.170961 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170860 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/01eb7d37-a76a-4933-87df-44cf8a8d3152-env-overrides\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.170961 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170880 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-var-lib-openvswitch\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.170961 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170884 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-etc-modprobe-d\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.170961 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.170909 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-etc-sysctl-d\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.171730 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171005 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-sys\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.171730 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171023 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9751ddaa-e412-4e4e-b97a-f4e5a7a050a9-cnibin\") pod \"multus-additional-cni-plugins-wfszv\" (UID: \"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9\") " pod="openshift-multus/multus-additional-cni-plugins-wfszv" Feb 17 12:46:35.171730 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171047 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-79v2q\" (UniqueName: \"kubernetes.io/projected/ba1edd82-a8f5-4e73-ac84-987cbbec0939-kube-api-access-79v2q\") pod \"iptables-alerter-fm5zs\" (UID: \"ba1edd82-a8f5-4e73-ac84-987cbbec0939\") " pod="openshift-network-operator/iptables-alerter-fm5zs" Feb 17 12:46:35.171730 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171066 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-run-systemd\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.171730 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171083 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-run-openvswitch\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.171730 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171099 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-host\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.171730 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171125 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-run-systemd\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.171730 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171134 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-run-openvswitch\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.171730 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171134 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-host-kubelet\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.171730 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171165 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-host-kubelet\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.171730 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171189 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8738b60c-920d-4bc5-933f-01d0fb8d68d3-serviceca\") pod \"node-ca-qjw6p\" (UID: \"8738b60c-920d-4bc5-933f-01d0fb8d68d3\") " pod="openshift-image-registry/node-ca-qjw6p" Feb 17 12:46:35.171730 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171235 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-2cz42\" (UniqueName: \"kubernetes.io/projected/8738b60c-920d-4bc5-933f-01d0fb8d68d3-kube-api-access-2cz42\") pod \"node-ca-qjw6p\" (UID: \"8738b60c-920d-4bc5-933f-01d0fb8d68d3\") " pod="openshift-image-registry/node-ca-qjw6p" Feb 17 12:46:35.171730 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171259 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-lib-modules\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.171730 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171277 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/e35ae435-3708-4b40-93e1-40c8b34c3fa4-tmp\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.171730 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171303 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9751ddaa-e412-4e4e-b97a-f4e5a7a050a9-os-release\") pod \"multus-additional-cni-plugins-wfszv\" (UID: \"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9\") " pod="openshift-multus/multus-additional-cni-plugins-wfszv" Feb 17 12:46:35.171730 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171331 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/ba1edd82-a8f5-4e73-ac84-987cbbec0939-iptables-alerter-script\") pod \"iptables-alerter-fm5zs\" (UID: \"ba1edd82-a8f5-4e73-ac84-987cbbec0939\") " pod="openshift-network-operator/iptables-alerter-fm5zs" Feb 17 12:46:35.171730 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171347 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-systemd-units\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.172549 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171370 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/22f3475b-a5d2-4d66-818f-5010e57aac8e-tmp-dir\") pod \"node-resolver-jqfj5\" (UID: \"22f3475b-a5d2-4d66-818f-5010e57aac8e\") " pod="openshift-dns/node-resolver-jqfj5" Feb 17 12:46:35.172549 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171389 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-host-run-netns\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.172549 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171404 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-host-cni-netd\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.172549 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171420 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/01eb7d37-a76a-4933-87df-44cf8a8d3152-env-overrides\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.172549 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171439 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-host-run-netns\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.172549 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171471 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-systemd-units\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.172549 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171533 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-host-cni-netd\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.172549 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171562 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/3d7ab4ec-e6ac-454b-8724-36a22d414dae-sys-fs\") pod \"aws-ebs-csi-driver-node-ghtvv\" (UID: \"3d7ab4ec-e6ac-454b-8724-36a22d414dae\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" Feb 17 12:46:35.172549 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171580 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8738b60c-920d-4bc5-933f-01d0fb8d68d3-serviceca\") pod \"node-ca-qjw6p\" (UID: \"8738b60c-920d-4bc5-933f-01d0fb8d68d3\") " pod="openshift-image-registry/node-ca-qjw6p" Feb 17 12:46:35.172549 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171598 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-system-cni-dir\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.172549 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171626 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-multus-cni-dir\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.172549 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171636 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys-fs\" (UniqueName: \"kubernetes.io/host-path/3d7ab4ec-e6ac-454b-8724-36a22d414dae-sys-fs\") pod \"aws-ebs-csi-driver-node-ghtvv\" (UID: \"3d7ab4ec-e6ac-454b-8724-36a22d414dae\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" Feb 17 12:46:35.172549 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171651 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/30e805d5-0857-42ef-a870-6492d444e008-cni-binary-copy\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.172549 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171684 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-etc-kubernetes\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.172549 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171687 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/22f3475b-a5d2-4d66-818f-5010e57aac8e-tmp-dir\") pod \"node-resolver-jqfj5\" (UID: \"22f3475b-a5d2-4d66-818f-5010e57aac8e\") " pod="openshift-dns/node-resolver-jqfj5" Feb 17 12:46:35.172549 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171716 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/e35ae435-3708-4b40-93e1-40c8b34c3fa4-etc-tuned\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.172549 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171755 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/22f3475b-a5d2-4d66-818f-5010e57aac8e-hosts-file\") pod \"node-resolver-jqfj5\" (UID: \"22f3475b-a5d2-4d66-818f-5010e57aac8e\") " pod="openshift-dns/node-resolver-jqfj5" Feb 17 12:46:35.172549 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171781 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-log-socket\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.173201 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171801 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/22f3475b-a5d2-4d66-818f-5010e57aac8e-hosts-file\") pod \"node-resolver-jqfj5\" (UID: \"22f3475b-a5d2-4d66-818f-5010e57aac8e\") " pod="openshift-dns/node-resolver-jqfj5" Feb 17 12:46:35.173201 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171806 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-host-run-ovn-kubernetes\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.173201 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171840 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/01eb7d37-a76a-4933-87df-44cf8a8d3152-ovnkube-config\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.173201 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171845 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-host-run-ovn-kubernetes\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.173201 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171846 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/ba1edd82-a8f5-4e73-ac84-987cbbec0939-iptables-alerter-script\") pod \"iptables-alerter-fm5zs\" (UID: \"ba1edd82-a8f5-4e73-ac84-987cbbec0939\") " pod="openshift-network-operator/iptables-alerter-fm5zs" Feb 17 12:46:35.173201 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171850 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-log-socket\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.173201 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171868 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/3d7ab4ec-e6ac-454b-8724-36a22d414dae-device-dir\") pod \"aws-ebs-csi-driver-node-ghtvv\" (UID: \"3d7ab4ec-e6ac-454b-8724-36a22d414dae\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" Feb 17 12:46:35.173201 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171903 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/3d7ab4ec-e6ac-454b-8724-36a22d414dae-device-dir\") pod \"aws-ebs-csi-driver-node-ghtvv\" (UID: \"3d7ab4ec-e6ac-454b-8724-36a22d414dae\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" Feb 17 12:46:35.173201 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171930 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-etc-kubernetes\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.173201 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171947 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-var-lib-kubelet\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.173201 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171965 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-x7jz4\" (UniqueName: \"kubernetes.io/projected/685523da-d834-463a-8a1a-ef70069c19c5-kube-api-access-x7jz4\") pod \"network-metrics-daemon-vmpwv\" (UID: \"685523da-d834-463a-8a1a-ef70069c19c5\") " pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:46:35.173201 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.171989 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-node-log\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.173201 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172013 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-etc-sysconfig\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.173201 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172036 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-etc-sysctl-conf\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.173201 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172061 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-run\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.173201 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172069 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-node-log\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.173201 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172085 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-host-run-netns\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.173678 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172125 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-hostroot\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.173678 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172153 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/01eb7d37-a76a-4933-87df-44cf8a8d3152-ovnkube-script-lib\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.173678 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172205 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-qksk7\" (UniqueName: \"kubernetes.io/projected/3d7ab4ec-e6ac-454b-8724-36a22d414dae-kube-api-access-qksk7\") pod \"aws-ebs-csi-driver-node-ghtvv\" (UID: \"3d7ab4ec-e6ac-454b-8724-36a22d414dae\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" Feb 17 12:46:35.173678 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172233 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9751ddaa-e412-4e4e-b97a-f4e5a7a050a9-cni-binary-copy\") pod \"multus-additional-cni-plugins-wfszv\" (UID: \"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9\") " pod="openshift-multus/multus-additional-cni-plugins-wfszv" Feb 17 12:46:35.173678 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172259 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m62fv\" (UniqueName: \"kubernetes.io/projected/30e805d5-0857-42ef-a870-6492d444e008-kube-api-access-m62fv\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.173678 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172288 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/30392cb9-7527-44af-833b-567cd508520e-agent-certs\") pod \"konnectivity-agent-kmsrd\" (UID: \"30392cb9-7527-44af-833b-567cd508520e\") " pod="kube-system/konnectivity-agent-kmsrd" Feb 17 12:46:35.173678 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172313 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/30392cb9-7527-44af-833b-567cd508520e-konnectivity-ca\") pod \"konnectivity-agent-kmsrd\" (UID: \"30392cb9-7527-44af-833b-567cd508520e\") " pod="kube-system/konnectivity-agent-kmsrd" Feb 17 12:46:35.173678 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172315 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/01eb7d37-a76a-4933-87df-44cf8a8d3152-ovnkube-config\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.173678 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172336 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9751ddaa-e412-4e4e-b97a-f4e5a7a050a9-tuning-conf-dir\") pod \"multus-additional-cni-plugins-wfszv\" (UID: \"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9\") " pod="openshift-multus/multus-additional-cni-plugins-wfszv" Feb 17 12:46:35.173678 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172363 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qq8s\" (UniqueName: \"kubernetes.io/projected/9751ddaa-e412-4e4e-b97a-f4e5a7a050a9-kube-api-access-7qq8s\") pod \"multus-additional-cni-plugins-wfszv\" (UID: \"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9\") " pod="openshift-multus/multus-additional-cni-plugins-wfszv" Feb 17 12:46:35.173678 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172413 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-multus-conf-dir\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.173678 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172455 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/30e805d5-0857-42ef-a870-6492d444e008-multus-daemon-config\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.173678 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172478 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-host-slash\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.173678 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172493 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-host-cni-bin\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.173678 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172513 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/01eb7d37-a76a-4933-87df-44cf8a8d3152-ovn-node-metrics-cert\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.173678 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172540 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-host-slash\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.173678 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172542 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rml2h\" (UniqueName: \"kubernetes.io/projected/e35ae435-3708-4b40-93e1-40c8b34c3fa4-kube-api-access-rml2h\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.174134 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172575 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9751ddaa-e412-4e4e-b97a-f4e5a7a050a9-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-wfszv\" (UID: \"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9\") " pod="openshift-multus/multus-additional-cni-plugins-wfszv" Feb 17 12:46:35.174134 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172592 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-cnibin\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.174134 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172610 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8738b60c-920d-4bc5-933f-01d0fb8d68d3-host\") pod \"node-ca-qjw6p\" (UID: \"8738b60c-920d-4bc5-933f-01d0fb8d68d3\") " pod="openshift-image-registry/node-ca-qjw6p" Feb 17 12:46:35.174134 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172544 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/01eb7d37-a76a-4933-87df-44cf8a8d3152-host-cni-bin\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.174134 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172635 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vdsf2\" (UniqueName: \"kubernetes.io/projected/f1e10a0f-13dd-44a0-9d79-e6375e12ac24-kube-api-access-vdsf2\") pod \"network-check-target-f2tlm\" (UID: \"f1e10a0f-13dd-44a0-9d79-e6375e12ac24\") " pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:46:35.174134 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172640 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/01eb7d37-a76a-4933-87df-44cf8a8d3152-ovnkube-script-lib\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.174134 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172660 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8738b60c-920d-4bc5-933f-01d0fb8d68d3-host\") pod \"node-ca-qjw6p\" (UID: \"8738b60c-920d-4bc5-933f-01d0fb8d68d3\") " pod="openshift-image-registry/node-ca-qjw6p" Feb 17 12:46:35.174134 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172661 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-etc-systemd\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.174134 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172680 2572 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Feb 17 12:46:35.174134 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172719 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9751ddaa-e412-4e4e-b97a-f4e5a7a050a9-system-cni-dir\") pod \"multus-additional-cni-plugins-wfszv\" (UID: \"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9\") " pod="openshift-multus/multus-additional-cni-plugins-wfszv" Feb 17 12:46:35.174134 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.172834 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"konnectivity-ca\" (UniqueName: \"kubernetes.io/configmap/30392cb9-7527-44af-833b-567cd508520e-konnectivity-ca\") pod \"konnectivity-agent-kmsrd\" (UID: \"30392cb9-7527-44af-833b-567cd508520e\") " pod="kube-system/konnectivity-agent-kmsrd" Feb 17 12:46:35.175371 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.175350 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/01eb7d37-a76a-4933-87df-44cf8a8d3152-ovn-node-metrics-cert\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.175459 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.175398 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"agent-certs\" (UniqueName: \"kubernetes.io/secret/30392cb9-7527-44af-833b-567cd508520e-agent-certs\") pod \"konnectivity-agent-kmsrd\" (UID: \"30392cb9-7527-44af-833b-567cd508520e\") " pod="kube-system/konnectivity-agent-kmsrd" Feb 17 12:46:35.178794 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:35.178771 2572 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 12:46:35.178892 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:35.178799 2572 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 12:46:35.178892 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:35.178813 2572 projected.go:194] Error preparing data for projected volume kube-api-access-vdsf2 for pod openshift-network-diagnostics/network-check-target-f2tlm: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 12:46:35.178892 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:35.178886 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f1e10a0f-13dd-44a0-9d79-e6375e12ac24-kube-api-access-vdsf2 podName:f1e10a0f-13dd-44a0-9d79-e6375e12ac24 nodeName:}" failed. No retries permitted until 2026-02-17 12:46:35.678867745 +0000 UTC m=+2.029278558 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-vdsf2" (UniqueName: "kubernetes.io/projected/f1e10a0f-13dd-44a0-9d79-e6375e12ac24-kube-api-access-vdsf2") pod "network-check-target-f2tlm" (UID: "f1e10a0f-13dd-44a0-9d79-e6375e12ac24") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 12:46:35.179799 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.179773 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7jz4\" (UniqueName: \"kubernetes.io/projected/685523da-d834-463a-8a1a-ef70069c19c5-kube-api-access-x7jz4\") pod \"network-metrics-daemon-vmpwv\" (UID: \"685523da-d834-463a-8a1a-ef70069c19c5\") " pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:46:35.179799 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.179782 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cz42\" (UniqueName: \"kubernetes.io/projected/8738b60c-920d-4bc5-933f-01d0fb8d68d3-kube-api-access-2cz42\") pod \"node-ca-qjw6p\" (UID: \"8738b60c-920d-4bc5-933f-01d0fb8d68d3\") " pod="openshift-image-registry/node-ca-qjw6p" Feb 17 12:46:35.179908 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.179811 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fcrg\" (UniqueName: \"kubernetes.io/projected/22f3475b-a5d2-4d66-818f-5010e57aac8e-kube-api-access-2fcrg\") pod \"node-resolver-jqfj5\" (UID: \"22f3475b-a5d2-4d66-818f-5010e57aac8e\") " pod="openshift-dns/node-resolver-jqfj5" Feb 17 12:46:35.180326 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.180308 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-qksk7\" (UniqueName: \"kubernetes.io/projected/3d7ab4ec-e6ac-454b-8724-36a22d414dae-kube-api-access-qksk7\") pod \"aws-ebs-csi-driver-node-ghtvv\" (UID: \"3d7ab4ec-e6ac-454b-8724-36a22d414dae\") " pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" Feb 17 12:46:35.180394 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.180309 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m6g7\" (UniqueName: \"kubernetes.io/projected/01eb7d37-a76a-4933-87df-44cf8a8d3152-kube-api-access-2m6g7\") pod \"ovnkube-node-lb8m9\" (UID: \"01eb7d37-a76a-4933-87df-44cf8a8d3152\") " pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.180394 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.180339 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-79v2q\" (UniqueName: \"kubernetes.io/projected/ba1edd82-a8f5-4e73-ac84-987cbbec0939-kube-api-access-79v2q\") pod \"iptables-alerter-fm5zs\" (UID: \"ba1edd82-a8f5-4e73-ac84-987cbbec0939\") " pod="openshift-network-operator/iptables-alerter-fm5zs" Feb 17 12:46:35.273502 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.273404 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-etc-sysconfig\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.273502 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.273444 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-etc-sysctl-conf\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.273502 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.273460 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-run\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.273756 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.273544 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-etc-sysconfig\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.273756 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.273567 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-host-run-netns\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.273756 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.273576 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-etc-sysctl-conf\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.273756 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.273596 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-hostroot\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.273756 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.273612 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-host-run-netns\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.273756 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.273613 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-run\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.273756 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.273621 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9751ddaa-e412-4e4e-b97a-f4e5a7a050a9-cni-binary-copy\") pod \"multus-additional-cni-plugins-wfszv\" (UID: \"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9\") " pod="openshift-multus/multus-additional-cni-plugins-wfszv" Feb 17 12:46:35.273756 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.273643 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-m62fv\" (UniqueName: \"kubernetes.io/projected/30e805d5-0857-42ef-a870-6492d444e008-kube-api-access-m62fv\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.273756 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.273649 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-hostroot\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.273756 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.273662 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9751ddaa-e412-4e4e-b97a-f4e5a7a050a9-tuning-conf-dir\") pod \"multus-additional-cni-plugins-wfszv\" (UID: \"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9\") " pod="openshift-multus/multus-additional-cni-plugins-wfszv" Feb 17 12:46:35.273756 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.273691 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-7qq8s\" (UniqueName: \"kubernetes.io/projected/9751ddaa-e412-4e4e-b97a-f4e5a7a050a9-kube-api-access-7qq8s\") pod \"multus-additional-cni-plugins-wfszv\" (UID: \"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9\") " pod="openshift-multus/multus-additional-cni-plugins-wfszv" Feb 17 12:46:35.273756 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.273708 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-multus-conf-dir\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.274326 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.273724 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/30e805d5-0857-42ef-a870-6492d444e008-multus-daemon-config\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.274326 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.273827 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-multus-conf-dir\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.274326 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.273842 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rml2h\" (UniqueName: \"kubernetes.io/projected/e35ae435-3708-4b40-93e1-40c8b34c3fa4-kube-api-access-rml2h\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.274326 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.273882 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9751ddaa-e412-4e4e-b97a-f4e5a7a050a9-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-wfszv\" (UID: \"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9\") " pod="openshift-multus/multus-additional-cni-plugins-wfszv" Feb 17 12:46:35.274326 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274085 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-cnibin\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.274326 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274153 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-etc-systemd\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.274326 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274171 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-cnibin\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.274326 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274193 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9751ddaa-e412-4e4e-b97a-f4e5a7a050a9-system-cni-dir\") pod \"multus-additional-cni-plugins-wfszv\" (UID: \"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9\") " pod="openshift-multus/multus-additional-cni-plugins-wfszv" Feb 17 12:46:35.274326 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274232 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-etc-systemd\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.274326 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274235 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-host-var-lib-cni-bin\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.274326 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274255 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9751ddaa-e412-4e4e-b97a-f4e5a7a050a9-cni-binary-copy\") pod \"multus-additional-cni-plugins-wfszv\" (UID: \"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9\") " pod="openshift-multus/multus-additional-cni-plugins-wfszv" Feb 17 12:46:35.274326 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274281 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-host-var-lib-cni-multus\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.274326 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274278 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-host-var-lib-cni-bin\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.274326 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274323 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-host-var-lib-cni-multus\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.274965 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274322 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/9751ddaa-e412-4e4e-b97a-f4e5a7a050a9-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-wfszv\" (UID: \"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9\") " pod="openshift-multus/multus-additional-cni-plugins-wfszv" Feb 17 12:46:35.274965 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274354 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9751ddaa-e412-4e4e-b97a-f4e5a7a050a9-tuning-conf-dir\") pod \"multus-additional-cni-plugins-wfszv\" (UID: \"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9\") " pod="openshift-multus/multus-additional-cni-plugins-wfszv" Feb 17 12:46:35.274965 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274364 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-os-release\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.274965 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274407 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-os-release\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.274965 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274414 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-host-var-lib-kubelet\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.274965 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274339 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9751ddaa-e412-4e4e-b97a-f4e5a7a050a9-system-cni-dir\") pod \"multus-additional-cni-plugins-wfszv\" (UID: \"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9\") " pod="openshift-multus/multus-additional-cni-plugins-wfszv" Feb 17 12:46:35.274965 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274440 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-multus-socket-dir-parent\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.274965 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274447 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/30e805d5-0857-42ef-a870-6492d444e008-multus-daemon-config\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.274965 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274380 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9751ddaa-e412-4e4e-b97a-f4e5a7a050a9-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-wfszv\" (UID: \"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9\") " pod="openshift-multus/multus-additional-cni-plugins-wfszv" Feb 17 12:46:35.274965 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274445 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-host-var-lib-kubelet\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.274965 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274528 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-host-run-k8s-cni-cncf-io\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.274965 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274551 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-host-run-multus-certs\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.274965 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274564 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-multus-socket-dir-parent\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.274965 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274573 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-etc-modprobe-d\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.274965 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274566 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-host-run-k8s-cni-cncf-io\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.274965 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274587 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-host-run-multus-certs\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.274965 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274596 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-etc-sysctl-d\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.275457 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274625 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-sys\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.275457 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274647 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9751ddaa-e412-4e4e-b97a-f4e5a7a050a9-cnibin\") pod \"multus-additional-cni-plugins-wfszv\" (UID: \"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9\") " pod="openshift-multus/multus-additional-cni-plugins-wfszv" Feb 17 12:46:35.275457 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274652 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-etc-modprobe-d\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.275457 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274672 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-host\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.275457 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274687 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-sys\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.275457 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274695 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9751ddaa-e412-4e4e-b97a-f4e5a7a050a9-cnibin\") pod \"multus-additional-cni-plugins-wfszv\" (UID: \"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9\") " pod="openshift-multus/multus-additional-cni-plugins-wfszv" Feb 17 12:46:35.275457 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274700 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-etc-sysctl-d\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.275457 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274722 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-lib-modules\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.275457 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274739 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/e35ae435-3708-4b40-93e1-40c8b34c3fa4-tmp\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.275457 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274750 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-host\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.275457 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274755 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9751ddaa-e412-4e4e-b97a-f4e5a7a050a9-os-release\") pod \"multus-additional-cni-plugins-wfszv\" (UID: \"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9\") " pod="openshift-multus/multus-additional-cni-plugins-wfszv" Feb 17 12:46:35.275457 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274792 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/9751ddaa-e412-4e4e-b97a-f4e5a7a050a9-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-wfszv\" (UID: \"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9\") " pod="openshift-multus/multus-additional-cni-plugins-wfszv" Feb 17 12:46:35.275457 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274812 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9751ddaa-e412-4e4e-b97a-f4e5a7a050a9-os-release\") pod \"multus-additional-cni-plugins-wfszv\" (UID: \"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9\") " pod="openshift-multus/multus-additional-cni-plugins-wfszv" Feb 17 12:46:35.275457 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274820 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-lib-modules\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.275457 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274835 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-system-cni-dir\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.275457 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274852 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-multus-cni-dir\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.275457 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274867 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/30e805d5-0857-42ef-a870-6492d444e008-cni-binary-copy\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.275457 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274865 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-system-cni-dir\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.275934 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274891 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-etc-kubernetes\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.275934 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274907 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/e35ae435-3708-4b40-93e1-40c8b34c3fa4-etc-tuned\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.275934 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274926 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-etc-kubernetes\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.275934 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274946 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-var-lib-kubelet\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.275934 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.274961 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-multus-cni-dir\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.275934 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.275014 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-var-lib-kubelet\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.275934 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.275015 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/30e805d5-0857-42ef-a870-6492d444e008-etc-kubernetes\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.275934 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.275065 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e35ae435-3708-4b40-93e1-40c8b34c3fa4-etc-kubernetes\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.275934 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.275317 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/30e805d5-0857-42ef-a870-6492d444e008-cni-binary-copy\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.276730 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.276702 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/e35ae435-3708-4b40-93e1-40c8b34c3fa4-tmp\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.276793 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.276779 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/e35ae435-3708-4b40-93e1-40c8b34c3fa4-etc-tuned\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.282551 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.282515 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-rml2h\" (UniqueName: \"kubernetes.io/projected/e35ae435-3708-4b40-93e1-40c8b34c3fa4-kube-api-access-rml2h\") pod \"tuned-48j6f\" (UID: \"e35ae435-3708-4b40-93e1-40c8b34c3fa4\") " pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.283526 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.283502 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qq8s\" (UniqueName: \"kubernetes.io/projected/9751ddaa-e412-4e4e-b97a-f4e5a7a050a9-kube-api-access-7qq8s\") pod \"multus-additional-cni-plugins-wfszv\" (UID: \"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9\") " pod="openshift-multus/multus-additional-cni-plugins-wfszv" Feb 17 12:46:35.283675 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.283656 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-m62fv\" (UniqueName: \"kubernetes.io/projected/30e805d5-0857-42ef-a870-6492d444e008-kube-api-access-m62fv\") pod \"multus-tsb7s\" (UID: \"30e805d5-0857-42ef-a870-6492d444e008\") " pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.362053 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.362013 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-jqfj5" Feb 17 12:46:35.368943 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.368917 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-qjw6p" Feb 17 12:46:35.369819 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:35.369731 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22f3475b_a5d2_4d66_818f_5010e57aac8e.slice/crio-f4e71d7c069bfafa5ef77227a771004039db45b833095db4896c78cafecbd216 WatchSource:0}: Error finding container f4e71d7c069bfafa5ef77227a771004039db45b833095db4896c78cafecbd216: Status 404 returned error can't find the container with id f4e71d7c069bfafa5ef77227a771004039db45b833095db4896c78cafecbd216 Feb 17 12:46:35.376451 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.376429 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-fm5zs" Feb 17 12:46:35.376744 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:35.376714 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8738b60c_920d_4bc5_933f_01d0fb8d68d3.slice/crio-4571f746fcfe0037446fcf9eac768fa1f1994a2d1c1e5fcae12a9d6db8de0f9c WatchSource:0}: Error finding container 4571f746fcfe0037446fcf9eac768fa1f1994a2d1c1e5fcae12a9d6db8de0f9c: Status 404 returned error can't find the container with id 4571f746fcfe0037446fcf9eac768fa1f1994a2d1c1e5fcae12a9d6db8de0f9c Feb 17 12:46:35.381597 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.381574 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:46:35.383857 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:35.383774 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podba1edd82_a8f5_4e73_ac84_987cbbec0939.slice/crio-03c1270b0c0d4c5e51c2db70e0e5b8f0a3c5f6c9b01734be2e4f1c0a1ba66c42 WatchSource:0}: Error finding container 03c1270b0c0d4c5e51c2db70e0e5b8f0a3c5f6c9b01734be2e4f1c0a1ba66c42: Status 404 returned error can't find the container with id 03c1270b0c0d4c5e51c2db70e0e5b8f0a3c5f6c9b01734be2e4f1c0a1ba66c42 Feb 17 12:46:35.386252 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.386236 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/konnectivity-agent-kmsrd" Feb 17 12:46:35.388943 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:35.388918 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01eb7d37_a76a_4933_87df_44cf8a8d3152.slice/crio-8541e115782d707d971c51eb769a80051b48c77301abf2e63ff96ef41d4c9063 WatchSource:0}: Error finding container 8541e115782d707d971c51eb769a80051b48c77301abf2e63ff96ef41d4c9063: Status 404 returned error can't find the container with id 8541e115782d707d971c51eb769a80051b48c77301abf2e63ff96ef41d4c9063 Feb 17 12:46:35.392193 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.392075 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" Feb 17 12:46:35.393814 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:35.393791 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30392cb9_7527_44af_833b_567cd508520e.slice/crio-ec6749e35e72aceb1158b1fa44bb79c182c4968c8d453081a8e146cacdddfea1 WatchSource:0}: Error finding container ec6749e35e72aceb1158b1fa44bb79c182c4968c8d453081a8e146cacdddfea1: Status 404 returned error can't find the container with id ec6749e35e72aceb1158b1fa44bb79c182c4968c8d453081a8e146cacdddfea1 Feb 17 12:46:35.400530 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.400155 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-48j6f" Feb 17 12:46:35.400530 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:35.400436 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d7ab4ec_e6ac_454b_8724_36a22d414dae.slice/crio-31df1cbb33ea39e41be83cb6c7a93196649b45000d0540030039cbfed3b5e08c WatchSource:0}: Error finding container 31df1cbb33ea39e41be83cb6c7a93196649b45000d0540030039cbfed3b5e08c: Status 404 returned error can't find the container with id 31df1cbb33ea39e41be83cb6c7a93196649b45000d0540030039cbfed3b5e08c Feb 17 12:46:35.406181 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.406159 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-wfszv" Feb 17 12:46:35.407206 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:35.407183 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode35ae435_3708_4b40_93e1_40c8b34c3fa4.slice/crio-1c65eff69c3dc7fdb029302b0b041fce46abece2bb4075434a938f916ec32dc7 WatchSource:0}: Error finding container 1c65eff69c3dc7fdb029302b0b041fce46abece2bb4075434a938f916ec32dc7: Status 404 returned error can't find the container with id 1c65eff69c3dc7fdb029302b0b041fce46abece2bb4075434a938f916ec32dc7 Feb 17 12:46:35.409552 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.409521 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-tsb7s" Feb 17 12:46:35.414466 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:35.414444 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9751ddaa_e412_4e4e_b97a_f4e5a7a050a9.slice/crio-10a58c23ecfe62d60e176fb8cccee70b36e4db5d663a6f00fa6fdb71960c34f0 WatchSource:0}: Error finding container 10a58c23ecfe62d60e176fb8cccee70b36e4db5d663a6f00fa6fdb71960c34f0: Status 404 returned error can't find the container with id 10a58c23ecfe62d60e176fb8cccee70b36e4db5d663a6f00fa6fdb71960c34f0 Feb 17 12:46:35.417417 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:46:35.417396 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30e805d5_0857_42ef_a870_6492d444e008.slice/crio-cc2a4ce3a3eff037b9ca00382b72adf6e64f53e908852b148cc60d83fe93c36f WatchSource:0}: Error finding container cc2a4ce3a3eff037b9ca00382b72adf6e64f53e908852b148cc60d83fe93c36f: Status 404 returned error can't find the container with id cc2a4ce3a3eff037b9ca00382b72adf6e64f53e908852b148cc60d83fe93c36f Feb 17 12:46:35.678863 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.678750 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs\") pod \"network-metrics-daemon-vmpwv\" (UID: \"685523da-d834-463a-8a1a-ef70069c19c5\") " pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:46:35.679018 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:35.678973 2572 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 12:46:35.679074 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:35.679051 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs podName:685523da-d834-463a-8a1a-ef70069c19c5 nodeName:}" failed. No retries permitted until 2026-02-17 12:46:36.679031045 +0000 UTC m=+3.029441840 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs") pod "network-metrics-daemon-vmpwv" (UID: "685523da-d834-463a-8a1a-ef70069c19c5") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 12:46:35.780761 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.780062 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vdsf2\" (UniqueName: \"kubernetes.io/projected/f1e10a0f-13dd-44a0-9d79-e6375e12ac24-kube-api-access-vdsf2\") pod \"network-check-target-f2tlm\" (UID: \"f1e10a0f-13dd-44a0-9d79-e6375e12ac24\") " pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:46:35.780761 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:35.780304 2572 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 12:46:35.780761 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:35.780328 2572 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 12:46:35.780761 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:35.780341 2572 projected.go:194] Error preparing data for projected volume kube-api-access-vdsf2 for pod openshift-network-diagnostics/network-check-target-f2tlm: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 12:46:35.780761 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:35.780404 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f1e10a0f-13dd-44a0-9d79-e6375e12ac24-kube-api-access-vdsf2 podName:f1e10a0f-13dd-44a0-9d79-e6375e12ac24 nodeName:}" failed. No retries permitted until 2026-02-17 12:46:36.780385238 +0000 UTC m=+3.130796042 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-vdsf2" (UniqueName: "kubernetes.io/projected/f1e10a0f-13dd-44a0-9d79-e6375e12ac24-kube-api-access-vdsf2") pod "network-check-target-f2tlm" (UID: "f1e10a0f-13dd-44a0-9d79-e6375e12ac24") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 12:46:35.894603 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.891232 2572 reflector.go:430] "Caches populated" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" Feb 17 12:46:35.923020 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:35.922859 2572 reflector.go:430] "Caches populated" type="*v1.CSIDriver" reflector="k8s.io/client-go/informers/factory.go:160" Feb 17 12:46:36.112421 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:36.112119 2572 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kubelet-serving" expiration="2028-02-17 12:41:35 +0000 UTC" deadline="2027-10-10 22:03:40.351794982 +0000 UTC" Feb 17 12:46:36.112421 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:36.112371 2572 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kubelet-serving" sleep="14409h17m4.239429539s" Feb 17 12:46:36.138522 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:36.138481 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" event={"ID":"3d7ab4ec-e6ac-454b-8724-36a22d414dae","Type":"ContainerStarted","Data":"31df1cbb33ea39e41be83cb6c7a93196649b45000d0540030039cbfed3b5e08c"} Feb 17 12:46:36.140877 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:36.140855 2572 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Feb 17 12:46:36.144300 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:36.144257 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" event={"ID":"01eb7d37-a76a-4933-87df-44cf8a8d3152","Type":"ContainerStarted","Data":"8541e115782d707d971c51eb769a80051b48c77301abf2e63ff96ef41d4c9063"} Feb 17 12:46:36.148301 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:36.148249 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-qjw6p" event={"ID":"8738b60c-920d-4bc5-933f-01d0fb8d68d3","Type":"ContainerStarted","Data":"4571f746fcfe0037446fcf9eac768fa1f1994a2d1c1e5fcae12a9d6db8de0f9c"} Feb 17 12:46:36.161047 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:36.160984 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-jqfj5" event={"ID":"22f3475b-a5d2-4d66-818f-5010e57aac8e","Type":"ContainerStarted","Data":"f4e71d7c069bfafa5ef77227a771004039db45b833095db4896c78cafecbd216"} Feb 17 12:46:36.170200 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:36.170168 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-tsb7s" event={"ID":"30e805d5-0857-42ef-a870-6492d444e008","Type":"ContainerStarted","Data":"cc2a4ce3a3eff037b9ca00382b72adf6e64f53e908852b148cc60d83fe93c36f"} Feb 17 12:46:36.174525 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:36.174493 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wfszv" event={"ID":"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9","Type":"ContainerStarted","Data":"10a58c23ecfe62d60e176fb8cccee70b36e4db5d663a6f00fa6fdb71960c34f0"} Feb 17 12:46:36.183203 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:36.183174 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-48j6f" event={"ID":"e35ae435-3708-4b40-93e1-40c8b34c3fa4","Type":"ContainerStarted","Data":"1c65eff69c3dc7fdb029302b0b041fce46abece2bb4075434a938f916ec32dc7"} Feb 17 12:46:36.189627 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:36.189596 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/konnectivity-agent-kmsrd" event={"ID":"30392cb9-7527-44af-833b-567cd508520e","Type":"ContainerStarted","Data":"ec6749e35e72aceb1158b1fa44bb79c182c4968c8d453081a8e146cacdddfea1"} Feb 17 12:46:36.196010 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:36.195978 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-fm5zs" event={"ID":"ba1edd82-a8f5-4e73-ac84-987cbbec0939","Type":"ContainerStarted","Data":"03c1270b0c0d4c5e51c2db70e0e5b8f0a3c5f6c9b01734be2e4f1c0a1ba66c42"} Feb 17 12:46:36.689626 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:36.689587 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs\") pod \"network-metrics-daemon-vmpwv\" (UID: \"685523da-d834-463a-8a1a-ef70069c19c5\") " pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:46:36.689820 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:36.689747 2572 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 12:46:36.689820 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:36.689812 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs podName:685523da-d834-463a-8a1a-ef70069c19c5 nodeName:}" failed. No retries permitted until 2026-02-17 12:46:38.689792653 +0000 UTC m=+5.040203468 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs") pod "network-metrics-daemon-vmpwv" (UID: "685523da-d834-463a-8a1a-ef70069c19c5") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 12:46:36.790475 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:36.790437 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vdsf2\" (UniqueName: \"kubernetes.io/projected/f1e10a0f-13dd-44a0-9d79-e6375e12ac24-kube-api-access-vdsf2\") pod \"network-check-target-f2tlm\" (UID: \"f1e10a0f-13dd-44a0-9d79-e6375e12ac24\") " pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:46:36.790650 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:36.790636 2572 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 12:46:36.790722 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:36.790654 2572 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 12:46:36.790722 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:36.790667 2572 projected.go:194] Error preparing data for projected volume kube-api-access-vdsf2 for pod openshift-network-diagnostics/network-check-target-f2tlm: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 12:46:36.790837 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:36.790729 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f1e10a0f-13dd-44a0-9d79-e6375e12ac24-kube-api-access-vdsf2 podName:f1e10a0f-13dd-44a0-9d79-e6375e12ac24 nodeName:}" failed. No retries permitted until 2026-02-17 12:46:38.790709663 +0000 UTC m=+5.141120467 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-vdsf2" (UniqueName: "kubernetes.io/projected/f1e10a0f-13dd-44a0-9d79-e6375e12ac24-kube-api-access-vdsf2") pod "network-check-target-f2tlm" (UID: "f1e10a0f-13dd-44a0-9d79-e6375e12ac24") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 12:46:37.113621 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:37.113507 2572 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kubelet-serving" expiration="2028-02-17 12:41:35 +0000 UTC" deadline="2027-09-17 05:10:45.400308689 +0000 UTC" Feb 17 12:46:37.113621 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:37.113574 2572 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kubelet-serving" sleep="13840h24m8.286738198s" Feb 17 12:46:37.125984 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:37.125938 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:46:37.126180 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:37.126097 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vmpwv" podUID="685523da-d834-463a-8a1a-ef70069c19c5" Feb 17 12:46:37.126778 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:37.126751 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:46:37.126922 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:37.126903 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-f2tlm" podUID="f1e10a0f-13dd-44a0-9d79-e6375e12ac24" Feb 17 12:46:38.707858 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:38.707810 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs\") pod \"network-metrics-daemon-vmpwv\" (UID: \"685523da-d834-463a-8a1a-ef70069c19c5\") " pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:46:38.708424 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:38.707967 2572 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 12:46:38.708424 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:38.708036 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs podName:685523da-d834-463a-8a1a-ef70069c19c5 nodeName:}" failed. No retries permitted until 2026-02-17 12:46:42.708016393 +0000 UTC m=+9.058427201 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs") pod "network-metrics-daemon-vmpwv" (UID: "685523da-d834-463a-8a1a-ef70069c19c5") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 12:46:38.808516 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:38.808459 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vdsf2\" (UniqueName: \"kubernetes.io/projected/f1e10a0f-13dd-44a0-9d79-e6375e12ac24-kube-api-access-vdsf2\") pod \"network-check-target-f2tlm\" (UID: \"f1e10a0f-13dd-44a0-9d79-e6375e12ac24\") " pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:46:38.808750 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:38.808667 2572 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 12:46:38.808750 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:38.808688 2572 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 12:46:38.808750 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:38.808702 2572 projected.go:194] Error preparing data for projected volume kube-api-access-vdsf2 for pod openshift-network-diagnostics/network-check-target-f2tlm: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 12:46:38.808915 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:38.808761 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f1e10a0f-13dd-44a0-9d79-e6375e12ac24-kube-api-access-vdsf2 podName:f1e10a0f-13dd-44a0-9d79-e6375e12ac24 nodeName:}" failed. No retries permitted until 2026-02-17 12:46:42.808743106 +0000 UTC m=+9.159153910 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-vdsf2" (UniqueName: "kubernetes.io/projected/f1e10a0f-13dd-44a0-9d79-e6375e12ac24-kube-api-access-vdsf2") pod "network-check-target-f2tlm" (UID: "f1e10a0f-13dd-44a0-9d79-e6375e12ac24") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 12:46:39.127064 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:39.126540 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:46:39.127064 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:39.126557 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:46:39.127064 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:39.126685 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vmpwv" podUID="685523da-d834-463a-8a1a-ef70069c19c5" Feb 17 12:46:39.127064 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:39.126821 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-f2tlm" podUID="f1e10a0f-13dd-44a0-9d79-e6375e12ac24" Feb 17 12:46:41.126152 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:41.126116 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:46:41.126607 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:41.126116 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:46:41.126607 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:41.126265 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vmpwv" podUID="685523da-d834-463a-8a1a-ef70069c19c5" Feb 17 12:46:41.126607 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:41.126352 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-f2tlm" podUID="f1e10a0f-13dd-44a0-9d79-e6375e12ac24" Feb 17 12:46:42.735087 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:42.734999 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs\") pod \"network-metrics-daemon-vmpwv\" (UID: \"685523da-d834-463a-8a1a-ef70069c19c5\") " pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:46:42.735468 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:42.735190 2572 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 12:46:42.735468 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:42.735277 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs podName:685523da-d834-463a-8a1a-ef70069c19c5 nodeName:}" failed. No retries permitted until 2026-02-17 12:46:50.73525648 +0000 UTC m=+17.085667290 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs") pod "network-metrics-daemon-vmpwv" (UID: "685523da-d834-463a-8a1a-ef70069c19c5") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 12:46:42.835718 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:42.835678 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vdsf2\" (UniqueName: \"kubernetes.io/projected/f1e10a0f-13dd-44a0-9d79-e6375e12ac24-kube-api-access-vdsf2\") pod \"network-check-target-f2tlm\" (UID: \"f1e10a0f-13dd-44a0-9d79-e6375e12ac24\") " pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:46:42.835893 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:42.835876 2572 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 12:46:42.835893 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:42.835894 2572 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 12:46:42.836026 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:42.835906 2572 projected.go:194] Error preparing data for projected volume kube-api-access-vdsf2 for pod openshift-network-diagnostics/network-check-target-f2tlm: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 12:46:42.836026 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:42.835963 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f1e10a0f-13dd-44a0-9d79-e6375e12ac24-kube-api-access-vdsf2 podName:f1e10a0f-13dd-44a0-9d79-e6375e12ac24 nodeName:}" failed. No retries permitted until 2026-02-17 12:46:50.835944184 +0000 UTC m=+17.186354994 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-vdsf2" (UniqueName: "kubernetes.io/projected/f1e10a0f-13dd-44a0-9d79-e6375e12ac24-kube-api-access-vdsf2") pod "network-check-target-f2tlm" (UID: "f1e10a0f-13dd-44a0-9d79-e6375e12ac24") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 12:46:43.126384 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:43.125821 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:46:43.126384 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:43.125828 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:46:43.126384 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:43.125968 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vmpwv" podUID="685523da-d834-463a-8a1a-ef70069c19c5" Feb 17 12:46:43.126384 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:43.126056 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-f2tlm" podUID="f1e10a0f-13dd-44a0-9d79-e6375e12ac24" Feb 17 12:46:45.125890 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:45.125771 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:46:45.126338 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:45.125919 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vmpwv" podUID="685523da-d834-463a-8a1a-ef70069c19c5" Feb 17 12:46:45.126338 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:45.125772 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:46:45.126338 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:45.126152 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-f2tlm" podUID="f1e10a0f-13dd-44a0-9d79-e6375e12ac24" Feb 17 12:46:47.125827 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:47.125784 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:46:47.126318 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:47.125781 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:46:47.126318 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:47.125933 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vmpwv" podUID="685523da-d834-463a-8a1a-ef70069c19c5" Feb 17 12:46:47.126318 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:47.126018 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-f2tlm" podUID="f1e10a0f-13dd-44a0-9d79-e6375e12ac24" Feb 17 12:46:49.126213 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:49.126184 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:46:49.126747 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:49.126185 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:46:49.126747 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:49.126312 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vmpwv" podUID="685523da-d834-463a-8a1a-ef70069c19c5" Feb 17 12:46:49.126747 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:49.126378 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-f2tlm" podUID="f1e10a0f-13dd-44a0-9d79-e6375e12ac24" Feb 17 12:46:50.796175 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:50.796136 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs\") pod \"network-metrics-daemon-vmpwv\" (UID: \"685523da-d834-463a-8a1a-ef70069c19c5\") " pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:46:50.796620 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:50.796262 2572 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 12:46:50.796620 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:50.796317 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs podName:685523da-d834-463a-8a1a-ef70069c19c5 nodeName:}" failed. No retries permitted until 2026-02-17 12:47:06.796301329 +0000 UTC m=+33.146712119 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs") pod "network-metrics-daemon-vmpwv" (UID: "685523da-d834-463a-8a1a-ef70069c19c5") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 12:46:50.897290 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:50.897255 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vdsf2\" (UniqueName: \"kubernetes.io/projected/f1e10a0f-13dd-44a0-9d79-e6375e12ac24-kube-api-access-vdsf2\") pod \"network-check-target-f2tlm\" (UID: \"f1e10a0f-13dd-44a0-9d79-e6375e12ac24\") " pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:46:50.897456 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:50.897435 2572 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 12:46:50.897508 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:50.897460 2572 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 12:46:50.897508 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:50.897473 2572 projected.go:194] Error preparing data for projected volume kube-api-access-vdsf2 for pod openshift-network-diagnostics/network-check-target-f2tlm: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 12:46:50.897568 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:50.897541 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f1e10a0f-13dd-44a0-9d79-e6375e12ac24-kube-api-access-vdsf2 podName:f1e10a0f-13dd-44a0-9d79-e6375e12ac24 nodeName:}" failed. No retries permitted until 2026-02-17 12:47:06.897523095 +0000 UTC m=+33.247933907 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-vdsf2" (UniqueName: "kubernetes.io/projected/f1e10a0f-13dd-44a0-9d79-e6375e12ac24-kube-api-access-vdsf2") pod "network-check-target-f2tlm" (UID: "f1e10a0f-13dd-44a0-9d79-e6375e12ac24") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 12:46:51.126161 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:51.126071 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:46:51.126327 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:51.126071 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:46:51.126327 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:51.126210 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-f2tlm" podUID="f1e10a0f-13dd-44a0-9d79-e6375e12ac24" Feb 17 12:46:51.126327 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:51.126313 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vmpwv" podUID="685523da-d834-463a-8a1a-ef70069c19c5" Feb 17 12:46:53.126377 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:53.126334 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:46:53.126826 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:53.126340 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:46:53.126826 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:53.126473 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-f2tlm" podUID="f1e10a0f-13dd-44a0-9d79-e6375e12ac24" Feb 17 12:46:53.126826 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:53.126535 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vmpwv" podUID="685523da-d834-463a-8a1a-ef70069c19c5" Feb 17 12:46:54.245767 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:54.245647 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lb8m9_01eb7d37-a76a-4933-87df-44cf8a8d3152/ovn-acl-logging/0.log" Feb 17 12:46:54.247089 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:54.247060 2572 generic.go:358] "Generic (PLEG): container finished" podID="01eb7d37-a76a-4933-87df-44cf8a8d3152" containerID="ac601fe3e98630376690a0c4d83331882461f183daa5bcb5669fcac5a0435563" exitCode=1 Feb 17 12:46:54.247218 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:54.247149 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" event={"ID":"01eb7d37-a76a-4933-87df-44cf8a8d3152","Type":"ContainerStarted","Data":"71fa04a3bbc2b829475505876417d91a035770080c26c9e23e1c9ad7d4eb3ca0"} Feb 17 12:46:54.247218 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:54.247186 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" event={"ID":"01eb7d37-a76a-4933-87df-44cf8a8d3152","Type":"ContainerStarted","Data":"a0e317050eb4d956d6645fd761640513e6697f43cae812b42cd2a8a77fd9ad09"} Feb 17 12:46:54.247218 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:54.247202 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" event={"ID":"01eb7d37-a76a-4933-87df-44cf8a8d3152","Type":"ContainerDied","Data":"ac601fe3e98630376690a0c4d83331882461f183daa5bcb5669fcac5a0435563"} Feb 17 12:46:54.247218 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:54.247217 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" event={"ID":"01eb7d37-a76a-4933-87df-44cf8a8d3152","Type":"ContainerStarted","Data":"5c4257e48651dd2479add7f74c76242f6064d2cc283f85f3242797c64ce3c73d"} Feb 17 12:46:54.249204 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:54.249177 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/kube-apiserver-proxy-ip-10-0-131-118.ec2.internal" event={"ID":"cde915e910705d98ba55e7806902bce3","Type":"ContainerStarted","Data":"d17f5d4f412066746d2bc6e0edd4c5a5e80c32785f358a5aab06661454b955c2"} Feb 17 12:46:54.251780 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:54.251684 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-tsb7s" event={"ID":"30e805d5-0857-42ef-a870-6492d444e008","Type":"ContainerStarted","Data":"b655c1695d418f307e176f723eb96550e1c21bfe3966c3522a351d749e9adf78"} Feb 17 12:46:54.254848 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:54.254819 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-48j6f" event={"ID":"e35ae435-3708-4b40-93e1-40c8b34c3fa4","Type":"ContainerStarted","Data":"51a9d5064bd58c4152aa32c1c4ef37e5ded38fea283768c88de4d645764d40e1"} Feb 17 12:46:54.264805 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:54.264753 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-proxy-ip-10-0-131-118.ec2.internal" podStartSLOduration=20.264713823 podStartE2EDuration="20.264713823s" podCreationTimestamp="2026-02-17 12:46:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 12:46:54.264373418 +0000 UTC m=+20.614784232" watchObservedRunningTime="2026-02-17 12:46:54.264713823 +0000 UTC m=+20.615124643" Feb 17 12:46:54.279238 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:54.279097 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-node-tuning-operator/tuned-48j6f" podStartSLOduration=1.9626744409999999 podStartE2EDuration="20.27907827s" podCreationTimestamp="2026-02-17 12:46:34 +0000 UTC" firstStartedPulling="2026-02-17 12:46:35.409674714 +0000 UTC m=+1.760085503" lastFinishedPulling="2026-02-17 12:46:53.726078531 +0000 UTC m=+20.076489332" observedRunningTime="2026-02-17 12:46:54.278786924 +0000 UTC m=+20.629197736" watchObservedRunningTime="2026-02-17 12:46:54.27907827 +0000 UTC m=+20.629489082" Feb 17 12:46:54.310884 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:54.310466 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-tsb7s" podStartSLOduration=1.737277478 podStartE2EDuration="20.310446533s" podCreationTimestamp="2026-02-17 12:46:34 +0000 UTC" firstStartedPulling="2026-02-17 12:46:35.419176975 +0000 UTC m=+1.769587764" lastFinishedPulling="2026-02-17 12:46:53.992346029 +0000 UTC m=+20.342756819" observedRunningTime="2026-02-17 12:46:54.310049356 +0000 UTC m=+20.660460169" watchObservedRunningTime="2026-02-17 12:46:54.310446533 +0000 UTC m=+20.660857346" Feb 17 12:46:55.126566 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:55.126329 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:46:55.126743 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:55.126342 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:46:55.126743 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:55.126629 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-f2tlm" podUID="f1e10a0f-13dd-44a0-9d79-e6375e12ac24" Feb 17 12:46:55.126743 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:55.126675 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vmpwv" podUID="685523da-d834-463a-8a1a-ef70069c19c5" Feb 17 12:46:55.260824 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:55.260792 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lb8m9_01eb7d37-a76a-4933-87df-44cf8a8d3152/ovn-acl-logging/0.log" Feb 17 12:46:55.261252 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:55.261220 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" event={"ID":"01eb7d37-a76a-4933-87df-44cf8a8d3152","Type":"ContainerStarted","Data":"6524adf32895001949323867d712d182f2ffb0beb7ad512bddcdd6dadaab19b9"} Feb 17 12:46:55.261381 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:55.261260 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" event={"ID":"01eb7d37-a76a-4933-87df-44cf8a8d3152","Type":"ContainerStarted","Data":"55ef80a60c3f8ccfce5bc479fdc99834e9743fc63686b55edde0997f0062027d"} Feb 17 12:46:55.263194 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:55.263145 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-qjw6p" event={"ID":"8738b60c-920d-4bc5-933f-01d0fb8d68d3","Type":"ContainerStarted","Data":"a81418965f22c96a4959ef1826f06ff6afdba6531beec33d3977e2f3263cf358"} Feb 17 12:46:55.264517 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:55.264494 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-jqfj5" event={"ID":"22f3475b-a5d2-4d66-818f-5010e57aac8e","Type":"ContainerStarted","Data":"60f762525cbd44fa9f5c4467f34996a541598f2cff16df5448736dda9f3f9012"} Feb 17 12:46:55.266063 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:55.266036 2572 generic.go:358] "Generic (PLEG): container finished" podID="9751ddaa-e412-4e4e-b97a-f4e5a7a050a9" containerID="e37e6ce4c336ba5536a12b365ef261e63d57dd041f5e0414cd74222ec6ebb987" exitCode=0 Feb 17 12:46:55.266182 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:55.266139 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wfszv" event={"ID":"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9","Type":"ContainerDied","Data":"e37e6ce4c336ba5536a12b365ef261e63d57dd041f5e0414cd74222ec6ebb987"} Feb 17 12:46:55.267623 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:55.267433 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/konnectivity-agent-kmsrd" event={"ID":"30392cb9-7527-44af-833b-567cd508520e","Type":"ContainerStarted","Data":"1d255ba9a35bdcf8598f2a3e23da615d6664a458ef2554e83421c7473d2f1904"} Feb 17 12:46:55.272040 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:55.272019 2572 generic.go:358] "Generic (PLEG): container finished" podID="9b1938025d97d9e77a432753c9d25dde" containerID="b3faf11d5d40579cb95095eaf4f3782c73adc33e67a8db09fa63537848227304" exitCode=0 Feb 17 12:46:55.272139 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:55.272087 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-131-118.ec2.internal" event={"ID":"9b1938025d97d9e77a432753c9d25dde","Type":"ContainerDied","Data":"b3faf11d5d40579cb95095eaf4f3782c73adc33e67a8db09fa63537848227304"} Feb 17 12:46:55.273530 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:55.273507 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" event={"ID":"3d7ab4ec-e6ac-454b-8724-36a22d414dae","Type":"ContainerStarted","Data":"f505345c488901ca1abda184b8abb091d2273a35bfa43b30b2c908f61c910cb9"} Feb 17 12:46:55.277353 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:55.277318 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-qjw6p" podStartSLOduration=2.947880737 podStartE2EDuration="21.277308104s" podCreationTimestamp="2026-02-17 12:46:34 +0000 UTC" firstStartedPulling="2026-02-17 12:46:35.38002411 +0000 UTC m=+1.730434903" lastFinishedPulling="2026-02-17 12:46:53.709451476 +0000 UTC m=+20.059862270" observedRunningTime="2026-02-17 12:46:55.276716655 +0000 UTC m=+21.627127464" watchObservedRunningTime="2026-02-17 12:46:55.277308104 +0000 UTC m=+21.627718916" Feb 17 12:46:55.288857 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:55.288810 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-jqfj5" podStartSLOduration=2.951738546 podStartE2EDuration="21.28879458s" podCreationTimestamp="2026-02-17 12:46:34 +0000 UTC" firstStartedPulling="2026-02-17 12:46:35.372379374 +0000 UTC m=+1.722790180" lastFinishedPulling="2026-02-17 12:46:53.70943541 +0000 UTC m=+20.059846214" observedRunningTime="2026-02-17 12:46:55.288482658 +0000 UTC m=+21.638893469" watchObservedRunningTime="2026-02-17 12:46:55.28879458 +0000 UTC m=+21.639205375" Feb 17 12:46:55.301681 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:55.301641 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-operator/iptables-alerter-fm5zs" podStartSLOduration=2.960553327 podStartE2EDuration="21.301630521s" podCreationTimestamp="2026-02-17 12:46:34 +0000 UTC" firstStartedPulling="2026-02-17 12:46:35.385559193 +0000 UTC m=+1.735969984" lastFinishedPulling="2026-02-17 12:46:53.726636374 +0000 UTC m=+20.077047178" observedRunningTime="2026-02-17 12:46:55.301520766 +0000 UTC m=+21.651931579" watchObservedRunningTime="2026-02-17 12:46:55.301630521 +0000 UTC m=+21.652041332" Feb 17 12:46:55.349187 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:55.349146 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/konnectivity-agent-kmsrd" podStartSLOduration=3.036423195 podStartE2EDuration="21.349132885s" podCreationTimestamp="2026-02-17 12:46:34 +0000 UTC" firstStartedPulling="2026-02-17 12:46:35.396729144 +0000 UTC m=+1.747139937" lastFinishedPulling="2026-02-17 12:46:53.709438819 +0000 UTC m=+20.059849627" observedRunningTime="2026-02-17 12:46:55.330546849 +0000 UTC m=+21.680957662" watchObservedRunningTime="2026-02-17 12:46:55.349132885 +0000 UTC m=+21.699543687" Feb 17 12:46:55.516748 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:55.516724 2572 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/ebs.csi.aws.com-reg.sock" Feb 17 12:46:56.137248 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:56.137147 2572 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/ebs.csi.aws.com-reg.sock","Timestamp":"2026-02-17T12:46:55.516742069Z","UUID":"5266dd21-13df-40de-9ec0-3dcddb3105ac","Handler":null,"Name":"","Endpoint":""} Feb 17 12:46:56.138990 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:56.138960 2572 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: ebs.csi.aws.com endpoint: /var/lib/kubelet/plugins/ebs.csi.aws.com/csi.sock versions: 1.0.0 Feb 17 12:46:56.138990 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:56.138991 2572 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: ebs.csi.aws.com at endpoint: /var/lib/kubelet/plugins/ebs.csi.aws.com/csi.sock Feb 17 12:46:56.276985 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:56.276916 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-fm5zs" event={"ID":"ba1edd82-a8f5-4e73-ac84-987cbbec0939","Type":"ContainerStarted","Data":"e2610b854171a266c78305bbc022888745541f2a3f5b18a8edd7b65539d78245"} Feb 17 12:46:56.278985 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:56.278958 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-131-118.ec2.internal" event={"ID":"9b1938025d97d9e77a432753c9d25dde","Type":"ContainerStarted","Data":"4e798b0d61f579172ba75d8172c1c91d43b7da4213ed59f3b4e14ea8e799fa89"} Feb 17 12:46:56.284229 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:56.284181 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" event={"ID":"3d7ab4ec-e6ac-454b-8724-36a22d414dae","Type":"ContainerStarted","Data":"e6716760a1af2bd9f83048b78bcde3fe49eda44caab131da02c1850cbbecb316"} Feb 17 12:46:56.292278 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:56.292233 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-ip-10-0-131-118.ec2.internal" podStartSLOduration=22.292221459 podStartE2EDuration="22.292221459s" podCreationTimestamp="2026-02-17 12:46:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 12:46:56.29221688 +0000 UTC m=+22.642627715" watchObservedRunningTime="2026-02-17 12:46:56.292221459 +0000 UTC m=+22.642632272" Feb 17 12:46:56.517648 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:56.517582 2572 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="kube-system/konnectivity-agent-kmsrd" Feb 17 12:46:56.518500 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:56.518480 2572 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="kube-system/konnectivity-agent-kmsrd" Feb 17 12:46:57.125864 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:57.125825 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:46:57.125864 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:57.125840 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:46:57.126096 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:57.125946 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vmpwv" podUID="685523da-d834-463a-8a1a-ef70069c19c5" Feb 17 12:46:57.126096 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:57.126081 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-f2tlm" podUID="f1e10a0f-13dd-44a0-9d79-e6375e12ac24" Feb 17 12:46:57.288125 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:57.288073 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" event={"ID":"3d7ab4ec-e6ac-454b-8724-36a22d414dae","Type":"ContainerStarted","Data":"5d134e2f06331c419c19e98adce179dbee62a2f9bf6f2ec5f28539ccd64da70e"} Feb 17 12:46:57.290854 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:57.290833 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lb8m9_01eb7d37-a76a-4933-87df-44cf8a8d3152/ovn-acl-logging/0.log" Feb 17 12:46:57.291251 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:57.291225 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" event={"ID":"01eb7d37-a76a-4933-87df-44cf8a8d3152","Type":"ContainerStarted","Data":"0e6c658913d854b20c2f46458f08c6fa5129a8efb96b65db49454ef5632fe8aa"} Feb 17 12:46:57.291509 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:57.291489 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="kube-system/konnectivity-agent-kmsrd" Feb 17 12:46:57.291999 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:57.291980 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="kube-system/konnectivity-agent-kmsrd" Feb 17 12:46:57.303950 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:57.303908 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-csi-drivers/aws-ebs-csi-driver-node-ghtvv" podStartSLOduration=2.31481324 podStartE2EDuration="23.303896315s" podCreationTimestamp="2026-02-17 12:46:34 +0000 UTC" firstStartedPulling="2026-02-17 12:46:35.403590044 +0000 UTC m=+1.754000834" lastFinishedPulling="2026-02-17 12:46:56.392673105 +0000 UTC m=+22.743083909" observedRunningTime="2026-02-17 12:46:57.303580226 +0000 UTC m=+23.653991040" watchObservedRunningTime="2026-02-17 12:46:57.303896315 +0000 UTC m=+23.654307126" Feb 17 12:46:59.126159 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:59.126121 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:46:59.126676 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:46:59.126121 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:46:59.126676 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:59.126245 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-f2tlm" podUID="f1e10a0f-13dd-44a0-9d79-e6375e12ac24" Feb 17 12:46:59.126676 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:46:59.126333 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vmpwv" podUID="685523da-d834-463a-8a1a-ef70069c19c5" Feb 17 12:47:00.300459 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:00.300280 2572 generic.go:358] "Generic (PLEG): container finished" podID="9751ddaa-e412-4e4e-b97a-f4e5a7a050a9" containerID="08a04f7360e4ffb60111807d46a060d5948e52bce05f26d1124163cf927bcf84" exitCode=0 Feb 17 12:47:00.301228 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:00.300371 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wfszv" event={"ID":"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9","Type":"ContainerDied","Data":"08a04f7360e4ffb60111807d46a060d5948e52bce05f26d1124163cf927bcf84"} Feb 17 12:47:00.303438 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:00.303422 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lb8m9_01eb7d37-a76a-4933-87df-44cf8a8d3152/ovn-acl-logging/0.log" Feb 17 12:47:00.303745 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:00.303712 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" event={"ID":"01eb7d37-a76a-4933-87df-44cf8a8d3152","Type":"ContainerStarted","Data":"e1bdc37dee8b0eb655c67f2eb1df8b5f79b02826f1bcc26580dd3b8c3675ef8d"} Feb 17 12:47:00.303964 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:00.303948 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:47:00.304050 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:00.304039 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:47:00.304203 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:00.304186 2572 scope.go:117] "RemoveContainer" containerID="ac601fe3e98630376690a0c4d83331882461f183daa5bcb5669fcac5a0435563" Feb 17 12:47:00.320098 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:00.319997 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:47:01.126002 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:01.125971 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:47:01.126195 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:01.125971 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:47:01.126195 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:01.126094 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vmpwv" podUID="685523da-d834-463a-8a1a-ef70069c19c5" Feb 17 12:47:01.126195 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:01.126181 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-f2tlm" podUID="f1e10a0f-13dd-44a0-9d79-e6375e12ac24" Feb 17 12:47:01.244265 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:01.244220 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-target-f2tlm"] Feb 17 12:47:01.246983 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:01.246852 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-vmpwv"] Feb 17 12:47:01.308169 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:01.308004 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lb8m9_01eb7d37-a76a-4933-87df-44cf8a8d3152/ovn-acl-logging/0.log" Feb 17 12:47:01.308486 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:01.308420 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" event={"ID":"01eb7d37-a76a-4933-87df-44cf8a8d3152","Type":"ContainerStarted","Data":"90502d106393a97c2fd553dc1a00a02ed12066d256789ecd01c306c55f9ce6ae"} Feb 17 12:47:01.308694 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:01.308678 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:47:01.310177 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:01.310152 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wfszv" event={"ID":"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9","Type":"ContainerStarted","Data":"ecc6eb727bc1af42435392ecf931a79591bce249c3c023d9e43aad6bf8792b34"} Feb 17 12:47:01.310279 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:01.310191 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:47:01.310279 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:01.310210 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:47:01.310428 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:01.310410 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vmpwv" podUID="685523da-d834-463a-8a1a-ef70069c19c5" Feb 17 12:47:01.310540 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:01.310521 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-f2tlm" podUID="f1e10a0f-13dd-44a0-9d79-e6375e12ac24" Feb 17 12:47:01.322917 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:01.322894 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:47:01.329883 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:01.329847 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" podStartSLOduration=8.965566875 podStartE2EDuration="27.329835239s" podCreationTimestamp="2026-02-17 12:46:34 +0000 UTC" firstStartedPulling="2026-02-17 12:46:35.390615458 +0000 UTC m=+1.741026249" lastFinishedPulling="2026-02-17 12:46:53.754883821 +0000 UTC m=+20.105294613" observedRunningTime="2026-02-17 12:47:01.328987529 +0000 UTC m=+27.679398340" watchObservedRunningTime="2026-02-17 12:47:01.329835239 +0000 UTC m=+27.680246051" Feb 17 12:47:02.313361 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:02.313331 2572 generic.go:358] "Generic (PLEG): container finished" podID="9751ddaa-e412-4e4e-b97a-f4e5a7a050a9" containerID="ecc6eb727bc1af42435392ecf931a79591bce249c3c023d9e43aad6bf8792b34" exitCode=0 Feb 17 12:47:02.313731 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:02.313419 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wfszv" event={"ID":"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9","Type":"ContainerDied","Data":"ecc6eb727bc1af42435392ecf931a79591bce249c3c023d9e43aad6bf8792b34"} Feb 17 12:47:03.126447 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:03.126413 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:47:03.126609 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:03.126426 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:47:03.126609 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:03.126536 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-f2tlm" podUID="f1e10a0f-13dd-44a0-9d79-e6375e12ac24" Feb 17 12:47:03.126609 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:03.126592 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vmpwv" podUID="685523da-d834-463a-8a1a-ef70069c19c5" Feb 17 12:47:03.317402 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:03.317369 2572 generic.go:358] "Generic (PLEG): container finished" podID="9751ddaa-e412-4e4e-b97a-f4e5a7a050a9" containerID="8da7f8881ac8d8e149b44070df099cff80dc65eb737b75219eeb4e7582ca60cc" exitCode=0 Feb 17 12:47:03.318183 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:03.317444 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wfszv" event={"ID":"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9","Type":"ContainerDied","Data":"8da7f8881ac8d8e149b44070df099cff80dc65eb737b75219eeb4e7582ca60cc"} Feb 17 12:47:05.126077 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:05.125998 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:47:05.126517 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:05.126001 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:47:05.126517 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:05.126141 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-f2tlm" podUID="f1e10a0f-13dd-44a0-9d79-e6375e12ac24" Feb 17 12:47:05.126517 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:05.126221 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vmpwv" podUID="685523da-d834-463a-8a1a-ef70069c19c5" Feb 17 12:47:06.817241 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:06.817196 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs\") pod \"network-metrics-daemon-vmpwv\" (UID: \"685523da-d834-463a-8a1a-ef70069c19c5\") " pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:47:06.817747 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:06.817354 2572 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 12:47:06.817747 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:06.817437 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs podName:685523da-d834-463a-8a1a-ef70069c19c5 nodeName:}" failed. No retries permitted until 2026-02-17 12:47:38.817409636 +0000 UTC m=+65.167820433 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs") pod "network-metrics-daemon-vmpwv" (UID: "685523da-d834-463a-8a1a-ef70069c19c5") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 17 12:47:06.917752 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:06.917667 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vdsf2\" (UniqueName: \"kubernetes.io/projected/f1e10a0f-13dd-44a0-9d79-e6375e12ac24-kube-api-access-vdsf2\") pod \"network-check-target-f2tlm\" (UID: \"f1e10a0f-13dd-44a0-9d79-e6375e12ac24\") " pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:47:06.917895 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:06.917855 2572 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 17 12:47:06.917895 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:06.917885 2572 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 17 12:47:06.917974 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:06.917898 2572 projected.go:194] Error preparing data for projected volume kube-api-access-vdsf2 for pod openshift-network-diagnostics/network-check-target-f2tlm: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 12:47:06.917974 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:06.917959 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f1e10a0f-13dd-44a0-9d79-e6375e12ac24-kube-api-access-vdsf2 podName:f1e10a0f-13dd-44a0-9d79-e6375e12ac24 nodeName:}" failed. No retries permitted until 2026-02-17 12:47:38.91794142 +0000 UTC m=+65.268352230 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-vdsf2" (UniqueName: "kubernetes.io/projected/f1e10a0f-13dd-44a0-9d79-e6375e12ac24-kube-api-access-vdsf2") pod "network-check-target-f2tlm" (UID: "f1e10a0f-13dd-44a0-9d79-e6375e12ac24") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 17 12:47:06.994474 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:06.994446 2572 kubelet_node_status.go:736] "Recording event message for node" node="ip-10-0-131-118.ec2.internal" event="NodeReady" Feb 17 12:47:06.994645 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:06.994593 2572 kubelet_node_status.go:550] "Fast updating node status as it just became ready" Feb 17 12:47:07.048172 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.048137 2572 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-b6hth"] Feb 17 12:47:07.079467 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.079435 2572 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-prcdr"] Feb 17 12:47:07.079762 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.079736 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-b6hth" Feb 17 12:47:07.081692 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.081606 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-default-metrics-tls\"" Feb 17 12:47:07.081692 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.081606 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-dockercfg-944wh\"" Feb 17 12:47:07.081863 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.081695 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"dns-default\"" Feb 17 12:47:07.095448 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.095422 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-b6hth"] Feb 17 12:47:07.095584 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.095453 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-prcdr"] Feb 17 12:47:07.095647 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.095592 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-prcdr" Feb 17 12:47:07.097500 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.097400 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"openshift-service-ca.crt\"" Feb 17 12:47:07.097500 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.097413 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"kube-root-ca.crt\"" Feb 17 12:47:07.097500 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.097437 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"default-dockercfg-9qj8f\"" Feb 17 12:47:07.097500 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.097406 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"canary-serving-cert\"" Feb 17 12:47:07.126358 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.126331 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:47:07.126495 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.126331 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:47:07.128213 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.128187 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Feb 17 12:47:07.128337 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.128191 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-b7dtp\"" Feb 17 12:47:07.128337 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.128283 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Feb 17 12:47:07.128337 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.128323 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Feb 17 12:47:07.128510 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.128191 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-diagnostics\"/\"default-dockercfg-cgwh6\"" Feb 17 12:47:07.220534 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.220496 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdn2q\" (UniqueName: \"kubernetes.io/projected/51d19194-106c-4b3f-ba37-a0a7eedbc75a-kube-api-access-sdn2q\") pod \"ingress-canary-prcdr\" (UID: \"51d19194-106c-4b3f-ba37-a0a7eedbc75a\") " pod="openshift-ingress-canary/ingress-canary-prcdr" Feb 17 12:47:07.220736 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.220544 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-metrics-tls\") pod \"dns-default-b6hth\" (UID: \"a01f1f0a-fa4d-4b3c-a85e-053603fecfc1\") " pod="openshift-dns/dns-default-b6hth" Feb 17 12:47:07.220736 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.220595 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/51d19194-106c-4b3f-ba37-a0a7eedbc75a-cert\") pod \"ingress-canary-prcdr\" (UID: \"51d19194-106c-4b3f-ba37-a0a7eedbc75a\") " pod="openshift-ingress-canary/ingress-canary-prcdr" Feb 17 12:47:07.220736 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.220659 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-tmp-dir\") pod \"dns-default-b6hth\" (UID: \"a01f1f0a-fa4d-4b3c-a85e-053603fecfc1\") " pod="openshift-dns/dns-default-b6hth" Feb 17 12:47:07.220736 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.220676 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs5rg\" (UniqueName: \"kubernetes.io/projected/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-kube-api-access-fs5rg\") pod \"dns-default-b6hth\" (UID: \"a01f1f0a-fa4d-4b3c-a85e-053603fecfc1\") " pod="openshift-dns/dns-default-b6hth" Feb 17 12:47:07.220736 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.220713 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-config-volume\") pod \"dns-default-b6hth\" (UID: \"a01f1f0a-fa4d-4b3c-a85e-053603fecfc1\") " pod="openshift-dns/dns-default-b6hth" Feb 17 12:47:07.321900 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.321853 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-config-volume\") pod \"dns-default-b6hth\" (UID: \"a01f1f0a-fa4d-4b3c-a85e-053603fecfc1\") " pod="openshift-dns/dns-default-b6hth" Feb 17 12:47:07.322086 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.321909 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-sdn2q\" (UniqueName: \"kubernetes.io/projected/51d19194-106c-4b3f-ba37-a0a7eedbc75a-kube-api-access-sdn2q\") pod \"ingress-canary-prcdr\" (UID: \"51d19194-106c-4b3f-ba37-a0a7eedbc75a\") " pod="openshift-ingress-canary/ingress-canary-prcdr" Feb 17 12:47:07.322086 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.321940 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-metrics-tls\") pod \"dns-default-b6hth\" (UID: \"a01f1f0a-fa4d-4b3c-a85e-053603fecfc1\") " pod="openshift-dns/dns-default-b6hth" Feb 17 12:47:07.322086 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.321981 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/51d19194-106c-4b3f-ba37-a0a7eedbc75a-cert\") pod \"ingress-canary-prcdr\" (UID: \"51d19194-106c-4b3f-ba37-a0a7eedbc75a\") " pod="openshift-ingress-canary/ingress-canary-prcdr" Feb 17 12:47:07.322086 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.322031 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-tmp-dir\") pod \"dns-default-b6hth\" (UID: \"a01f1f0a-fa4d-4b3c-a85e-053603fecfc1\") " pod="openshift-dns/dns-default-b6hth" Feb 17 12:47:07.322086 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.322053 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-fs5rg\" (UniqueName: \"kubernetes.io/projected/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-kube-api-access-fs5rg\") pod \"dns-default-b6hth\" (UID: \"a01f1f0a-fa4d-4b3c-a85e-053603fecfc1\") " pod="openshift-dns/dns-default-b6hth" Feb 17 12:47:07.322360 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:07.322092 2572 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Feb 17 12:47:07.322360 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:07.322160 2572 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Feb 17 12:47:07.322360 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:07.322199 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-metrics-tls podName:a01f1f0a-fa4d-4b3c-a85e-053603fecfc1 nodeName:}" failed. No retries permitted until 2026-02-17 12:47:07.82217518 +0000 UTC m=+34.172585992 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-metrics-tls") pod "dns-default-b6hth" (UID: "a01f1f0a-fa4d-4b3c-a85e-053603fecfc1") : secret "dns-default-metrics-tls" not found Feb 17 12:47:07.322360 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:07.322229 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/51d19194-106c-4b3f-ba37-a0a7eedbc75a-cert podName:51d19194-106c-4b3f-ba37-a0a7eedbc75a nodeName:}" failed. No retries permitted until 2026-02-17 12:47:07.822210641 +0000 UTC m=+34.172621437 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/51d19194-106c-4b3f-ba37-a0a7eedbc75a-cert") pod "ingress-canary-prcdr" (UID: "51d19194-106c-4b3f-ba37-a0a7eedbc75a") : secret "canary-serving-cert" not found Feb 17 12:47:07.322519 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.322491 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-tmp-dir\") pod \"dns-default-b6hth\" (UID: \"a01f1f0a-fa4d-4b3c-a85e-053603fecfc1\") " pod="openshift-dns/dns-default-b6hth" Feb 17 12:47:07.322569 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.322549 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-config-volume\") pod \"dns-default-b6hth\" (UID: \"a01f1f0a-fa4d-4b3c-a85e-053603fecfc1\") " pod="openshift-dns/dns-default-b6hth" Feb 17 12:47:07.334414 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.334388 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs5rg\" (UniqueName: \"kubernetes.io/projected/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-kube-api-access-fs5rg\") pod \"dns-default-b6hth\" (UID: \"a01f1f0a-fa4d-4b3c-a85e-053603fecfc1\") " pod="openshift-dns/dns-default-b6hth" Feb 17 12:47:07.334572 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.334546 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdn2q\" (UniqueName: \"kubernetes.io/projected/51d19194-106c-4b3f-ba37-a0a7eedbc75a-kube-api-access-sdn2q\") pod \"ingress-canary-prcdr\" (UID: \"51d19194-106c-4b3f-ba37-a0a7eedbc75a\") " pod="openshift-ingress-canary/ingress-canary-prcdr" Feb 17 12:47:07.825876 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.825659 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/51d19194-106c-4b3f-ba37-a0a7eedbc75a-cert\") pod \"ingress-canary-prcdr\" (UID: \"51d19194-106c-4b3f-ba37-a0a7eedbc75a\") " pod="openshift-ingress-canary/ingress-canary-prcdr" Feb 17 12:47:07.826357 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:07.825800 2572 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Feb 17 12:47:07.826357 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:07.826008 2572 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Feb 17 12:47:07.826357 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:07.825948 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-metrics-tls\") pod \"dns-default-b6hth\" (UID: \"a01f1f0a-fa4d-4b3c-a85e-053603fecfc1\") " pod="openshift-dns/dns-default-b6hth" Feb 17 12:47:07.826357 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:07.826036 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/51d19194-106c-4b3f-ba37-a0a7eedbc75a-cert podName:51d19194-106c-4b3f-ba37-a0a7eedbc75a nodeName:}" failed. No retries permitted until 2026-02-17 12:47:08.826008737 +0000 UTC m=+35.176419553 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/51d19194-106c-4b3f-ba37-a0a7eedbc75a-cert") pod "ingress-canary-prcdr" (UID: "51d19194-106c-4b3f-ba37-a0a7eedbc75a") : secret "canary-serving-cert" not found Feb 17 12:47:07.826357 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:07.826114 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-metrics-tls podName:a01f1f0a-fa4d-4b3c-a85e-053603fecfc1 nodeName:}" failed. No retries permitted until 2026-02-17 12:47:08.826076136 +0000 UTC m=+35.176486940 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-metrics-tls") pod "dns-default-b6hth" (UID: "a01f1f0a-fa4d-4b3c-a85e-053603fecfc1") : secret "dns-default-metrics-tls" not found Feb 17 12:47:08.833684 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:08.833640 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-metrics-tls\") pod \"dns-default-b6hth\" (UID: \"a01f1f0a-fa4d-4b3c-a85e-053603fecfc1\") " pod="openshift-dns/dns-default-b6hth" Feb 17 12:47:08.834164 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:08.833694 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/51d19194-106c-4b3f-ba37-a0a7eedbc75a-cert\") pod \"ingress-canary-prcdr\" (UID: \"51d19194-106c-4b3f-ba37-a0a7eedbc75a\") " pod="openshift-ingress-canary/ingress-canary-prcdr" Feb 17 12:47:08.834164 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:08.833790 2572 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Feb 17 12:47:08.834164 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:08.833811 2572 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Feb 17 12:47:08.834164 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:08.833859 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-metrics-tls podName:a01f1f0a-fa4d-4b3c-a85e-053603fecfc1 nodeName:}" failed. No retries permitted until 2026-02-17 12:47:10.833841461 +0000 UTC m=+37.184252258 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-metrics-tls") pod "dns-default-b6hth" (UID: "a01f1f0a-fa4d-4b3c-a85e-053603fecfc1") : secret "dns-default-metrics-tls" not found Feb 17 12:47:08.834164 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:08.833873 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/51d19194-106c-4b3f-ba37-a0a7eedbc75a-cert podName:51d19194-106c-4b3f-ba37-a0a7eedbc75a nodeName:}" failed. No retries permitted until 2026-02-17 12:47:10.833867007 +0000 UTC m=+37.184277797 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/51d19194-106c-4b3f-ba37-a0a7eedbc75a-cert") pod "ingress-canary-prcdr" (UID: "51d19194-106c-4b3f-ba37-a0a7eedbc75a") : secret "canary-serving-cert" not found Feb 17 12:47:09.332194 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:09.332159 2572 generic.go:358] "Generic (PLEG): container finished" podID="9751ddaa-e412-4e4e-b97a-f4e5a7a050a9" containerID="9f6c8ebfd89e4994e28e3644636a9662aee3928294a01f842bb065055e0c09e0" exitCode=0 Feb 17 12:47:09.332349 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:09.332219 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wfszv" event={"ID":"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9","Type":"ContainerDied","Data":"9f6c8ebfd89e4994e28e3644636a9662aee3928294a01f842bb065055e0c09e0"} Feb 17 12:47:10.336910 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:10.336868 2572 generic.go:358] "Generic (PLEG): container finished" podID="9751ddaa-e412-4e4e-b97a-f4e5a7a050a9" containerID="7e9f5d2e8573662a9d1b6bc10d79354c9a5611ac6f04bc859f51e107521cf0fe" exitCode=0 Feb 17 12:47:10.337268 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:10.336934 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wfszv" event={"ID":"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9","Type":"ContainerDied","Data":"7e9f5d2e8573662a9d1b6bc10d79354c9a5611ac6f04bc859f51e107521cf0fe"} Feb 17 12:47:10.850063 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:10.850027 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-metrics-tls\") pod \"dns-default-b6hth\" (UID: \"a01f1f0a-fa4d-4b3c-a85e-053603fecfc1\") " pod="openshift-dns/dns-default-b6hth" Feb 17 12:47:10.850225 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:10.850076 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/51d19194-106c-4b3f-ba37-a0a7eedbc75a-cert\") pod \"ingress-canary-prcdr\" (UID: \"51d19194-106c-4b3f-ba37-a0a7eedbc75a\") " pod="openshift-ingress-canary/ingress-canary-prcdr" Feb 17 12:47:10.850225 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:10.850195 2572 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Feb 17 12:47:10.850225 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:10.850222 2572 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Feb 17 12:47:10.850334 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:10.850277 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/51d19194-106c-4b3f-ba37-a0a7eedbc75a-cert podName:51d19194-106c-4b3f-ba37-a0a7eedbc75a nodeName:}" failed. No retries permitted until 2026-02-17 12:47:14.850261512 +0000 UTC m=+41.200672318 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/51d19194-106c-4b3f-ba37-a0a7eedbc75a-cert") pod "ingress-canary-prcdr" (UID: "51d19194-106c-4b3f-ba37-a0a7eedbc75a") : secret "canary-serving-cert" not found Feb 17 12:47:10.850334 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:10.850290 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-metrics-tls podName:a01f1f0a-fa4d-4b3c-a85e-053603fecfc1 nodeName:}" failed. No retries permitted until 2026-02-17 12:47:14.850285115 +0000 UTC m=+41.200695905 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-metrics-tls") pod "dns-default-b6hth" (UID: "a01f1f0a-fa4d-4b3c-a85e-053603fecfc1") : secret "dns-default-metrics-tls" not found Feb 17 12:47:11.341128 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:11.341083 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wfszv" event={"ID":"9751ddaa-e412-4e4e-b97a-f4e5a7a050a9","Type":"ContainerStarted","Data":"61fbf9baff0176b4db56967d9ad2b79e821968d28f08edf8d7697f3d8f316a7a"} Feb 17 12:47:11.364009 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:11.363960 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-wfszv" podStartSLOduration=3.780835298 podStartE2EDuration="37.363946577s" podCreationTimestamp="2026-02-17 12:46:34 +0000 UTC" firstStartedPulling="2026-02-17 12:46:35.416623929 +0000 UTC m=+1.767034734" lastFinishedPulling="2026-02-17 12:47:08.999735223 +0000 UTC m=+35.350146013" observedRunningTime="2026-02-17 12:47:11.362810818 +0000 UTC m=+37.713221630" watchObservedRunningTime="2026-02-17 12:47:11.363946577 +0000 UTC m=+37.714357385" Feb 17 12:47:14.875749 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:14.875706 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-metrics-tls\") pod \"dns-default-b6hth\" (UID: \"a01f1f0a-fa4d-4b3c-a85e-053603fecfc1\") " pod="openshift-dns/dns-default-b6hth" Feb 17 12:47:14.875749 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:14.875755 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/51d19194-106c-4b3f-ba37-a0a7eedbc75a-cert\") pod \"ingress-canary-prcdr\" (UID: \"51d19194-106c-4b3f-ba37-a0a7eedbc75a\") " pod="openshift-ingress-canary/ingress-canary-prcdr" Feb 17 12:47:14.876226 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:14.875855 2572 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Feb 17 12:47:14.876226 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:14.875868 2572 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Feb 17 12:47:14.876226 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:14.875905 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/51d19194-106c-4b3f-ba37-a0a7eedbc75a-cert podName:51d19194-106c-4b3f-ba37-a0a7eedbc75a nodeName:}" failed. No retries permitted until 2026-02-17 12:47:22.875891737 +0000 UTC m=+49.226302527 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/51d19194-106c-4b3f-ba37-a0a7eedbc75a-cert") pod "ingress-canary-prcdr" (UID: "51d19194-106c-4b3f-ba37-a0a7eedbc75a") : secret "canary-serving-cert" not found Feb 17 12:47:14.876226 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:14.875937 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-metrics-tls podName:a01f1f0a-fa4d-4b3c-a85e-053603fecfc1 nodeName:}" failed. No retries permitted until 2026-02-17 12:47:22.875917593 +0000 UTC m=+49.226328398 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-metrics-tls") pod "dns-default-b6hth" (UID: "a01f1f0a-fa4d-4b3c-a85e-053603fecfc1") : secret "dns-default-metrics-tls" not found Feb 17 12:47:22.924280 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:22.924232 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-metrics-tls\") pod \"dns-default-b6hth\" (UID: \"a01f1f0a-fa4d-4b3c-a85e-053603fecfc1\") " pod="openshift-dns/dns-default-b6hth" Feb 17 12:47:22.924868 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:22.924295 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/51d19194-106c-4b3f-ba37-a0a7eedbc75a-cert\") pod \"ingress-canary-prcdr\" (UID: \"51d19194-106c-4b3f-ba37-a0a7eedbc75a\") " pod="openshift-ingress-canary/ingress-canary-prcdr" Feb 17 12:47:22.924868 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:22.924392 2572 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Feb 17 12:47:22.924868 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:22.924442 2572 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Feb 17 12:47:22.924868 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:22.924463 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-metrics-tls podName:a01f1f0a-fa4d-4b3c-a85e-053603fecfc1 nodeName:}" failed. No retries permitted until 2026-02-17 12:47:38.924446826 +0000 UTC m=+65.274857616 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-metrics-tls") pod "dns-default-b6hth" (UID: "a01f1f0a-fa4d-4b3c-a85e-053603fecfc1") : secret "dns-default-metrics-tls" not found Feb 17 12:47:22.924868 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:22.924504 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/51d19194-106c-4b3f-ba37-a0a7eedbc75a-cert podName:51d19194-106c-4b3f-ba37-a0a7eedbc75a nodeName:}" failed. No retries permitted until 2026-02-17 12:47:38.924486482 +0000 UTC m=+65.274897289 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/51d19194-106c-4b3f-ba37-a0a7eedbc75a-cert") pod "ingress-canary-prcdr" (UID: "51d19194-106c-4b3f-ba37-a0a7eedbc75a") : secret "canary-serving-cert" not found Feb 17 12:47:33.327600 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:33.327570 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lb8m9" Feb 17 12:47:38.835594 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:38.835552 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs\") pod \"network-metrics-daemon-vmpwv\" (UID: \"685523da-d834-463a-8a1a-ef70069c19c5\") " pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:47:38.837783 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:38.837760 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Feb 17 12:47:38.846248 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:38.846227 2572 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: secret "metrics-daemon-secret" not found Feb 17 12:47:38.846322 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:38.846299 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs podName:685523da-d834-463a-8a1a-ef70069c19c5 nodeName:}" failed. No retries permitted until 2026-02-17 12:48:42.84627552 +0000 UTC m=+129.196686328 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs") pod "network-metrics-daemon-vmpwv" (UID: "685523da-d834-463a-8a1a-ef70069c19c5") : secret "metrics-daemon-secret" not found Feb 17 12:47:38.936722 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:38.936691 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-metrics-tls\") pod \"dns-default-b6hth\" (UID: \"a01f1f0a-fa4d-4b3c-a85e-053603fecfc1\") " pod="openshift-dns/dns-default-b6hth" Feb 17 12:47:38.936722 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:38.936726 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vdsf2\" (UniqueName: \"kubernetes.io/projected/f1e10a0f-13dd-44a0-9d79-e6375e12ac24-kube-api-access-vdsf2\") pod \"network-check-target-f2tlm\" (UID: \"f1e10a0f-13dd-44a0-9d79-e6375e12ac24\") " pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:47:38.936900 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:38.936748 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/51d19194-106c-4b3f-ba37-a0a7eedbc75a-cert\") pod \"ingress-canary-prcdr\" (UID: \"51d19194-106c-4b3f-ba37-a0a7eedbc75a\") " pod="openshift-ingress-canary/ingress-canary-prcdr" Feb 17 12:47:38.936900 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:38.936842 2572 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Feb 17 12:47:38.936900 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:38.936857 2572 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Feb 17 12:47:38.936989 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:38.936909 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-metrics-tls podName:a01f1f0a-fa4d-4b3c-a85e-053603fecfc1 nodeName:}" failed. No retries permitted until 2026-02-17 12:48:10.936892602 +0000 UTC m=+97.287303392 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-metrics-tls") pod "dns-default-b6hth" (UID: "a01f1f0a-fa4d-4b3c-a85e-053603fecfc1") : secret "dns-default-metrics-tls" not found Feb 17 12:47:38.936989 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:47:38.936924 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/51d19194-106c-4b3f-ba37-a0a7eedbc75a-cert podName:51d19194-106c-4b3f-ba37-a0a7eedbc75a nodeName:}" failed. No retries permitted until 2026-02-17 12:48:10.936917671 +0000 UTC m=+97.287328462 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/51d19194-106c-4b3f-ba37-a0a7eedbc75a-cert") pod "ingress-canary-prcdr" (UID: "51d19194-106c-4b3f-ba37-a0a7eedbc75a") : secret "canary-serving-cert" not found Feb 17 12:47:38.938867 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:38.938847 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Feb 17 12:47:38.948629 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:38.948612 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Feb 17 12:47:38.961203 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:38.961178 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdsf2\" (UniqueName: \"kubernetes.io/projected/f1e10a0f-13dd-44a0-9d79-e6375e12ac24-kube-api-access-vdsf2\") pod \"network-check-target-f2tlm\" (UID: \"f1e10a0f-13dd-44a0-9d79-e6375e12ac24\") " pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:47:39.239505 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:39.239478 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-diagnostics\"/\"default-dockercfg-cgwh6\"" Feb 17 12:47:39.248021 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:39.248002 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:47:39.413925 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:39.413897 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-target-f2tlm"] Feb 17 12:47:39.417702 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:47:39.417670 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1e10a0f_13dd_44a0_9d79_e6375e12ac24.slice/crio-23b07d6db45d3b755a1b3c846148096ffd448412413d087eae6b98d1704746be WatchSource:0}: Error finding container 23b07d6db45d3b755a1b3c846148096ffd448412413d087eae6b98d1704746be: Status 404 returned error can't find the container with id 23b07d6db45d3b755a1b3c846148096ffd448412413d087eae6b98d1704746be Feb 17 12:47:40.398390 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:40.398350 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-f2tlm" event={"ID":"f1e10a0f-13dd-44a0-9d79-e6375e12ac24","Type":"ContainerStarted","Data":"23b07d6db45d3b755a1b3c846148096ffd448412413d087eae6b98d1704746be"} Feb 17 12:47:42.403798 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:42.403715 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-f2tlm" event={"ID":"f1e10a0f-13dd-44a0-9d79-e6375e12ac24","Type":"ContainerStarted","Data":"288a376bc6d1a31a3d7f10f39b6c6053401a1cbf01923b7f73212cc42456fbbe"} Feb 17 12:47:42.404157 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:42.403829 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:47:42.417458 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:47:42.417318 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-diagnostics/network-check-target-f2tlm" podStartSLOduration=65.745932418 podStartE2EDuration="1m8.417304604s" podCreationTimestamp="2026-02-17 12:46:34 +0000 UTC" firstStartedPulling="2026-02-17 12:47:39.419331311 +0000 UTC m=+65.769742100" lastFinishedPulling="2026-02-17 12:47:42.090703482 +0000 UTC m=+68.441114286" observedRunningTime="2026-02-17 12:47:42.417089739 +0000 UTC m=+68.767500561" watchObservedRunningTime="2026-02-17 12:47:42.417304604 +0000 UTC m=+68.767715453" Feb 17 12:48:10.955570 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:10.955517 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-metrics-tls\") pod \"dns-default-b6hth\" (UID: \"a01f1f0a-fa4d-4b3c-a85e-053603fecfc1\") " pod="openshift-dns/dns-default-b6hth" Feb 17 12:48:10.955570 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:10.955577 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/51d19194-106c-4b3f-ba37-a0a7eedbc75a-cert\") pod \"ingress-canary-prcdr\" (UID: \"51d19194-106c-4b3f-ba37-a0a7eedbc75a\") " pod="openshift-ingress-canary/ingress-canary-prcdr" Feb 17 12:48:10.956089 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:10.955657 2572 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Feb 17 12:48:10.956089 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:10.955675 2572 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Feb 17 12:48:10.956089 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:10.955719 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/51d19194-106c-4b3f-ba37-a0a7eedbc75a-cert podName:51d19194-106c-4b3f-ba37-a0a7eedbc75a nodeName:}" failed. No retries permitted until 2026-02-17 12:49:14.955705118 +0000 UTC m=+161.306115921 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/51d19194-106c-4b3f-ba37-a0a7eedbc75a-cert") pod "ingress-canary-prcdr" (UID: "51d19194-106c-4b3f-ba37-a0a7eedbc75a") : secret "canary-serving-cert" not found Feb 17 12:48:10.956089 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:10.955792 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-metrics-tls podName:a01f1f0a-fa4d-4b3c-a85e-053603fecfc1 nodeName:}" failed. No retries permitted until 2026-02-17 12:49:14.955768659 +0000 UTC m=+161.306179462 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-metrics-tls") pod "dns-default-b6hth" (UID: "a01f1f0a-fa4d-4b3c-a85e-053603fecfc1") : secret "dns-default-metrics-tls" not found Feb 17 12:48:13.408486 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:13.408458 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-f2tlm" Feb 17 12:48:42.866510 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:42.866464 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs\") pod \"network-metrics-daemon-vmpwv\" (UID: \"685523da-d834-463a-8a1a-ef70069c19c5\") " pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:48:42.867004 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:42.866574 2572 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: secret "metrics-daemon-secret" not found Feb 17 12:48:42.867004 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:42.866630 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs podName:685523da-d834-463a-8a1a-ef70069c19c5 nodeName:}" failed. No retries permitted until 2026-02-17 12:50:44.866613324 +0000 UTC m=+251.217024114 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs") pod "network-metrics-daemon-vmpwv" (UID: "685523da-d834-463a-8a1a-ef70069c19c5") : secret "metrics-daemon-secret" not found Feb 17 12:48:51.000739 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.000708 2572 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-7bf5dd6d69-nwb99"] Feb 17 12:48:51.003511 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.003487 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-7bf5dd6d69-nwb99" Feb 17 12:48:51.005210 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.005188 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"samples-operator-tls\"" Feb 17 12:48:51.005328 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.005246 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-samples-operator\"/\"openshift-service-ca.crt\"" Feb 17 12:48:51.005593 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.005577 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"cluster-samples-operator-dockercfg-r4hwq\"" Feb 17 12:48:51.005639 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.005581 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-samples-operator\"/\"kube-root-ca.crt\"" Feb 17 12:48:51.013051 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.013020 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-7bf5dd6d69-nwb99"] Feb 17 12:48:51.099994 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.099954 2572 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-7764898fbb-x52mf"] Feb 17 12:48:51.102688 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.102670 2572 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-68598f8b47-r5pmz"] Feb 17 12:48:51.102824 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.102804 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:48:51.104701 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.104635 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-metrics-certs-default\"" Feb 17 12:48:51.104701 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.104679 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"kube-root-ca.crt\"" Feb 17 12:48:51.104926 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.104643 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"service-ca-bundle\"" Feb 17 12:48:51.104926 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.104646 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"openshift-service-ca.crt\"" Feb 17 12:48:51.104926 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.104679 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"default-ingress-cert\"" Feb 17 12:48:51.104926 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.104790 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-dockercfg-pbjp9\"" Feb 17 12:48:51.104926 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.104635 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-stats-default\"" Feb 17 12:48:51.105475 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.105460 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:51.106759 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.106743 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-tls\"" Feb 17 12:48:51.106759 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.106758 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-private-configuration\"" Feb 17 12:48:51.107237 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.107223 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"registry-dockercfg-68g2c\"" Feb 17 12:48:51.107533 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.107503 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"installation-pull-secrets\"" Feb 17 12:48:51.112536 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.112498 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"trusted-ca\"" Feb 17 12:48:51.113976 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.113939 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress/router-default-7764898fbb-x52mf"] Feb 17 12:48:51.116455 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.116433 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-68598f8b47-r5pmz"] Feb 17 12:48:51.119080 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.119063 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/1ac4bab6-4158-4d4e-91f1-ff2c43225f7d-samples-operator-tls\") pod \"cluster-samples-operator-7bf5dd6d69-nwb99\" (UID: \"1ac4bab6-4158-4d4e-91f1-ff2c43225f7d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-7bf5dd6d69-nwb99" Feb 17 12:48:51.119186 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.119095 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swh42\" (UniqueName: \"kubernetes.io/projected/1ac4bab6-4158-4d4e-91f1-ff2c43225f7d-kube-api-access-swh42\") pod \"cluster-samples-operator-7bf5dd6d69-nwb99\" (UID: \"1ac4bab6-4158-4d4e-91f1-ff2c43225f7d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-7bf5dd6d69-nwb99" Feb 17 12:48:51.219883 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.219855 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-bound-sa-token\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:51.220039 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.219887 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cf9b\" (UniqueName: \"kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-kube-api-access-4cf9b\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:51.220039 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.219908 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34367000-c224-43d7-9e8b-916d67860ee4-metrics-certs\") pod \"router-default-7764898fbb-x52mf\" (UID: \"34367000-c224-43d7-9e8b-916d67860ee4\") " pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:48:51.220039 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.219950 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/756003d3-7d4c-4fec-b738-1d8d45bf048c-registry-certificates\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:51.220039 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.219975 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/756003d3-7d4c-4fec-b738-1d8d45bf048c-image-registry-private-configuration\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:51.220039 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.220002 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-registry-tls\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:51.220039 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.220036 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flgd9\" (UniqueName: \"kubernetes.io/projected/34367000-c224-43d7-9e8b-916d67860ee4-kube-api-access-flgd9\") pod \"router-default-7764898fbb-x52mf\" (UID: \"34367000-c224-43d7-9e8b-916d67860ee4\") " pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:48:51.220257 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.220063 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/1ac4bab6-4158-4d4e-91f1-ff2c43225f7d-samples-operator-tls\") pod \"cluster-samples-operator-7bf5dd6d69-nwb99\" (UID: \"1ac4bab6-4158-4d4e-91f1-ff2c43225f7d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-7bf5dd6d69-nwb99" Feb 17 12:48:51.220257 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.220084 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-swh42\" (UniqueName: \"kubernetes.io/projected/1ac4bab6-4158-4d4e-91f1-ff2c43225f7d-kube-api-access-swh42\") pod \"cluster-samples-operator-7bf5dd6d69-nwb99\" (UID: \"1ac4bab6-4158-4d4e-91f1-ff2c43225f7d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-7bf5dd6d69-nwb99" Feb 17 12:48:51.220257 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.220127 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/756003d3-7d4c-4fec-b738-1d8d45bf048c-trusted-ca\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:51.220257 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.220160 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/756003d3-7d4c-4fec-b738-1d8d45bf048c-ca-trust-extracted\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:51.220257 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.220188 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/34367000-c224-43d7-9e8b-916d67860ee4-default-certificate\") pod \"router-default-7764898fbb-x52mf\" (UID: \"34367000-c224-43d7-9e8b-916d67860ee4\") " pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:48:51.220257 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:51.220196 2572 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Feb 17 12:48:51.220257 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.220244 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/756003d3-7d4c-4fec-b738-1d8d45bf048c-installation-pull-secrets\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:51.220257 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:51.220253 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1ac4bab6-4158-4d4e-91f1-ff2c43225f7d-samples-operator-tls podName:1ac4bab6-4158-4d4e-91f1-ff2c43225f7d nodeName:}" failed. No retries permitted until 2026-02-17 12:48:51.720237494 +0000 UTC m=+138.070648283 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/1ac4bab6-4158-4d4e-91f1-ff2c43225f7d-samples-operator-tls") pod "cluster-samples-operator-7bf5dd6d69-nwb99" (UID: "1ac4bab6-4158-4d4e-91f1-ff2c43225f7d") : secret "samples-operator-tls" not found Feb 17 12:48:51.220494 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.220281 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/34367000-c224-43d7-9e8b-916d67860ee4-service-ca-bundle\") pod \"router-default-7764898fbb-x52mf\" (UID: \"34367000-c224-43d7-9e8b-916d67860ee4\") " pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:48:51.220494 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.220303 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/34367000-c224-43d7-9e8b-916d67860ee4-stats-auth\") pod \"router-default-7764898fbb-x52mf\" (UID: \"34367000-c224-43d7-9e8b-916d67860ee4\") " pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:48:51.231136 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.231116 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-swh42\" (UniqueName: \"kubernetes.io/projected/1ac4bab6-4158-4d4e-91f1-ff2c43225f7d-kube-api-access-swh42\") pod \"cluster-samples-operator-7bf5dd6d69-nwb99\" (UID: \"1ac4bab6-4158-4d4e-91f1-ff2c43225f7d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-7bf5dd6d69-nwb99" Feb 17 12:48:51.321079 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.321000 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/756003d3-7d4c-4fec-b738-1d8d45bf048c-registry-certificates\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:51.321079 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.321049 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/756003d3-7d4c-4fec-b738-1d8d45bf048c-image-registry-private-configuration\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:51.321269 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.321095 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-registry-tls\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:51.321269 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.321162 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-flgd9\" (UniqueName: \"kubernetes.io/projected/34367000-c224-43d7-9e8b-916d67860ee4-kube-api-access-flgd9\") pod \"router-default-7764898fbb-x52mf\" (UID: \"34367000-c224-43d7-9e8b-916d67860ee4\") " pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:48:51.321269 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.321215 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/756003d3-7d4c-4fec-b738-1d8d45bf048c-trusted-ca\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:51.321269 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.321246 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/756003d3-7d4c-4fec-b738-1d8d45bf048c-ca-trust-extracted\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:51.321471 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.321274 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/34367000-c224-43d7-9e8b-916d67860ee4-default-certificate\") pod \"router-default-7764898fbb-x52mf\" (UID: \"34367000-c224-43d7-9e8b-916d67860ee4\") " pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:48:51.321471 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:51.321280 2572 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Feb 17 12:48:51.321471 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:51.321302 2572 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-68598f8b47-r5pmz: secret "image-registry-tls" not found Feb 17 12:48:51.321471 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.321305 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/756003d3-7d4c-4fec-b738-1d8d45bf048c-installation-pull-secrets\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:51.321471 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.321333 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/34367000-c224-43d7-9e8b-916d67860ee4-service-ca-bundle\") pod \"router-default-7764898fbb-x52mf\" (UID: \"34367000-c224-43d7-9e8b-916d67860ee4\") " pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:48:51.321471 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.321358 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/34367000-c224-43d7-9e8b-916d67860ee4-stats-auth\") pod \"router-default-7764898fbb-x52mf\" (UID: \"34367000-c224-43d7-9e8b-916d67860ee4\") " pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:48:51.321471 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:51.321371 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-registry-tls podName:756003d3-7d4c-4fec-b738-1d8d45bf048c nodeName:}" failed. No retries permitted until 2026-02-17 12:48:51.821351311 +0000 UTC m=+138.171762101 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-registry-tls") pod "image-registry-68598f8b47-r5pmz" (UID: "756003d3-7d4c-4fec-b738-1d8d45bf048c") : secret "image-registry-tls" not found Feb 17 12:48:51.321471 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.321428 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-bound-sa-token\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:51.321471 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.321456 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-4cf9b\" (UniqueName: \"kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-kube-api-access-4cf9b\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:51.321921 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.321481 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34367000-c224-43d7-9e8b-916d67860ee4-metrics-certs\") pod \"router-default-7764898fbb-x52mf\" (UID: \"34367000-c224-43d7-9e8b-916d67860ee4\") " pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:48:51.321921 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:51.321573 2572 secret.go:189] Couldn't get secret openshift-ingress/router-metrics-certs-default: secret "router-metrics-certs-default" not found Feb 17 12:48:51.321921 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:51.321919 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/34367000-c224-43d7-9e8b-916d67860ee4-metrics-certs podName:34367000-c224-43d7-9e8b-916d67860ee4 nodeName:}" failed. No retries permitted until 2026-02-17 12:48:51.821898806 +0000 UTC m=+138.172309618 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/34367000-c224-43d7-9e8b-916d67860ee4-metrics-certs") pod "router-default-7764898fbb-x52mf" (UID: "34367000-c224-43d7-9e8b-916d67860ee4") : secret "router-metrics-certs-default" not found Feb 17 12:48:51.322084 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:51.321947 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/34367000-c224-43d7-9e8b-916d67860ee4-service-ca-bundle podName:34367000-c224-43d7-9e8b-916d67860ee4 nodeName:}" failed. No retries permitted until 2026-02-17 12:48:51.821937132 +0000 UTC m=+138.172347951 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/34367000-c224-43d7-9e8b-916d67860ee4-service-ca-bundle") pod "router-default-7764898fbb-x52mf" (UID: "34367000-c224-43d7-9e8b-916d67860ee4") : configmap references non-existent config key: service-ca.crt Feb 17 12:48:51.322219 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.322166 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/756003d3-7d4c-4fec-b738-1d8d45bf048c-ca-trust-extracted\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:51.322328 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.322307 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/756003d3-7d4c-4fec-b738-1d8d45bf048c-registry-certificates\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:51.322883 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.322857 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/756003d3-7d4c-4fec-b738-1d8d45bf048c-trusted-ca\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:51.323800 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.323776 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/756003d3-7d4c-4fec-b738-1d8d45bf048c-image-registry-private-configuration\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:51.324228 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.324206 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/34367000-c224-43d7-9e8b-916d67860ee4-stats-auth\") pod \"router-default-7764898fbb-x52mf\" (UID: \"34367000-c224-43d7-9e8b-916d67860ee4\") " pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:48:51.324283 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.324211 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/756003d3-7d4c-4fec-b738-1d8d45bf048c-installation-pull-secrets\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:51.324283 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.324245 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/34367000-c224-43d7-9e8b-916d67860ee4-default-certificate\") pod \"router-default-7764898fbb-x52mf\" (UID: \"34367000-c224-43d7-9e8b-916d67860ee4\") " pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:48:51.329000 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.328978 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-bound-sa-token\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:51.329155 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.329135 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-flgd9\" (UniqueName: \"kubernetes.io/projected/34367000-c224-43d7-9e8b-916d67860ee4-kube-api-access-flgd9\") pod \"router-default-7764898fbb-x52mf\" (UID: \"34367000-c224-43d7-9e8b-916d67860ee4\") " pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:48:51.329611 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.329585 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cf9b\" (UniqueName: \"kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-kube-api-access-4cf9b\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:51.724861 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.724813 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/1ac4bab6-4158-4d4e-91f1-ff2c43225f7d-samples-operator-tls\") pod \"cluster-samples-operator-7bf5dd6d69-nwb99\" (UID: \"1ac4bab6-4158-4d4e-91f1-ff2c43225f7d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-7bf5dd6d69-nwb99" Feb 17 12:48:51.725030 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:51.724957 2572 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Feb 17 12:48:51.725074 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:51.725033 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1ac4bab6-4158-4d4e-91f1-ff2c43225f7d-samples-operator-tls podName:1ac4bab6-4158-4d4e-91f1-ff2c43225f7d nodeName:}" failed. No retries permitted until 2026-02-17 12:48:52.725015771 +0000 UTC m=+139.075426562 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/1ac4bab6-4158-4d4e-91f1-ff2c43225f7d-samples-operator-tls") pod "cluster-samples-operator-7bf5dd6d69-nwb99" (UID: "1ac4bab6-4158-4d4e-91f1-ff2c43225f7d") : secret "samples-operator-tls" not found Feb 17 12:48:51.825882 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.825847 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/34367000-c224-43d7-9e8b-916d67860ee4-service-ca-bundle\") pod \"router-default-7764898fbb-x52mf\" (UID: \"34367000-c224-43d7-9e8b-916d67860ee4\") " pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:48:51.825882 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.825890 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34367000-c224-43d7-9e8b-916d67860ee4-metrics-certs\") pod \"router-default-7764898fbb-x52mf\" (UID: \"34367000-c224-43d7-9e8b-916d67860ee4\") " pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:48:51.826130 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:51.825926 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-registry-tls\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:51.826130 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:51.826001 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/34367000-c224-43d7-9e8b-916d67860ee4-service-ca-bundle podName:34367000-c224-43d7-9e8b-916d67860ee4 nodeName:}" failed. No retries permitted until 2026-02-17 12:48:52.825983095 +0000 UTC m=+139.176393915 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/34367000-c224-43d7-9e8b-916d67860ee4-service-ca-bundle") pod "router-default-7764898fbb-x52mf" (UID: "34367000-c224-43d7-9e8b-916d67860ee4") : configmap references non-existent config key: service-ca.crt Feb 17 12:48:51.826130 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:51.826046 2572 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Feb 17 12:48:51.826130 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:51.826061 2572 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-68598f8b47-r5pmz: secret "image-registry-tls" not found Feb 17 12:48:51.826130 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:51.826067 2572 secret.go:189] Couldn't get secret openshift-ingress/router-metrics-certs-default: secret "router-metrics-certs-default" not found Feb 17 12:48:51.826130 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:51.826124 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-registry-tls podName:756003d3-7d4c-4fec-b738-1d8d45bf048c nodeName:}" failed. No retries permitted until 2026-02-17 12:48:52.826094244 +0000 UTC m=+139.176505038 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-registry-tls") pod "image-registry-68598f8b47-r5pmz" (UID: "756003d3-7d4c-4fec-b738-1d8d45bf048c") : secret "image-registry-tls" not found Feb 17 12:48:51.826341 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:51.826141 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/34367000-c224-43d7-9e8b-916d67860ee4-metrics-certs podName:34367000-c224-43d7-9e8b-916d67860ee4 nodeName:}" failed. No retries permitted until 2026-02-17 12:48:52.826135114 +0000 UTC m=+139.176545903 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/34367000-c224-43d7-9e8b-916d67860ee4-metrics-certs") pod "router-default-7764898fbb-x52mf" (UID: "34367000-c224-43d7-9e8b-916d67860ee4") : secret "router-metrics-certs-default" not found Feb 17 12:48:52.732518 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:52.732474 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/1ac4bab6-4158-4d4e-91f1-ff2c43225f7d-samples-operator-tls\") pod \"cluster-samples-operator-7bf5dd6d69-nwb99\" (UID: \"1ac4bab6-4158-4d4e-91f1-ff2c43225f7d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-7bf5dd6d69-nwb99" Feb 17 12:48:52.732999 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:52.732606 2572 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Feb 17 12:48:52.732999 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:52.732684 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1ac4bab6-4158-4d4e-91f1-ff2c43225f7d-samples-operator-tls podName:1ac4bab6-4158-4d4e-91f1-ff2c43225f7d nodeName:}" failed. No retries permitted until 2026-02-17 12:48:54.732660736 +0000 UTC m=+141.083071532 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/1ac4bab6-4158-4d4e-91f1-ff2c43225f7d-samples-operator-tls") pod "cluster-samples-operator-7bf5dd6d69-nwb99" (UID: "1ac4bab6-4158-4d4e-91f1-ff2c43225f7d") : secret "samples-operator-tls" not found Feb 17 12:48:52.833034 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:52.832993 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/34367000-c224-43d7-9e8b-916d67860ee4-service-ca-bundle\") pod \"router-default-7764898fbb-x52mf\" (UID: \"34367000-c224-43d7-9e8b-916d67860ee4\") " pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:48:52.833214 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:52.833044 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34367000-c224-43d7-9e8b-916d67860ee4-metrics-certs\") pod \"router-default-7764898fbb-x52mf\" (UID: \"34367000-c224-43d7-9e8b-916d67860ee4\") " pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:48:52.833214 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:52.833082 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-registry-tls\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:52.833214 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:52.833161 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/34367000-c224-43d7-9e8b-916d67860ee4-service-ca-bundle podName:34367000-c224-43d7-9e8b-916d67860ee4 nodeName:}" failed. No retries permitted until 2026-02-17 12:48:54.833143626 +0000 UTC m=+141.183554421 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/34367000-c224-43d7-9e8b-916d67860ee4-service-ca-bundle") pod "router-default-7764898fbb-x52mf" (UID: "34367000-c224-43d7-9e8b-916d67860ee4") : configmap references non-existent config key: service-ca.crt Feb 17 12:48:52.833214 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:52.833198 2572 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Feb 17 12:48:52.833214 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:52.833211 2572 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-68598f8b47-r5pmz: secret "image-registry-tls" not found Feb 17 12:48:52.833214 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:52.833199 2572 secret.go:189] Couldn't get secret openshift-ingress/router-metrics-certs-default: secret "router-metrics-certs-default" not found Feb 17 12:48:52.833444 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:52.833247 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-registry-tls podName:756003d3-7d4c-4fec-b738-1d8d45bf048c nodeName:}" failed. No retries permitted until 2026-02-17 12:48:54.833234854 +0000 UTC m=+141.183645644 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-registry-tls") pod "image-registry-68598f8b47-r5pmz" (UID: "756003d3-7d4c-4fec-b738-1d8d45bf048c") : secret "image-registry-tls" not found Feb 17 12:48:52.833444 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:52.833272 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/34367000-c224-43d7-9e8b-916d67860ee4-metrics-certs podName:34367000-c224-43d7-9e8b-916d67860ee4 nodeName:}" failed. No retries permitted until 2026-02-17 12:48:54.833252051 +0000 UTC m=+141.183662843 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/34367000-c224-43d7-9e8b-916d67860ee4-metrics-certs") pod "router-default-7764898fbb-x52mf" (UID: "34367000-c224-43d7-9e8b-916d67860ee4") : secret "router-metrics-certs-default" not found Feb 17 12:48:54.747123 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:54.747058 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/1ac4bab6-4158-4d4e-91f1-ff2c43225f7d-samples-operator-tls\") pod \"cluster-samples-operator-7bf5dd6d69-nwb99\" (UID: \"1ac4bab6-4158-4d4e-91f1-ff2c43225f7d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-7bf5dd6d69-nwb99" Feb 17 12:48:54.747505 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:54.747211 2572 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Feb 17 12:48:54.747505 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:54.747279 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1ac4bab6-4158-4d4e-91f1-ff2c43225f7d-samples-operator-tls podName:1ac4bab6-4158-4d4e-91f1-ff2c43225f7d nodeName:}" failed. No retries permitted until 2026-02-17 12:48:58.747261229 +0000 UTC m=+145.097672036 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/1ac4bab6-4158-4d4e-91f1-ff2c43225f7d-samples-operator-tls") pod "cluster-samples-operator-7bf5dd6d69-nwb99" (UID: "1ac4bab6-4158-4d4e-91f1-ff2c43225f7d") : secret "samples-operator-tls" not found Feb 17 12:48:54.847638 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:54.847593 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34367000-c224-43d7-9e8b-916d67860ee4-metrics-certs\") pod \"router-default-7764898fbb-x52mf\" (UID: \"34367000-c224-43d7-9e8b-916d67860ee4\") " pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:48:54.847809 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:54.847656 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-registry-tls\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:54.847809 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:54.847706 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/34367000-c224-43d7-9e8b-916d67860ee4-service-ca-bundle\") pod \"router-default-7764898fbb-x52mf\" (UID: \"34367000-c224-43d7-9e8b-916d67860ee4\") " pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:48:54.847809 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:54.847750 2572 secret.go:189] Couldn't get secret openshift-ingress/router-metrics-certs-default: secret "router-metrics-certs-default" not found Feb 17 12:48:54.847809 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:54.847809 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/34367000-c224-43d7-9e8b-916d67860ee4-metrics-certs podName:34367000-c224-43d7-9e8b-916d67860ee4 nodeName:}" failed. No retries permitted until 2026-02-17 12:48:58.84779364 +0000 UTC m=+145.198204430 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/34367000-c224-43d7-9e8b-916d67860ee4-metrics-certs") pod "router-default-7764898fbb-x52mf" (UID: "34367000-c224-43d7-9e8b-916d67860ee4") : secret "router-metrics-certs-default" not found Feb 17 12:48:54.847937 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:54.847826 2572 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Feb 17 12:48:54.847937 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:54.847846 2572 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-68598f8b47-r5pmz: secret "image-registry-tls" not found Feb 17 12:48:54.847937 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:54.847910 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-registry-tls podName:756003d3-7d4c-4fec-b738-1d8d45bf048c nodeName:}" failed. No retries permitted until 2026-02-17 12:48:58.847897426 +0000 UTC m=+145.198308228 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-registry-tls") pod "image-registry-68598f8b47-r5pmz" (UID: "756003d3-7d4c-4fec-b738-1d8d45bf048c") : secret "image-registry-tls" not found Feb 17 12:48:54.847937 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:54.847924 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/34367000-c224-43d7-9e8b-916d67860ee4-service-ca-bundle podName:34367000-c224-43d7-9e8b-916d67860ee4 nodeName:}" failed. No retries permitted until 2026-02-17 12:48:58.847918513 +0000 UTC m=+145.198329303 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/34367000-c224-43d7-9e8b-916d67860ee4-service-ca-bundle") pod "router-default-7764898fbb-x52mf" (UID: "34367000-c224-43d7-9e8b-916d67860ee4") : configmap references non-existent config key: service-ca.crt Feb 17 12:48:57.677154 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:57.677116 2572 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-75f6d74b65-gtbxg"] Feb 17 12:48:57.680954 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:57.680925 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-75f6d74b65-gtbxg" Feb 17 12:48:57.682654 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:57.682633 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator\"/\"openshift-service-ca.crt\"" Feb 17 12:48:57.682654 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:57.682644 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator\"/\"kube-storage-version-migrator-sa-dockercfg-dqtkp\"" Feb 17 12:48:57.682834 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:57.682820 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator\"/\"kube-root-ca.crt\"" Feb 17 12:48:57.686580 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:57.686553 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-75f6d74b65-gtbxg"] Feb 17 12:48:57.768034 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:57.767999 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6l7r\" (UniqueName: \"kubernetes.io/projected/3dfbc9a4-8a6f-4466-88ee-8a276d190f2e-kube-api-access-h6l7r\") pod \"migrator-75f6d74b65-gtbxg\" (UID: \"3dfbc9a4-8a6f-4466-88ee-8a276d190f2e\") " pod="openshift-kube-storage-version-migrator/migrator-75f6d74b65-gtbxg" Feb 17 12:48:57.868728 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:57.868677 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-h6l7r\" (UniqueName: \"kubernetes.io/projected/3dfbc9a4-8a6f-4466-88ee-8a276d190f2e-kube-api-access-h6l7r\") pod \"migrator-75f6d74b65-gtbxg\" (UID: \"3dfbc9a4-8a6f-4466-88ee-8a276d190f2e\") " pod="openshift-kube-storage-version-migrator/migrator-75f6d74b65-gtbxg" Feb 17 12:48:57.877354 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:57.877319 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6l7r\" (UniqueName: \"kubernetes.io/projected/3dfbc9a4-8a6f-4466-88ee-8a276d190f2e-kube-api-access-h6l7r\") pod \"migrator-75f6d74b65-gtbxg\" (UID: \"3dfbc9a4-8a6f-4466-88ee-8a276d190f2e\") " pod="openshift-kube-storage-version-migrator/migrator-75f6d74b65-gtbxg" Feb 17 12:48:57.990004 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:57.989977 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-75f6d74b65-gtbxg" Feb 17 12:48:58.098794 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:58.098761 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-75f6d74b65-gtbxg"] Feb 17 12:48:58.101893 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:48:58.101863 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dfbc9a4_8a6f_4466_88ee_8a276d190f2e.slice/crio-13e663f42f7c616f555b244c51e05c2f93a9b94b20a506d54244e6a117abd51b WatchSource:0}: Error finding container 13e663f42f7c616f555b244c51e05c2f93a9b94b20a506d54244e6a117abd51b: Status 404 returned error can't find the container with id 13e663f42f7c616f555b244c51e05c2f93a9b94b20a506d54244e6a117abd51b Feb 17 12:48:58.542035 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:58.541994 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-75f6d74b65-gtbxg" event={"ID":"3dfbc9a4-8a6f-4466-88ee-8a276d190f2e","Type":"ContainerStarted","Data":"13e663f42f7c616f555b244c51e05c2f93a9b94b20a506d54244e6a117abd51b"} Feb 17 12:48:58.777547 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:58.777508 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/1ac4bab6-4158-4d4e-91f1-ff2c43225f7d-samples-operator-tls\") pod \"cluster-samples-operator-7bf5dd6d69-nwb99\" (UID: \"1ac4bab6-4158-4d4e-91f1-ff2c43225f7d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-7bf5dd6d69-nwb99" Feb 17 12:48:58.777977 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:58.777658 2572 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: secret "samples-operator-tls" not found Feb 17 12:48:58.777977 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:58.777725 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1ac4bab6-4158-4d4e-91f1-ff2c43225f7d-samples-operator-tls podName:1ac4bab6-4158-4d4e-91f1-ff2c43225f7d nodeName:}" failed. No retries permitted until 2026-02-17 12:49:06.77770916 +0000 UTC m=+153.128119950 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/1ac4bab6-4158-4d4e-91f1-ff2c43225f7d-samples-operator-tls") pod "cluster-samples-operator-7bf5dd6d69-nwb99" (UID: "1ac4bab6-4158-4d4e-91f1-ff2c43225f7d") : secret "samples-operator-tls" not found Feb 17 12:48:58.878832 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:58.878745 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-registry-tls\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:48:58.879025 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:58.878870 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/34367000-c224-43d7-9e8b-916d67860ee4-service-ca-bundle\") pod \"router-default-7764898fbb-x52mf\" (UID: \"34367000-c224-43d7-9e8b-916d67860ee4\") " pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:48:58.879025 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:58.878877 2572 projected.go:264] Couldn't get secret openshift-image-registry/image-registry-tls: secret "image-registry-tls" not found Feb 17 12:48:58.879025 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:58.878898 2572 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-68598f8b47-r5pmz: secret "image-registry-tls" not found Feb 17 12:48:58.879025 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:58.878922 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34367000-c224-43d7-9e8b-916d67860ee4-metrics-certs\") pod \"router-default-7764898fbb-x52mf\" (UID: \"34367000-c224-43d7-9e8b-916d67860ee4\") " pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:48:58.879025 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:58.878959 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-registry-tls podName:756003d3-7d4c-4fec-b738-1d8d45bf048c nodeName:}" failed. No retries permitted until 2026-02-17 12:49:06.878939017 +0000 UTC m=+153.229349807 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-registry-tls") pod "image-registry-68598f8b47-r5pmz" (UID: "756003d3-7d4c-4fec-b738-1d8d45bf048c") : secret "image-registry-tls" not found Feb 17 12:48:58.879025 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:58.879016 2572 secret.go:189] Couldn't get secret openshift-ingress/router-metrics-certs-default: secret "router-metrics-certs-default" not found Feb 17 12:48:58.879362 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:58.879045 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/34367000-c224-43d7-9e8b-916d67860ee4-service-ca-bundle podName:34367000-c224-43d7-9e8b-916d67860ee4 nodeName:}" failed. No retries permitted until 2026-02-17 12:49:06.879027841 +0000 UTC m=+153.229438645 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/34367000-c224-43d7-9e8b-916d67860ee4-service-ca-bundle") pod "router-default-7764898fbb-x52mf" (UID: "34367000-c224-43d7-9e8b-916d67860ee4") : configmap references non-existent config key: service-ca.crt Feb 17 12:48:58.879362 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:48:58.879068 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/34367000-c224-43d7-9e8b-916d67860ee4-metrics-certs podName:34367000-c224-43d7-9e8b-916d67860ee4 nodeName:}" failed. No retries permitted until 2026-02-17 12:49:06.87905744 +0000 UTC m=+153.229468233 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/34367000-c224-43d7-9e8b-916d67860ee4-metrics-certs") pod "router-default-7764898fbb-x52mf" (UID: "34367000-c224-43d7-9e8b-916d67860ee4") : secret "router-metrics-certs-default" not found Feb 17 12:48:58.970070 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:58.970039 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_node-resolver-jqfj5_22f3475b-a5d2-4d66-818f-5010e57aac8e/dns-node-resolver/0.log" Feb 17 12:48:59.545348 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:59.545313 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-75f6d74b65-gtbxg" event={"ID":"3dfbc9a4-8a6f-4466-88ee-8a276d190f2e","Type":"ContainerStarted","Data":"e0b724e98347296b709959a77545f30f03fc532cc60120cf4487b9c7dca5cfbd"} Feb 17 12:48:59.545348 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:59.545352 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-75f6d74b65-gtbxg" event={"ID":"3dfbc9a4-8a6f-4466-88ee-8a276d190f2e","Type":"ContainerStarted","Data":"7ac7f960dabc8a1124f48e189e50aad1bc3e103f423faf12f808f9e809b45570"} Feb 17 12:48:59.561119 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:59.561068 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-75f6d74b65-gtbxg" podStartSLOduration=1.4966567290000001 podStartE2EDuration="2.561054388s" podCreationTimestamp="2026-02-17 12:48:57 +0000 UTC" firstStartedPulling="2026-02-17 12:48:58.10363316 +0000 UTC m=+144.454043963" lastFinishedPulling="2026-02-17 12:48:59.168030828 +0000 UTC m=+145.518441622" observedRunningTime="2026-02-17 12:48:59.560431946 +0000 UTC m=+145.910842758" watchObservedRunningTime="2026-02-17 12:48:59.561054388 +0000 UTC m=+145.911465192" Feb 17 12:48:59.969207 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:48:59.969180 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_node-ca-qjw6p_8738b60c-920d-4bc5-933f-01d0fb8d68d3/node-ca/0.log" Feb 17 12:49:06.839903 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:06.839863 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/1ac4bab6-4158-4d4e-91f1-ff2c43225f7d-samples-operator-tls\") pod \"cluster-samples-operator-7bf5dd6d69-nwb99\" (UID: \"1ac4bab6-4158-4d4e-91f1-ff2c43225f7d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-7bf5dd6d69-nwb99" Feb 17 12:49:06.842219 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:06.842201 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/1ac4bab6-4158-4d4e-91f1-ff2c43225f7d-samples-operator-tls\") pod \"cluster-samples-operator-7bf5dd6d69-nwb99\" (UID: \"1ac4bab6-4158-4d4e-91f1-ff2c43225f7d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-7bf5dd6d69-nwb99" Feb 17 12:49:06.913016 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:06.912974 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-7bf5dd6d69-nwb99" Feb 17 12:49:06.940928 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:06.940893 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/34367000-c224-43d7-9e8b-916d67860ee4-service-ca-bundle\") pod \"router-default-7764898fbb-x52mf\" (UID: \"34367000-c224-43d7-9e8b-916d67860ee4\") " pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:49:06.941072 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:06.940942 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34367000-c224-43d7-9e8b-916d67860ee4-metrics-certs\") pod \"router-default-7764898fbb-x52mf\" (UID: \"34367000-c224-43d7-9e8b-916d67860ee4\") " pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:49:06.941072 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:06.941048 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-registry-tls\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:49:06.941528 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:06.941507 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/34367000-c224-43d7-9e8b-916d67860ee4-service-ca-bundle\") pod \"router-default-7764898fbb-x52mf\" (UID: \"34367000-c224-43d7-9e8b-916d67860ee4\") " pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:49:06.943445 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:06.943422 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-registry-tls\") pod \"image-registry-68598f8b47-r5pmz\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:49:06.943653 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:06.943631 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34367000-c224-43d7-9e8b-916d67860ee4-metrics-certs\") pod \"router-default-7764898fbb-x52mf\" (UID: \"34367000-c224-43d7-9e8b-916d67860ee4\") " pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:49:07.013778 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:07.013739 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:49:07.019507 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:07.019479 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:49:07.028155 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:07.028093 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-7bf5dd6d69-nwb99"] Feb 17 12:49:07.137549 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:07.137518 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress/router-default-7764898fbb-x52mf"] Feb 17 12:49:07.140679 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:49:07.140640 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34367000_c224_43d7_9e8b_916d67860ee4.slice/crio-357e71327b0acd8de87d99d19e85708c350456d25ea636297ff75f64ab2affa6 WatchSource:0}: Error finding container 357e71327b0acd8de87d99d19e85708c350456d25ea636297ff75f64ab2affa6: Status 404 returned error can't find the container with id 357e71327b0acd8de87d99d19e85708c350456d25ea636297ff75f64ab2affa6 Feb 17 12:49:07.152347 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:07.152323 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-68598f8b47-r5pmz"] Feb 17 12:49:07.156300 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:49:07.156271 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod756003d3_7d4c_4fec_b738_1d8d45bf048c.slice/crio-cd4901d2c05a7ae08422f57471d55a04ab1d86d8d7ae104d4933efde3574e670 WatchSource:0}: Error finding container cd4901d2c05a7ae08422f57471d55a04ab1d86d8d7ae104d4933efde3574e670: Status 404 returned error can't find the container with id cd4901d2c05a7ae08422f57471d55a04ab1d86d8d7ae104d4933efde3574e670 Feb 17 12:49:07.562812 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:07.562764 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-7bf5dd6d69-nwb99" event={"ID":"1ac4bab6-4158-4d4e-91f1-ff2c43225f7d","Type":"ContainerStarted","Data":"e07ebaa4f9efce0b8d8dfe628800ce83c3bacaea67250b9a96c35226f425e13b"} Feb 17 12:49:07.564242 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:07.564208 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" event={"ID":"756003d3-7d4c-4fec-b738-1d8d45bf048c","Type":"ContainerStarted","Data":"c4999988293f533ed9aad40d6664284b6887f8904a29bdfa9aa3650002705574"} Feb 17 12:49:07.564394 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:07.564248 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" event={"ID":"756003d3-7d4c-4fec-b738-1d8d45bf048c","Type":"ContainerStarted","Data":"cd4901d2c05a7ae08422f57471d55a04ab1d86d8d7ae104d4933efde3574e670"} Feb 17 12:49:07.564394 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:07.564330 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:49:07.565564 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:07.565541 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-7764898fbb-x52mf" event={"ID":"34367000-c224-43d7-9e8b-916d67860ee4","Type":"ContainerStarted","Data":"6376f3c32843208c1affc005da24035750b8b67a18ccee87c9939c03c797dbd4"} Feb 17 12:49:07.565564 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:07.565569 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-7764898fbb-x52mf" event={"ID":"34367000-c224-43d7-9e8b-916d67860ee4","Type":"ContainerStarted","Data":"357e71327b0acd8de87d99d19e85708c350456d25ea636297ff75f64ab2affa6"} Feb 17 12:49:07.580749 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:07.580685 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" podStartSLOduration=16.580665548 podStartE2EDuration="16.580665548s" podCreationTimestamp="2026-02-17 12:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 12:49:07.580128947 +0000 UTC m=+153.930539759" watchObservedRunningTime="2026-02-17 12:49:07.580665548 +0000 UTC m=+153.931076364" Feb 17 12:49:07.595728 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:07.595671 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-7764898fbb-x52mf" podStartSLOduration=16.595651027 podStartE2EDuration="16.595651027s" podCreationTimestamp="2026-02-17 12:48:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 12:49:07.594981412 +0000 UTC m=+153.945392226" watchObservedRunningTime="2026-02-17 12:49:07.595651027 +0000 UTC m=+153.946061840" Feb 17 12:49:08.015005 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:08.014962 2572 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:49:08.017757 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:08.017731 2572 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:49:08.567950 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:08.567916 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:49:08.569149 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:08.569128 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-7764898fbb-x52mf" Feb 17 12:49:09.571582 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:09.571548 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-7bf5dd6d69-nwb99" event={"ID":"1ac4bab6-4158-4d4e-91f1-ff2c43225f7d","Type":"ContainerStarted","Data":"8f8977f2b708413ac740aa94f8de9b257161a08abc85807b56177549ff7fd24b"} Feb 17 12:49:09.571582 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:09.571584 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-7bf5dd6d69-nwb99" event={"ID":"1ac4bab6-4158-4d4e-91f1-ff2c43225f7d","Type":"ContainerStarted","Data":"a6e1ada685f446ce3caa09d4b708e047d077fbde64d551def4429d9fda3c18fc"} Feb 17 12:49:09.585694 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:09.585645 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-7bf5dd6d69-nwb99" podStartSLOduration=18.062423721 podStartE2EDuration="19.585631437s" podCreationTimestamp="2026-02-17 12:48:50 +0000 UTC" firstStartedPulling="2026-02-17 12:49:07.07027635 +0000 UTC m=+153.420687143" lastFinishedPulling="2026-02-17 12:49:08.593484054 +0000 UTC m=+154.943894859" observedRunningTime="2026-02-17 12:49:09.585119417 +0000 UTC m=+155.935530225" watchObservedRunningTime="2026-02-17 12:49:09.585631437 +0000 UTC m=+155.936042248" Feb 17 12:49:10.091732 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:49:10.091680 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[metrics-tls], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-dns/dns-default-b6hth" podUID="a01f1f0a-fa4d-4b3c-a85e-053603fecfc1" Feb 17 12:49:10.107500 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:49:10.107472 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[cert], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-ingress-canary/ingress-canary-prcdr" podUID="51d19194-106c-4b3f-ba37-a0a7eedbc75a" Feb 17 12:49:10.142456 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:49:10.142425 2572 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[metrics-certs], unattached volumes=[], failed to process volumes=[]: context deadline exceeded" pod="openshift-multus/network-metrics-daemon-vmpwv" podUID="685523da-d834-463a-8a1a-ef70069c19c5" Feb 17 12:49:10.574373 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:10.574344 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-b6hth" Feb 17 12:49:15.003675 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:15.003635 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-metrics-tls\") pod \"dns-default-b6hth\" (UID: \"a01f1f0a-fa4d-4b3c-a85e-053603fecfc1\") " pod="openshift-dns/dns-default-b6hth" Feb 17 12:49:15.003675 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:15.003678 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/51d19194-106c-4b3f-ba37-a0a7eedbc75a-cert\") pod \"ingress-canary-prcdr\" (UID: \"51d19194-106c-4b3f-ba37-a0a7eedbc75a\") " pod="openshift-ingress-canary/ingress-canary-prcdr" Feb 17 12:49:15.006003 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:15.005978 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a01f1f0a-fa4d-4b3c-a85e-053603fecfc1-metrics-tls\") pod \"dns-default-b6hth\" (UID: \"a01f1f0a-fa4d-4b3c-a85e-053603fecfc1\") " pod="openshift-dns/dns-default-b6hth" Feb 17 12:49:15.006258 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:15.006236 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/51d19194-106c-4b3f-ba37-a0a7eedbc75a-cert\") pod \"ingress-canary-prcdr\" (UID: \"51d19194-106c-4b3f-ba37-a0a7eedbc75a\") " pod="openshift-ingress-canary/ingress-canary-prcdr" Feb 17 12:49:15.077701 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:15.077668 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-dockercfg-944wh\"" Feb 17 12:49:15.085424 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:15.085406 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-b6hth" Feb 17 12:49:15.197737 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:15.197706 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-b6hth"] Feb 17 12:49:15.200888 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:49:15.200852 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda01f1f0a_fa4d_4b3c_a85e_053603fecfc1.slice/crio-d598d07fee96d5838c4f09350b89927507e795d7b9266c5a523b42efebc1449f WatchSource:0}: Error finding container d598d07fee96d5838c4f09350b89927507e795d7b9266c5a523b42efebc1449f: Status 404 returned error can't find the container with id d598d07fee96d5838c4f09350b89927507e795d7b9266c5a523b42efebc1449f Feb 17 12:49:15.587383 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:15.587347 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-b6hth" event={"ID":"a01f1f0a-fa4d-4b3c-a85e-053603fecfc1","Type":"ContainerStarted","Data":"d598d07fee96d5838c4f09350b89927507e795d7b9266c5a523b42efebc1449f"} Feb 17 12:49:16.596527 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:16.596494 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-b6hth" event={"ID":"a01f1f0a-fa4d-4b3c-a85e-053603fecfc1","Type":"ContainerStarted","Data":"c22da0e9287c9c5c66c9e374d1bee8f1a8bd7b17e0cc421129436b185f12d47c"} Feb 17 12:49:17.401260 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.401178 2572 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-insights/insights-runtime-extractor-hgpw7"] Feb 17 12:49:17.404200 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.404180 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-runtime-extractor-hgpw7" Feb 17 12:49:17.406703 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.406652 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"insights-runtime-extractor-tls\"" Feb 17 12:49:17.406703 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.406669 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"openshift-service-ca.crt\"" Feb 17 12:49:17.406866 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.406654 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"kube-root-ca.crt\"" Feb 17 12:49:17.406866 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.406653 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-insights\"/\"insights-runtime-extractor-sa-dockercfg-hv7tp\"" Feb 17 12:49:17.406866 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.406653 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-insights\"/\"kube-rbac-proxy\"" Feb 17 12:49:17.414142 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.414122 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-runtime-extractor-hgpw7"] Feb 17 12:49:17.422390 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.422366 2572 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-68598f8b47-r5pmz"] Feb 17 12:49:17.522926 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.522898 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/72276741-8f08-4c0c-98ca-d1614d018232-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-hgpw7\" (UID: \"72276741-8f08-4c0c-98ca-d1614d018232\") " pod="openshift-insights/insights-runtime-extractor-hgpw7" Feb 17 12:49:17.523067 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.522935 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/72276741-8f08-4c0c-98ca-d1614d018232-data-volume\") pod \"insights-runtime-extractor-hgpw7\" (UID: \"72276741-8f08-4c0c-98ca-d1614d018232\") " pod="openshift-insights/insights-runtime-extractor-hgpw7" Feb 17 12:49:17.523067 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.522997 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/72276741-8f08-4c0c-98ca-d1614d018232-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-hgpw7\" (UID: \"72276741-8f08-4c0c-98ca-d1614d018232\") " pod="openshift-insights/insights-runtime-extractor-hgpw7" Feb 17 12:49:17.523067 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.523019 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dbvt\" (UniqueName: \"kubernetes.io/projected/72276741-8f08-4c0c-98ca-d1614d018232-kube-api-access-5dbvt\") pod \"insights-runtime-extractor-hgpw7\" (UID: \"72276741-8f08-4c0c-98ca-d1614d018232\") " pod="openshift-insights/insights-runtime-extractor-hgpw7" Feb 17 12:49:17.523207 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.523140 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/72276741-8f08-4c0c-98ca-d1614d018232-crio-socket\") pod \"insights-runtime-extractor-hgpw7\" (UID: \"72276741-8f08-4c0c-98ca-d1614d018232\") " pod="openshift-insights/insights-runtime-extractor-hgpw7" Feb 17 12:49:17.601124 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.601079 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-b6hth" event={"ID":"a01f1f0a-fa4d-4b3c-a85e-053603fecfc1","Type":"ContainerStarted","Data":"e48ebbf82d6fc7d05f030873088827e87771918e85fa85f9cd6773b9f271a437"} Feb 17 12:49:17.601508 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.601241 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-dns/dns-default-b6hth" Feb 17 12:49:17.617869 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.617803 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-b6hth" podStartSLOduration=129.359576939 podStartE2EDuration="2m10.617785725s" podCreationTimestamp="2026-02-17 12:47:07 +0000 UTC" firstStartedPulling="2026-02-17 12:49:15.202753864 +0000 UTC m=+161.553164667" lastFinishedPulling="2026-02-17 12:49:16.460962663 +0000 UTC m=+162.811373453" observedRunningTime="2026-02-17 12:49:17.616629091 +0000 UTC m=+163.967039903" watchObservedRunningTime="2026-02-17 12:49:17.617785725 +0000 UTC m=+163.968196540" Feb 17 12:49:17.623595 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.623570 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/72276741-8f08-4c0c-98ca-d1614d018232-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-hgpw7\" (UID: \"72276741-8f08-4c0c-98ca-d1614d018232\") " pod="openshift-insights/insights-runtime-extractor-hgpw7" Feb 17 12:49:17.623708 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.623605 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/72276741-8f08-4c0c-98ca-d1614d018232-data-volume\") pod \"insights-runtime-extractor-hgpw7\" (UID: \"72276741-8f08-4c0c-98ca-d1614d018232\") " pod="openshift-insights/insights-runtime-extractor-hgpw7" Feb 17 12:49:17.623708 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.623647 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/72276741-8f08-4c0c-98ca-d1614d018232-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-hgpw7\" (UID: \"72276741-8f08-4c0c-98ca-d1614d018232\") " pod="openshift-insights/insights-runtime-extractor-hgpw7" Feb 17 12:49:17.623708 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.623676 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-5dbvt\" (UniqueName: \"kubernetes.io/projected/72276741-8f08-4c0c-98ca-d1614d018232-kube-api-access-5dbvt\") pod \"insights-runtime-extractor-hgpw7\" (UID: \"72276741-8f08-4c0c-98ca-d1614d018232\") " pod="openshift-insights/insights-runtime-extractor-hgpw7" Feb 17 12:49:17.623866 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.623718 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/72276741-8f08-4c0c-98ca-d1614d018232-crio-socket\") pod \"insights-runtime-extractor-hgpw7\" (UID: \"72276741-8f08-4c0c-98ca-d1614d018232\") " pod="openshift-insights/insights-runtime-extractor-hgpw7" Feb 17 12:49:17.623866 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.623816 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"crio-socket\" (UniqueName: \"kubernetes.io/host-path/72276741-8f08-4c0c-98ca-d1614d018232-crio-socket\") pod \"insights-runtime-extractor-hgpw7\" (UID: \"72276741-8f08-4c0c-98ca-d1614d018232\") " pod="openshift-insights/insights-runtime-extractor-hgpw7" Feb 17 12:49:17.624001 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.623972 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"data-volume\" (UniqueName: \"kubernetes.io/empty-dir/72276741-8f08-4c0c-98ca-d1614d018232-data-volume\") pod \"insights-runtime-extractor-hgpw7\" (UID: \"72276741-8f08-4c0c-98ca-d1614d018232\") " pod="openshift-insights/insights-runtime-extractor-hgpw7" Feb 17 12:49:17.624262 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.624243 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-rbac-proxy-cm\" (UniqueName: \"kubernetes.io/configmap/72276741-8f08-4c0c-98ca-d1614d018232-kube-rbac-proxy-cm\") pod \"insights-runtime-extractor-hgpw7\" (UID: \"72276741-8f08-4c0c-98ca-d1614d018232\") " pod="openshift-insights/insights-runtime-extractor-hgpw7" Feb 17 12:49:17.626350 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.626332 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"insights-runtime-extractor-tls\" (UniqueName: \"kubernetes.io/secret/72276741-8f08-4c0c-98ca-d1614d018232-insights-runtime-extractor-tls\") pod \"insights-runtime-extractor-hgpw7\" (UID: \"72276741-8f08-4c0c-98ca-d1614d018232\") " pod="openshift-insights/insights-runtime-extractor-hgpw7" Feb 17 12:49:17.639048 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.639025 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dbvt\" (UniqueName: \"kubernetes.io/projected/72276741-8f08-4c0c-98ca-d1614d018232-kube-api-access-5dbvt\") pod \"insights-runtime-extractor-hgpw7\" (UID: \"72276741-8f08-4c0c-98ca-d1614d018232\") " pod="openshift-insights/insights-runtime-extractor-hgpw7" Feb 17 12:49:17.713086 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.713051 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-runtime-extractor-hgpw7" Feb 17 12:49:17.830839 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:17.830807 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-runtime-extractor-hgpw7"] Feb 17 12:49:17.834193 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:49:17.834162 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72276741_8f08_4c0c_98ca_d1614d018232.slice/crio-5a00f0fdd80bc6b6f53fd987d67963915b1c8d51073b86a613c8e6a64c187ee6 WatchSource:0}: Error finding container 5a00f0fdd80bc6b6f53fd987d67963915b1c8d51073b86a613c8e6a64c187ee6: Status 404 returned error can't find the container with id 5a00f0fdd80bc6b6f53fd987d67963915b1c8d51073b86a613c8e6a64c187ee6 Feb 17 12:49:18.605406 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:18.605377 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-hgpw7" event={"ID":"72276741-8f08-4c0c-98ca-d1614d018232","Type":"ContainerStarted","Data":"118edd2efea98b1549e7c9fa7575d452d36b5f4f613ebb6ec8950f8fbd7291ca"} Feb 17 12:49:18.605788 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:18.605412 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-hgpw7" event={"ID":"72276741-8f08-4c0c-98ca-d1614d018232","Type":"ContainerStarted","Data":"5a00f0fdd80bc6b6f53fd987d67963915b1c8d51073b86a613c8e6a64c187ee6"} Feb 17 12:49:19.609938 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:19.609905 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-hgpw7" event={"ID":"72276741-8f08-4c0c-98ca-d1614d018232","Type":"ContainerStarted","Data":"26e05174f2b3848981a23364342c719bc46d7174c5d9132f9a2e52ad6c1b3219"} Feb 17 12:49:20.613477 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:20.613444 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-runtime-extractor-hgpw7" event={"ID":"72276741-8f08-4c0c-98ca-d1614d018232","Type":"ContainerStarted","Data":"992211ea7a7b9eadd202dd64c6e9c246fd921058969003909887964ab075e02c"} Feb 17 12:49:20.629332 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:20.629289 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-insights/insights-runtime-extractor-hgpw7" podStartSLOduration=1.656684139 podStartE2EDuration="3.629275786s" podCreationTimestamp="2026-02-17 12:49:17 +0000 UTC" firstStartedPulling="2026-02-17 12:49:17.892493777 +0000 UTC m=+164.242904567" lastFinishedPulling="2026-02-17 12:49:19.865085424 +0000 UTC m=+166.215496214" observedRunningTime="2026-02-17 12:49:20.628133738 +0000 UTC m=+166.978544550" watchObservedRunningTime="2026-02-17 12:49:20.629275786 +0000 UTC m=+166.979686598" Feb 17 12:49:22.126268 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:22.126235 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:49:25.125783 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:25.125725 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-prcdr" Feb 17 12:49:25.127645 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:25.127625 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"default-dockercfg-9qj8f\"" Feb 17 12:49:25.136860 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:25.136847 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-prcdr" Feb 17 12:49:25.260407 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:25.260375 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-prcdr"] Feb 17 12:49:25.263566 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:49:25.263540 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51d19194_106c_4b3f_ba37_a0a7eedbc75a.slice/crio-717cc9506448bb0bf05d2cc63b3ba13ecd43ba5053c692217df4632bfbb260e6 WatchSource:0}: Error finding container 717cc9506448bb0bf05d2cc63b3ba13ecd43ba5053c692217df4632bfbb260e6: Status 404 returned error can't find the container with id 717cc9506448bb0bf05d2cc63b3ba13ecd43ba5053c692217df4632bfbb260e6 Feb 17 12:49:25.488077 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:25.488046 2572 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-7b7b594589-sk6tq"] Feb 17 12:49:25.492440 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:25.492424 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-7b7b594589-sk6tq" Feb 17 12:49:25.494490 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:25.494465 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-operator-admission-webhook-tls\"" Feb 17 12:49:25.494571 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:25.494466 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-operator-admission-webhook-dockercfg-mpr9j\"" Feb 17 12:49:25.497881 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:25.497861 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-7b7b594589-sk6tq"] Feb 17 12:49:25.628120 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:25.628067 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-prcdr" event={"ID":"51d19194-106c-4b3f-ba37-a0a7eedbc75a","Type":"ContainerStarted","Data":"717cc9506448bb0bf05d2cc63b3ba13ecd43ba5053c692217df4632bfbb260e6"} Feb 17 12:49:25.684451 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:25.684413 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/37490693-a051-4312-8ffc-772adf00e51b-tls-certificates\") pod \"prometheus-operator-admission-webhook-7b7b594589-sk6tq\" (UID: \"37490693-a051-4312-8ffc-772adf00e51b\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-7b7b594589-sk6tq" Feb 17 12:49:25.785489 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:25.785403 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/37490693-a051-4312-8ffc-772adf00e51b-tls-certificates\") pod \"prometheus-operator-admission-webhook-7b7b594589-sk6tq\" (UID: \"37490693-a051-4312-8ffc-772adf00e51b\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-7b7b594589-sk6tq" Feb 17 12:49:25.787912 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:25.787888 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/37490693-a051-4312-8ffc-772adf00e51b-tls-certificates\") pod \"prometheus-operator-admission-webhook-7b7b594589-sk6tq\" (UID: \"37490693-a051-4312-8ffc-772adf00e51b\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-7b7b594589-sk6tq" Feb 17 12:49:25.801756 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:25.801719 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-7b7b594589-sk6tq" Feb 17 12:49:25.956654 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:25.956621 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-7b7b594589-sk6tq"] Feb 17 12:49:25.961782 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:49:25.961751 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37490693_a051_4312_8ffc_772adf00e51b.slice/crio-30f8c7e6367846d636e00855c89266a9abb24a95c2ff79166ad6fd11ee498e5c WatchSource:0}: Error finding container 30f8c7e6367846d636e00855c89266a9abb24a95c2ff79166ad6fd11ee498e5c: Status 404 returned error can't find the container with id 30f8c7e6367846d636e00855c89266a9abb24a95c2ff79166ad6fd11ee498e5c Feb 17 12:49:26.632136 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:26.632087 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-7b7b594589-sk6tq" event={"ID":"37490693-a051-4312-8ffc-772adf00e51b","Type":"ContainerStarted","Data":"30f8c7e6367846d636e00855c89266a9abb24a95c2ff79166ad6fd11ee498e5c"} Feb 17 12:49:27.428005 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:27.427965 2572 patch_prober.go:28] interesting pod/image-registry-68598f8b47-r5pmz container/registry namespace/openshift-image-registry: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body={"errors":[{"code":"UNAVAILABLE","message":"service unavailable","detail":"health check failed: please see /debug/health"}]} Feb 17 12:49:27.428118 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:27.428024 2572 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" podUID="756003d3-7d4c-4fec-b738-1d8d45bf048c" containerName="registry" probeResult="failure" output="HTTP probe failed with statuscode: 503" Feb 17 12:49:27.607827 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:27.607797 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-b6hth" Feb 17 12:49:27.640271 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:27.640223 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-prcdr" event={"ID":"51d19194-106c-4b3f-ba37-a0a7eedbc75a","Type":"ContainerStarted","Data":"b5d3e53602a4ac59d9bd09cbee036458b00e3fcd0efc486d6e7bb3d880d26209"} Feb 17 12:49:27.642214 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:27.642183 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-7b7b594589-sk6tq" event={"ID":"37490693-a051-4312-8ffc-772adf00e51b","Type":"ContainerStarted","Data":"2ab06af733d4baf5f63a6a4b69c4c90b817d7565b379d24f2344ceffe619e9c7"} Feb 17 12:49:27.642452 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:27.642433 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-7b7b594589-sk6tq" Feb 17 12:49:27.648082 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:27.648056 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-7b7b594589-sk6tq" Feb 17 12:49:27.653378 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:27.653331 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-prcdr" podStartSLOduration=139.155578523 podStartE2EDuration="2m20.653319298s" podCreationTimestamp="2026-02-17 12:47:07 +0000 UTC" firstStartedPulling="2026-02-17 12:49:25.265503718 +0000 UTC m=+171.615914508" lastFinishedPulling="2026-02-17 12:49:26.76324449 +0000 UTC m=+173.113655283" observedRunningTime="2026-02-17 12:49:27.653269377 +0000 UTC m=+174.003680201" watchObservedRunningTime="2026-02-17 12:49:27.653319298 +0000 UTC m=+174.003730111" Feb 17 12:49:27.665832 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:27.665758 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-admission-webhook-7b7b594589-sk6tq" podStartSLOduration=1.220346356 podStartE2EDuration="2.665745423s" podCreationTimestamp="2026-02-17 12:49:25 +0000 UTC" firstStartedPulling="2026-02-17 12:49:25.964165348 +0000 UTC m=+172.314576137" lastFinishedPulling="2026-02-17 12:49:27.409564411 +0000 UTC m=+173.759975204" observedRunningTime="2026-02-17 12:49:27.665005946 +0000 UTC m=+174.015416758" watchObservedRunningTime="2026-02-17 12:49:27.665745423 +0000 UTC m=+174.016156235" Feb 17 12:49:28.549890 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:28.549861 2572 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-6567ccfb9-dxxms"] Feb 17 12:49:28.552986 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:28.552970 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-6567ccfb9-dxxms" Feb 17 12:49:28.557194 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:28.555156 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"metrics-client-ca\"" Feb 17 12:49:28.557194 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:28.555403 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-operator-tls\"" Feb 17 12:49:28.557194 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:28.555645 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"kube-root-ca.crt\"" Feb 17 12:49:28.557194 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:28.555847 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-operator-dockercfg-cf4js\"" Feb 17 12:49:28.557194 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:28.556074 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"openshift-service-ca.crt\"" Feb 17 12:49:28.557194 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:28.556316 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-operator-kube-rbac-proxy-config\"" Feb 17 12:49:28.560549 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:28.560497 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-6567ccfb9-dxxms"] Feb 17 12:49:28.605415 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:28.605376 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/e928aa3b-de14-4d54-aad7-4f8bb32ef910-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-6567ccfb9-dxxms\" (UID: \"e928aa3b-de14-4d54-aad7-4f8bb32ef910\") " pod="openshift-monitoring/prometheus-operator-6567ccfb9-dxxms" Feb 17 12:49:28.605588 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:28.605452 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6k2d\" (UniqueName: \"kubernetes.io/projected/e928aa3b-de14-4d54-aad7-4f8bb32ef910-kube-api-access-x6k2d\") pod \"prometheus-operator-6567ccfb9-dxxms\" (UID: \"e928aa3b-de14-4d54-aad7-4f8bb32ef910\") " pod="openshift-monitoring/prometheus-operator-6567ccfb9-dxxms" Feb 17 12:49:28.605588 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:28.605475 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/e928aa3b-de14-4d54-aad7-4f8bb32ef910-metrics-client-ca\") pod \"prometheus-operator-6567ccfb9-dxxms\" (UID: \"e928aa3b-de14-4d54-aad7-4f8bb32ef910\") " pod="openshift-monitoring/prometheus-operator-6567ccfb9-dxxms" Feb 17 12:49:28.605588 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:28.605501 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/e928aa3b-de14-4d54-aad7-4f8bb32ef910-prometheus-operator-tls\") pod \"prometheus-operator-6567ccfb9-dxxms\" (UID: \"e928aa3b-de14-4d54-aad7-4f8bb32ef910\") " pod="openshift-monitoring/prometheus-operator-6567ccfb9-dxxms" Feb 17 12:49:28.706228 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:28.706187 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/e928aa3b-de14-4d54-aad7-4f8bb32ef910-metrics-client-ca\") pod \"prometheus-operator-6567ccfb9-dxxms\" (UID: \"e928aa3b-de14-4d54-aad7-4f8bb32ef910\") " pod="openshift-monitoring/prometheus-operator-6567ccfb9-dxxms" Feb 17 12:49:28.706674 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:28.706251 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/e928aa3b-de14-4d54-aad7-4f8bb32ef910-prometheus-operator-tls\") pod \"prometheus-operator-6567ccfb9-dxxms\" (UID: \"e928aa3b-de14-4d54-aad7-4f8bb32ef910\") " pod="openshift-monitoring/prometheus-operator-6567ccfb9-dxxms" Feb 17 12:49:28.706674 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:28.706288 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/e928aa3b-de14-4d54-aad7-4f8bb32ef910-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-6567ccfb9-dxxms\" (UID: \"e928aa3b-de14-4d54-aad7-4f8bb32ef910\") " pod="openshift-monitoring/prometheus-operator-6567ccfb9-dxxms" Feb 17 12:49:28.706674 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:28.706647 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-x6k2d\" (UniqueName: \"kubernetes.io/projected/e928aa3b-de14-4d54-aad7-4f8bb32ef910-kube-api-access-x6k2d\") pod \"prometheus-operator-6567ccfb9-dxxms\" (UID: \"e928aa3b-de14-4d54-aad7-4f8bb32ef910\") " pod="openshift-monitoring/prometheus-operator-6567ccfb9-dxxms" Feb 17 12:49:28.707006 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:28.706979 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/e928aa3b-de14-4d54-aad7-4f8bb32ef910-metrics-client-ca\") pod \"prometheus-operator-6567ccfb9-dxxms\" (UID: \"e928aa3b-de14-4d54-aad7-4f8bb32ef910\") " pod="openshift-monitoring/prometheus-operator-6567ccfb9-dxxms" Feb 17 12:49:28.708811 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:28.708788 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/e928aa3b-de14-4d54-aad7-4f8bb32ef910-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-6567ccfb9-dxxms\" (UID: \"e928aa3b-de14-4d54-aad7-4f8bb32ef910\") " pod="openshift-monitoring/prometheus-operator-6567ccfb9-dxxms" Feb 17 12:49:28.708811 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:28.708802 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/e928aa3b-de14-4d54-aad7-4f8bb32ef910-prometheus-operator-tls\") pod \"prometheus-operator-6567ccfb9-dxxms\" (UID: \"e928aa3b-de14-4d54-aad7-4f8bb32ef910\") " pod="openshift-monitoring/prometheus-operator-6567ccfb9-dxxms" Feb 17 12:49:28.715197 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:28.715179 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6k2d\" (UniqueName: \"kubernetes.io/projected/e928aa3b-de14-4d54-aad7-4f8bb32ef910-kube-api-access-x6k2d\") pod \"prometheus-operator-6567ccfb9-dxxms\" (UID: \"e928aa3b-de14-4d54-aad7-4f8bb32ef910\") " pod="openshift-monitoring/prometheus-operator-6567ccfb9-dxxms" Feb 17 12:49:28.866074 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:28.865978 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-6567ccfb9-dxxms" Feb 17 12:49:28.980213 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:28.980177 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-6567ccfb9-dxxms"] Feb 17 12:49:28.984095 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:49:28.984071 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode928aa3b_de14_4d54_aad7_4f8bb32ef910.slice/crio-53825d1bf94df8de3dd4a272b7c582c7eabe1deb9403fe2d8d80c01245c164d7 WatchSource:0}: Error finding container 53825d1bf94df8de3dd4a272b7c582c7eabe1deb9403fe2d8d80c01245c164d7: Status 404 returned error can't find the container with id 53825d1bf94df8de3dd4a272b7c582c7eabe1deb9403fe2d8d80c01245c164d7 Feb 17 12:49:29.647912 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:29.647871 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-6567ccfb9-dxxms" event={"ID":"e928aa3b-de14-4d54-aad7-4f8bb32ef910","Type":"ContainerStarted","Data":"53825d1bf94df8de3dd4a272b7c582c7eabe1deb9403fe2d8d80c01245c164d7"} Feb 17 12:49:30.651672 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:30.651641 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-6567ccfb9-dxxms" event={"ID":"e928aa3b-de14-4d54-aad7-4f8bb32ef910","Type":"ContainerStarted","Data":"8276abbdf510a940562eb4337408698dfa7da10159b5ab322375282808e8c2a0"} Feb 17 12:49:30.651672 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:30.651674 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-6567ccfb9-dxxms" event={"ID":"e928aa3b-de14-4d54-aad7-4f8bb32ef910","Type":"ContainerStarted","Data":"4a2b6452209ca199135c4bf907da286cb64b6fc2e6ad003d34a9dc658f4bdb17"} Feb 17 12:49:30.665639 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:30.665587 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-6567ccfb9-dxxms" podStartSLOduration=1.474819281 podStartE2EDuration="2.665565309s" podCreationTimestamp="2026-02-17 12:49:28 +0000 UTC" firstStartedPulling="2026-02-17 12:49:28.985825603 +0000 UTC m=+175.336236393" lastFinishedPulling="2026-02-17 12:49:30.176571628 +0000 UTC m=+176.526982421" observedRunningTime="2026-02-17 12:49:30.665071908 +0000 UTC m=+177.015482720" watchObservedRunningTime="2026-02-17 12:49:30.665565309 +0000 UTC m=+177.015976122" Feb 17 12:49:32.902268 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:32.902231 2572 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/node-exporter-6r5m2"] Feb 17 12:49:32.905694 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:32.905667 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:32.907516 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:32.907499 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-tls\"" Feb 17 12:49:32.908145 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:32.908098 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-kube-rbac-proxy-config\"" Feb 17 12:49:32.908355 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:32.908340 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"node-exporter-dockercfg-4ggjd\"" Feb 17 12:49:32.908510 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:32.908492 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"node-exporter-accelerators-collector-config\"" Feb 17 12:49:32.934952 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:32.934930 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c32ad535-2861-4d6f-83a7-8852563190a5-sys\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:32.935062 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:32.934963 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4rvl\" (UniqueName: \"kubernetes.io/projected/c32ad535-2861-4d6f-83a7-8852563190a5-kube-api-access-z4rvl\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:32.935062 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:32.934987 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/c32ad535-2861-4d6f-83a7-8852563190a5-node-exporter-wtmp\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:32.935062 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:32.935006 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/c32ad535-2861-4d6f-83a7-8852563190a5-node-exporter-textfile\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:32.935062 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:32.935042 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/c32ad535-2861-4d6f-83a7-8852563190a5-node-exporter-tls\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:32.935224 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:32.935121 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/c32ad535-2861-4d6f-83a7-8852563190a5-metrics-client-ca\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:32.935224 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:32.935152 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/c32ad535-2861-4d6f-83a7-8852563190a5-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:32.935224 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:32.935184 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/c32ad535-2861-4d6f-83a7-8852563190a5-root\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:32.935224 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:32.935214 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/c32ad535-2861-4d6f-83a7-8852563190a5-node-exporter-accelerators-collector-config\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:33.035537 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:33.035502 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-z4rvl\" (UniqueName: \"kubernetes.io/projected/c32ad535-2861-4d6f-83a7-8852563190a5-kube-api-access-z4rvl\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:33.035537 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:33.035540 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/c32ad535-2861-4d6f-83a7-8852563190a5-node-exporter-wtmp\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:33.035769 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:33.035560 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/c32ad535-2861-4d6f-83a7-8852563190a5-node-exporter-textfile\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:33.035769 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:33.035585 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/c32ad535-2861-4d6f-83a7-8852563190a5-node-exporter-tls\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:33.035769 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:33.035613 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/c32ad535-2861-4d6f-83a7-8852563190a5-metrics-client-ca\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:33.035769 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:33.035646 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/c32ad535-2861-4d6f-83a7-8852563190a5-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:33.035769 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:33.035693 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/c32ad535-2861-4d6f-83a7-8852563190a5-root\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:33.035769 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:49:33.035696 2572 secret.go:189] Couldn't get secret openshift-monitoring/node-exporter-tls: secret "node-exporter-tls" not found Feb 17 12:49:33.035769 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:33.035739 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/c32ad535-2861-4d6f-83a7-8852563190a5-node-exporter-accelerators-collector-config\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:33.035769 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:49:33.035764 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c32ad535-2861-4d6f-83a7-8852563190a5-node-exporter-tls podName:c32ad535-2861-4d6f-83a7-8852563190a5 nodeName:}" failed. No retries permitted until 2026-02-17 12:49:33.535744993 +0000 UTC m=+179.886155787 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-exporter-tls" (UniqueName: "kubernetes.io/secret/c32ad535-2861-4d6f-83a7-8852563190a5-node-exporter-tls") pod "node-exporter-6r5m2" (UID: "c32ad535-2861-4d6f-83a7-8852563190a5") : secret "node-exporter-tls" not found Feb 17 12:49:33.036181 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:33.035797 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/c32ad535-2861-4d6f-83a7-8852563190a5-root\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:33.036181 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:33.035821 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c32ad535-2861-4d6f-83a7-8852563190a5-sys\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:33.036181 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:33.035704 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/c32ad535-2861-4d6f-83a7-8852563190a5-node-exporter-wtmp\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:33.036181 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:33.035885 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c32ad535-2861-4d6f-83a7-8852563190a5-sys\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:33.036181 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:33.035950 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/c32ad535-2861-4d6f-83a7-8852563190a5-node-exporter-textfile\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:33.036427 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:33.036373 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-accelerators-collector-config\" (UniqueName: \"kubernetes.io/configmap/c32ad535-2861-4d6f-83a7-8852563190a5-node-exporter-accelerators-collector-config\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:33.036427 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:33.036413 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/c32ad535-2861-4d6f-83a7-8852563190a5-metrics-client-ca\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:33.037939 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:33.037918 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/c32ad535-2861-4d6f-83a7-8852563190a5-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:33.042931 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:33.042912 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4rvl\" (UniqueName: \"kubernetes.io/projected/c32ad535-2861-4d6f-83a7-8852563190a5-kube-api-access-z4rvl\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:33.539389 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:33.539337 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/c32ad535-2861-4d6f-83a7-8852563190a5-node-exporter-tls\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:33.541645 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:33.541627 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/c32ad535-2861-4d6f-83a7-8852563190a5-node-exporter-tls\") pod \"node-exporter-6r5m2\" (UID: \"c32ad535-2861-4d6f-83a7-8852563190a5\") " pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:33.814626 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:33.814542 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-6r5m2" Feb 17 12:49:33.822954 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:49:33.822926 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc32ad535_2861_4d6f_83a7_8852563190a5.slice/crio-0d99bb2e19b80fd64c931d46bbc8ff01295452672ea8aa188193beff57eaef25 WatchSource:0}: Error finding container 0d99bb2e19b80fd64c931d46bbc8ff01295452672ea8aa188193beff57eaef25: Status 404 returned error can't find the container with id 0d99bb2e19b80fd64c931d46bbc8ff01295452672ea8aa188193beff57eaef25 Feb 17 12:49:33.998559 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:33.998527 2572 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Feb 17 12:49:34.006503 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.006474 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.010139 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.009956 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy\"" Feb 17 12:49:34.010139 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.009978 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-tls\"" Feb 17 12:49:34.010139 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.010015 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-cluster-tls-config\"" Feb 17 12:49:34.010139 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.009956 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"alertmanager-trusted-ca-bundle\"" Feb 17 12:49:34.010139 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.009956 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-web-config\"" Feb 17 12:49:34.010447 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.010210 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-generated\"" Feb 17 12:49:34.010447 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.010244 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy-web\"" Feb 17 12:49:34.010447 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.010259 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-tls-assets-0\"" Feb 17 12:49:34.010447 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.010398 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-dockercfg-6qj8b\"" Feb 17 12:49:34.010740 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.010723 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy-metric\"" Feb 17 12:49:34.017299 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.017277 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Feb 17 12:49:34.044277 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.044243 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-config-volume\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.044429 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.044287 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.044429 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.044318 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/35751e92-fed8-4a26-931d-1e9070e8181b-config-out\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.044429 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.044348 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/35751e92-fed8-4a26-931d-1e9070e8181b-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.044429 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.044385 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-web-config\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.044429 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.044426 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-cluster-tls-config\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.044664 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.044451 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/35751e92-fed8-4a26-931d-1e9070e8181b-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.044664 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.044500 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kgrv\" (UniqueName: \"kubernetes.io/projected/35751e92-fed8-4a26-931d-1e9070e8181b-kube-api-access-8kgrv\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.044664 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.044527 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.044664 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.044551 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/35751e92-fed8-4a26-931d-1e9070e8181b-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.044664 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.044613 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.044664 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.044641 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/35751e92-fed8-4a26-931d-1e9070e8181b-tls-assets\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.044864 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.044670 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.145241 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.145146 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-config-volume\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.145241 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.145206 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.145449 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.145241 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/35751e92-fed8-4a26-931d-1e9070e8181b-config-out\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.145449 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.145274 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/35751e92-fed8-4a26-931d-1e9070e8181b-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.145546 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.145446 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-web-config\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.145593 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.145567 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-cluster-tls-config\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.145647 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.145610 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/35751e92-fed8-4a26-931d-1e9070e8181b-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.145696 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.145645 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8kgrv\" (UniqueName: \"kubernetes.io/projected/35751e92-fed8-4a26-931d-1e9070e8181b-kube-api-access-8kgrv\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.145696 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.145681 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.145790 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.145708 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/35751e92-fed8-4a26-931d-1e9070e8181b-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.145790 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.145745 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.145790 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.145782 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/35751e92-fed8-4a26-931d-1e9070e8181b-tls-assets\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.145923 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.145828 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.145997 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.145778 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/35751e92-fed8-4a26-931d-1e9070e8181b-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.146929 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.146902 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/35751e92-fed8-4a26-931d-1e9070e8181b-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.147005 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.146942 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-generated\"" Feb 17 12:49:34.147005 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.146988 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy-web\"" Feb 17 12:49:34.147256 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.147234 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-web-config\"" Feb 17 12:49:34.147256 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.147234 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"alertmanager-trusted-ca-bundle\"" Feb 17 12:49:34.147545 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.147311 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-cluster-tls-config\"" Feb 17 12:49:34.147613 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.147545 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-tls-assets-0\"" Feb 17 12:49:34.147688 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.147620 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-tls\"" Feb 17 12:49:34.147688 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.147679 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy-metric\"" Feb 17 12:49:34.147790 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.147751 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy\"" Feb 17 12:49:34.147959 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.147939 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/35751e92-fed8-4a26-931d-1e9070e8181b-config-out\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.152895 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.152874 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kgrv\" (UniqueName: \"kubernetes.io/projected/35751e92-fed8-4a26-931d-1e9070e8181b-kube-api-access-8kgrv\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.156004 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:49:34.155976 2572 secret.go:189] Couldn't get secret openshift-monitoring/alertmanager-main-tls: secret "alertmanager-main-tls" not found Feb 17 12:49:34.156432 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:49:34.156248 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/35751e92-fed8-4a26-931d-1e9070e8181b-alertmanager-trusted-ca-bundle podName:35751e92-fed8-4a26-931d-1e9070e8181b nodeName:}" failed. No retries permitted until 2026-02-17 12:49:34.656223745 +0000 UTC m=+181.006634549 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "alertmanager-trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/35751e92-fed8-4a26-931d-1e9070e8181b-alertmanager-trusted-ca-bundle") pod "alertmanager-main-0" (UID: "35751e92-fed8-4a26-931d-1e9070e8181b") : configmap references non-existent config key: ca-bundle.crt Feb 17 12:49:34.156538 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:49:34.156460 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-secret-alertmanager-main-tls podName:35751e92-fed8-4a26-931d-1e9070e8181b nodeName:}" failed. No retries permitted until 2026-02-17 12:49:34.656430143 +0000 UTC m=+181.006840934 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-alertmanager-main-tls" (UniqueName: "kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-secret-alertmanager-main-tls") pod "alertmanager-main-0" (UID: "35751e92-fed8-4a26-931d-1e9070e8181b") : secret "alertmanager-main-tls" not found Feb 17 12:49:34.158038 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.158008 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.158741 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.158720 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-config-volume\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.158872 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.158846 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/35751e92-fed8-4a26-931d-1e9070e8181b-tls-assets\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.159000 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.158981 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-cluster-tls-config\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.159041 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.158983 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-web-config\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.159117 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.159087 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.160584 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.160563 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.663397 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.663370 2572 generic.go:358] "Generic (PLEG): container finished" podID="c32ad535-2861-4d6f-83a7-8852563190a5" containerID="c50d4294cdc9bf8637f4cda08a9157cf75500ca3b35f56e9cb73a7d5aab3e874" exitCode=0 Feb 17 12:49:34.663518 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.663440 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-6r5m2" event={"ID":"c32ad535-2861-4d6f-83a7-8852563190a5","Type":"ContainerDied","Data":"c50d4294cdc9bf8637f4cda08a9157cf75500ca3b35f56e9cb73a7d5aab3e874"} Feb 17 12:49:34.663518 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.663478 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-6r5m2" event={"ID":"c32ad535-2861-4d6f-83a7-8852563190a5","Type":"ContainerStarted","Data":"0d99bb2e19b80fd64c931d46bbc8ff01295452672ea8aa188193beff57eaef25"} Feb 17 12:49:34.750236 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.750213 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/35751e92-fed8-4a26-931d-1e9070e8181b-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.750361 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.750294 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.750987 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.750965 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/35751e92-fed8-4a26-931d-1e9070e8181b-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.752633 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.752610 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:34.919191 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.919162 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-dockercfg-6qj8b\"" Feb 17 12:49:34.928078 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:34.928050 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:49:35.054427 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:35.054397 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Feb 17 12:49:35.058503 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:49:35.058479 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod35751e92_fed8_4a26_931d_1e9070e8181b.slice/crio-64694402cdb3bdfea6e7850d714e8f6bc96bc653a2fb8ff93ffb8597fee9273e WatchSource:0}: Error finding container 64694402cdb3bdfea6e7850d714e8f6bc96bc653a2fb8ff93ffb8597fee9273e: Status 404 returned error can't find the container with id 64694402cdb3bdfea6e7850d714e8f6bc96bc653a2fb8ff93ffb8597fee9273e Feb 17 12:49:35.668949 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:35.668904 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-6r5m2" event={"ID":"c32ad535-2861-4d6f-83a7-8852563190a5","Type":"ContainerStarted","Data":"3a9a7ae61eabab8c9e9439c2e4eb3d39eb5be4653fb70de33d69cff02de749b0"} Feb 17 12:49:35.668949 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:35.668954 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-6r5m2" event={"ID":"c32ad535-2861-4d6f-83a7-8852563190a5","Type":"ContainerStarted","Data":"759b169c721f6819f5364c1b91153ca795ca9773bce9f7ac3fd5f83f0b192d7e"} Feb 17 12:49:35.670170 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:35.670143 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"35751e92-fed8-4a26-931d-1e9070e8181b","Type":"ContainerStarted","Data":"64694402cdb3bdfea6e7850d714e8f6bc96bc653a2fb8ff93ffb8597fee9273e"} Feb 17 12:49:35.691940 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:35.691887 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/node-exporter-6r5m2" podStartSLOduration=3.051083853 podStartE2EDuration="3.691871262s" podCreationTimestamp="2026-02-17 12:49:32 +0000 UTC" firstStartedPulling="2026-02-17 12:49:33.824644807 +0000 UTC m=+180.175055609" lastFinishedPulling="2026-02-17 12:49:34.465432224 +0000 UTC m=+180.815843018" observedRunningTime="2026-02-17 12:49:35.689580314 +0000 UTC m=+182.039991137" watchObservedRunningTime="2026-02-17 12:49:35.691871262 +0000 UTC m=+182.042282076" Feb 17 12:49:36.673643 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:36.673612 2572 generic.go:358] "Generic (PLEG): container finished" podID="35751e92-fed8-4a26-931d-1e9070e8181b" containerID="6f5bb117d629c48ad1ffcb2dc8b5c7fc16e173a1673d071f0ec52f4b482e6b44" exitCode=0 Feb 17 12:49:36.674064 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:36.673675 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"35751e92-fed8-4a26-931d-1e9070e8181b","Type":"ContainerDied","Data":"6f5bb117d629c48ad1ffcb2dc8b5c7fc16e173a1673d071f0ec52f4b482e6b44"} Feb 17 12:49:37.393360 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.393326 2572 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/metrics-server-86f7684bf5-m2r9q"] Feb 17 12:49:37.396744 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.396717 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" Feb 17 12:49:37.398659 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.398636 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-server-client-certs\"" Feb 17 12:49:37.399265 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.399244 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-server-tls\"" Feb 17 12:49:37.399395 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.399297 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"metrics-server-audit-profiles\"" Feb 17 12:49:37.399395 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.399344 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"kubelet-serving-ca-bundle\"" Feb 17 12:49:37.399395 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.399365 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-server-dockercfg-7vxcc\"" Feb 17 12:49:37.399395 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.399247 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-server-90ec7voomrm0f\"" Feb 17 12:49:37.408222 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.408199 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-86f7684bf5-m2r9q"] Feb 17 12:49:37.427637 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.427613 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:49:37.473521 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.473490 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/2766fd9f-4249-4473-8285-5dcd03d2b37c-metrics-server-audit-profiles\") pod \"metrics-server-86f7684bf5-m2r9q\" (UID: \"2766fd9f-4249-4473-8285-5dcd03d2b37c\") " pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" Feb 17 12:49:37.473667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.473542 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2766fd9f-4249-4473-8285-5dcd03d2b37c-client-ca-bundle\") pod \"metrics-server-86f7684bf5-m2r9q\" (UID: \"2766fd9f-4249-4473-8285-5dcd03d2b37c\") " pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" Feb 17 12:49:37.473667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.473563 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2766fd9f-4249-4473-8285-5dcd03d2b37c-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-86f7684bf5-m2r9q\" (UID: \"2766fd9f-4249-4473-8285-5dcd03d2b37c\") " pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" Feb 17 12:49:37.473667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.473606 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-server-client-certs\" (UniqueName: \"kubernetes.io/secret/2766fd9f-4249-4473-8285-5dcd03d2b37c-secret-metrics-server-client-certs\") pod \"metrics-server-86f7684bf5-m2r9q\" (UID: \"2766fd9f-4249-4473-8285-5dcd03d2b37c\") " pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" Feb 17 12:49:37.473667 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.473625 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/2766fd9f-4249-4473-8285-5dcd03d2b37c-secret-metrics-server-tls\") pod \"metrics-server-86f7684bf5-m2r9q\" (UID: \"2766fd9f-4249-4473-8285-5dcd03d2b37c\") " pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" Feb 17 12:49:37.473829 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.473669 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r5vb\" (UniqueName: \"kubernetes.io/projected/2766fd9f-4249-4473-8285-5dcd03d2b37c-kube-api-access-7r5vb\") pod \"metrics-server-86f7684bf5-m2r9q\" (UID: \"2766fd9f-4249-4473-8285-5dcd03d2b37c\") " pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" Feb 17 12:49:37.473829 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.473685 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/2766fd9f-4249-4473-8285-5dcd03d2b37c-audit-log\") pod \"metrics-server-86f7684bf5-m2r9q\" (UID: \"2766fd9f-4249-4473-8285-5dcd03d2b37c\") " pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" Feb 17 12:49:37.574770 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.574739 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2766fd9f-4249-4473-8285-5dcd03d2b37c-client-ca-bundle\") pod \"metrics-server-86f7684bf5-m2r9q\" (UID: \"2766fd9f-4249-4473-8285-5dcd03d2b37c\") " pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" Feb 17 12:49:37.574770 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.574785 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2766fd9f-4249-4473-8285-5dcd03d2b37c-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-86f7684bf5-m2r9q\" (UID: \"2766fd9f-4249-4473-8285-5dcd03d2b37c\") " pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" Feb 17 12:49:37.575006 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.574845 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-metrics-server-client-certs\" (UniqueName: \"kubernetes.io/secret/2766fd9f-4249-4473-8285-5dcd03d2b37c-secret-metrics-server-client-certs\") pod \"metrics-server-86f7684bf5-m2r9q\" (UID: \"2766fd9f-4249-4473-8285-5dcd03d2b37c\") " pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" Feb 17 12:49:37.575006 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.574881 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/2766fd9f-4249-4473-8285-5dcd03d2b37c-secret-metrics-server-tls\") pod \"metrics-server-86f7684bf5-m2r9q\" (UID: \"2766fd9f-4249-4473-8285-5dcd03d2b37c\") " pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" Feb 17 12:49:37.575006 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.574932 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-7r5vb\" (UniqueName: \"kubernetes.io/projected/2766fd9f-4249-4473-8285-5dcd03d2b37c-kube-api-access-7r5vb\") pod \"metrics-server-86f7684bf5-m2r9q\" (UID: \"2766fd9f-4249-4473-8285-5dcd03d2b37c\") " pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" Feb 17 12:49:37.575179 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.575068 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/2766fd9f-4249-4473-8285-5dcd03d2b37c-audit-log\") pod \"metrics-server-86f7684bf5-m2r9q\" (UID: \"2766fd9f-4249-4473-8285-5dcd03d2b37c\") " pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" Feb 17 12:49:37.575179 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.575136 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/2766fd9f-4249-4473-8285-5dcd03d2b37c-metrics-server-audit-profiles\") pod \"metrics-server-86f7684bf5-m2r9q\" (UID: \"2766fd9f-4249-4473-8285-5dcd03d2b37c\") " pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" Feb 17 12:49:37.575454 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.575429 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/2766fd9f-4249-4473-8285-5dcd03d2b37c-audit-log\") pod \"metrics-server-86f7684bf5-m2r9q\" (UID: \"2766fd9f-4249-4473-8285-5dcd03d2b37c\") " pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" Feb 17 12:49:37.575688 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.575666 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2766fd9f-4249-4473-8285-5dcd03d2b37c-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-86f7684bf5-m2r9q\" (UID: \"2766fd9f-4249-4473-8285-5dcd03d2b37c\") " pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" Feb 17 12:49:37.576172 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.576149 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/2766fd9f-4249-4473-8285-5dcd03d2b37c-metrics-server-audit-profiles\") pod \"metrics-server-86f7684bf5-m2r9q\" (UID: \"2766fd9f-4249-4473-8285-5dcd03d2b37c\") " pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" Feb 17 12:49:37.578165 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.578142 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/2766fd9f-4249-4473-8285-5dcd03d2b37c-secret-metrics-server-tls\") pod \"metrics-server-86f7684bf5-m2r9q\" (UID: \"2766fd9f-4249-4473-8285-5dcd03d2b37c\") " pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" Feb 17 12:49:37.578306 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.578288 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-metrics-server-client-certs\" (UniqueName: \"kubernetes.io/secret/2766fd9f-4249-4473-8285-5dcd03d2b37c-secret-metrics-server-client-certs\") pod \"metrics-server-86f7684bf5-m2r9q\" (UID: \"2766fd9f-4249-4473-8285-5dcd03d2b37c\") " pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" Feb 17 12:49:37.578431 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.578406 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2766fd9f-4249-4473-8285-5dcd03d2b37c-client-ca-bundle\") pod \"metrics-server-86f7684bf5-m2r9q\" (UID: \"2766fd9f-4249-4473-8285-5dcd03d2b37c\") " pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" Feb 17 12:49:37.583479 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.583459 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r5vb\" (UniqueName: \"kubernetes.io/projected/2766fd9f-4249-4473-8285-5dcd03d2b37c-kube-api-access-7r5vb\") pod \"metrics-server-86f7684bf5-m2r9q\" (UID: \"2766fd9f-4249-4473-8285-5dcd03d2b37c\") " pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" Feb 17 12:49:37.678072 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.677987 2572 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/monitoring-plugin-69574b87bc-x78dp"] Feb 17 12:49:37.685695 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.685668 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-69574b87bc-x78dp" Feb 17 12:49:37.687711 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.687686 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"default-dockercfg-v28g8\"" Feb 17 12:49:37.687842 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.687686 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"monitoring-plugin-cert\"" Feb 17 12:49:37.688756 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.688732 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-69574b87bc-x78dp"] Feb 17 12:49:37.709572 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.709533 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" Feb 17 12:49:37.776406 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.776322 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/b0db52e3-5296-4f94-95e3-b3867884108a-monitoring-plugin-cert\") pod \"monitoring-plugin-69574b87bc-x78dp\" (UID: \"b0db52e3-5296-4f94-95e3-b3867884108a\") " pod="openshift-monitoring/monitoring-plugin-69574b87bc-x78dp" Feb 17 12:49:37.831674 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.831647 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-86f7684bf5-m2r9q"] Feb 17 12:49:37.834334 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:49:37.834305 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2766fd9f_4249_4473_8285_5dcd03d2b37c.slice/crio-c66f9d4404c35922950753af12732f1c123011f4dffd59270673f8eaf056ce55 WatchSource:0}: Error finding container c66f9d4404c35922950753af12732f1c123011f4dffd59270673f8eaf056ce55: Status 404 returned error can't find the container with id c66f9d4404c35922950753af12732f1c123011f4dffd59270673f8eaf056ce55 Feb 17 12:49:37.876962 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:37.876921 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/b0db52e3-5296-4f94-95e3-b3867884108a-monitoring-plugin-cert\") pod \"monitoring-plugin-69574b87bc-x78dp\" (UID: \"b0db52e3-5296-4f94-95e3-b3867884108a\") " pod="openshift-monitoring/monitoring-plugin-69574b87bc-x78dp" Feb 17 12:49:37.877144 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:49:37.877077 2572 secret.go:189] Couldn't get secret openshift-monitoring/monitoring-plugin-cert: secret "monitoring-plugin-cert" not found Feb 17 12:49:37.877211 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:49:37.877191 2572 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b0db52e3-5296-4f94-95e3-b3867884108a-monitoring-plugin-cert podName:b0db52e3-5296-4f94-95e3-b3867884108a nodeName:}" failed. No retries permitted until 2026-02-17 12:49:38.377168403 +0000 UTC m=+184.727579196 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "monitoring-plugin-cert" (UniqueName: "kubernetes.io/secret/b0db52e3-5296-4f94-95e3-b3867884108a-monitoring-plugin-cert") pod "monitoring-plugin-69574b87bc-x78dp" (UID: "b0db52e3-5296-4f94-95e3-b3867884108a") : secret "monitoring-plugin-cert" not found Feb 17 12:49:38.382615 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:38.382589 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/b0db52e3-5296-4f94-95e3-b3867884108a-monitoring-plugin-cert\") pod \"monitoring-plugin-69574b87bc-x78dp\" (UID: \"b0db52e3-5296-4f94-95e3-b3867884108a\") " pod="openshift-monitoring/monitoring-plugin-69574b87bc-x78dp" Feb 17 12:49:38.385354 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:38.385327 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/b0db52e3-5296-4f94-95e3-b3867884108a-monitoring-plugin-cert\") pod \"monitoring-plugin-69574b87bc-x78dp\" (UID: \"b0db52e3-5296-4f94-95e3-b3867884108a\") " pod="openshift-monitoring/monitoring-plugin-69574b87bc-x78dp" Feb 17 12:49:38.597592 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:38.597516 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-69574b87bc-x78dp" Feb 17 12:49:38.684801 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:38.684714 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"35751e92-fed8-4a26-931d-1e9070e8181b","Type":"ContainerStarted","Data":"98fa9cd1d827c70e3d844432f4b6ddbb6a65e85fe8daf94445a1134a168ac445"} Feb 17 12:49:38.684801 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:38.684752 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"35751e92-fed8-4a26-931d-1e9070e8181b","Type":"ContainerStarted","Data":"fd8b3b87a83fc0ca14a6d30a829101f488a3d8c20593ee20c27bd20a02195c4f"} Feb 17 12:49:38.684801 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:38.684769 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"35751e92-fed8-4a26-931d-1e9070e8181b","Type":"ContainerStarted","Data":"d1c51b0dcc092df1a64c98743fddd1c3a1961955929a955cae250827703f882c"} Feb 17 12:49:38.684801 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:38.684782 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"35751e92-fed8-4a26-931d-1e9070e8181b","Type":"ContainerStarted","Data":"a3b2f792f814616f4f04b46a1e85d624eeff05eb1c1c0100c30721cadf92d6e8"} Feb 17 12:49:38.686468 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:38.686382 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" event={"ID":"2766fd9f-4249-4473-8285-5dcd03d2b37c","Type":"ContainerStarted","Data":"c66f9d4404c35922950753af12732f1c123011f4dffd59270673f8eaf056ce55"} Feb 17 12:49:38.737137 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:38.736711 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-69574b87bc-x78dp"] Feb 17 12:49:38.741561 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:49:38.741529 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb0db52e3_5296_4f94_95e3_b3867884108a.slice/crio-f456162b090a992aeeb2cca0c1f3bcb07b99ff79a02930a3bc01ed83fdfac0d9 WatchSource:0}: Error finding container f456162b090a992aeeb2cca0c1f3bcb07b99ff79a02930a3bc01ed83fdfac0d9: Status 404 returned error can't find the container with id f456162b090a992aeeb2cca0c1f3bcb07b99ff79a02930a3bc01ed83fdfac0d9 Feb 17 12:49:39.115759 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.115704 2572 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Feb 17 12:49:39.121451 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.121423 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.123815 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.123640 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-thanos-sidecar-tls\"" Feb 17 12:49:39.123815 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.123640 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"prometheus-trusted-ca-bundle\"" Feb 17 12:49:39.123815 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.123809 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"serving-certs-ca-bundle\"" Feb 17 12:49:39.124061 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.123957 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-thanos-prometheus-http-client-file\"" Feb 17 12:49:39.124061 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.124019 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"kube-rbac-proxy\"" Feb 17 12:49:39.124389 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.124204 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-client-certs\"" Feb 17 12:49:39.124389 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.124295 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-dockercfg-wdjcs\"" Feb 17 12:49:39.124389 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.124375 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-tls\"" Feb 17 12:49:39.124581 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.124443 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-kube-rbac-proxy-web\"" Feb 17 12:49:39.124581 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.124494 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-grpc-tls-5l1v479posc61\"" Feb 17 12:49:39.124674 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.124588 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s\"" Feb 17 12:49:39.124674 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.124634 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-tls-assets-0\"" Feb 17 12:49:39.124674 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.124643 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-web-config\"" Feb 17 12:49:39.128028 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.128003 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-rulefiles-0\"" Feb 17 12:49:39.130975 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.130935 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Feb 17 12:49:39.192373 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.192262 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.192373 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.192309 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.192373 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.192335 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.192373 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.192360 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.192736 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.192386 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.192736 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.192419 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/5d6ebab3-fa58-444e-8b5e-e9d52e188752-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.192736 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.192448 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5d6ebab3-fa58-444e-8b5e-e9d52e188752-config-out\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.192736 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.192468 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.192736 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.192492 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.192736 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.192518 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.192736 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.192545 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.192736 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.192570 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5d6ebab3-fa58-444e-8b5e-e9d52e188752-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.192736 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.192605 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.192736 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.192642 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-web-config\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.192736 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.192668 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-config\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.192736 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.192690 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stdxg\" (UniqueName: \"kubernetes.io/projected/5d6ebab3-fa58-444e-8b5e-e9d52e188752-kube-api-access-stdxg\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.192736 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.192739 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.193331 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.192773 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.294237 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.294205 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.294410 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.294246 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.294410 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.294264 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.294410 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.294325 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.294410 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.294369 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.294634 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.294419 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/5d6ebab3-fa58-444e-8b5e-e9d52e188752-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.294634 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.294456 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5d6ebab3-fa58-444e-8b5e-e9d52e188752-config-out\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.294634 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.294485 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.294634 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.294508 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.294634 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.294538 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.294634 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.294593 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.294922 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.294656 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5d6ebab3-fa58-444e-8b5e-e9d52e188752-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.294922 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.294699 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.294922 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.294739 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-web-config\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.294922 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.294771 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-config\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.294922 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.294796 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-stdxg\" (UniqueName: \"kubernetes.io/projected/5d6ebab3-fa58-444e-8b5e-e9d52e188752-kube-api-access-stdxg\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.294922 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.294851 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.294922 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.294904 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.295360 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.295229 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.295360 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.295306 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.295488 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.295463 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/5d6ebab3-fa58-444e-8b5e-e9d52e188752-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.296060 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.296028 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.298296 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.298270 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.299142 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.298697 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.299286 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.298826 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.299441 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.298987 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5d6ebab3-fa58-444e-8b5e-e9d52e188752-config-out\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.299441 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.299055 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.299441 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.299057 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.299801 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.299756 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-config\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.300397 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.300356 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-web-config\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.300620 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.300595 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.300764 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.300638 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.301079 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.301058 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.301560 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.301535 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.302081 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.302066 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5d6ebab3-fa58-444e-8b5e-e9d52e188752-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.308148 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.308090 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-stdxg\" (UniqueName: \"kubernetes.io/projected/5d6ebab3-fa58-444e-8b5e-e9d52e188752-kube-api-access-stdxg\") pod \"prometheus-k8s-0\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.438386 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.438356 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:39.627192 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.627158 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Feb 17 12:49:39.641197 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:49:39.641122 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5d6ebab3_fa58_444e_8b5e_e9d52e188752.slice/crio-eb814d2a53086156ca7779bf67372e992ce49afeb4608f64f1d0f216d0d98e1b WatchSource:0}: Error finding container eb814d2a53086156ca7779bf67372e992ce49afeb4608f64f1d0f216d0d98e1b: Status 404 returned error can't find the container with id eb814d2a53086156ca7779bf67372e992ce49afeb4608f64f1d0f216d0d98e1b Feb 17 12:49:39.693190 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.693145 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"35751e92-fed8-4a26-931d-1e9070e8181b","Type":"ContainerStarted","Data":"b9ccfdcb7b936cf256bb6ba8141cc1d830719805be1ac916d15da1dff87968aa"} Feb 17 12:49:39.694918 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.694818 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" event={"ID":"2766fd9f-4249-4473-8285-5dcd03d2b37c","Type":"ContainerStarted","Data":"85fd497b87434d5f2ad3cd14a61c0d7bb25a58c0bd100d256898afa528eae697"} Feb 17 12:49:39.696452 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.696401 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"5d6ebab3-fa58-444e-8b5e-e9d52e188752","Type":"ContainerStarted","Data":"eb814d2a53086156ca7779bf67372e992ce49afeb4608f64f1d0f216d0d98e1b"} Feb 17 12:49:39.698502 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:39.698465 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-69574b87bc-x78dp" event={"ID":"b0db52e3-5296-4f94-95e3-b3867884108a","Type":"ContainerStarted","Data":"f456162b090a992aeeb2cca0c1f3bcb07b99ff79a02930a3bc01ed83fdfac0d9"} Feb 17 12:49:40.704180 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:40.704139 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"35751e92-fed8-4a26-931d-1e9070e8181b","Type":"ContainerStarted","Data":"ce42649490e3041f1dca08fb8b38628f86a6f7c94c0dc8ff19460069ff25a516"} Feb 17 12:49:40.705416 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:40.705392 2572 generic.go:358] "Generic (PLEG): container finished" podID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerID="6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d" exitCode=0 Feb 17 12:49:40.705496 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:40.705460 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"5d6ebab3-fa58-444e-8b5e-e9d52e188752","Type":"ContainerDied","Data":"6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d"} Feb 17 12:49:40.706687 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:40.706665 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-69574b87bc-x78dp" event={"ID":"b0db52e3-5296-4f94-95e3-b3867884108a","Type":"ContainerStarted","Data":"15e84f4d0b8d5aa443cd26fa93371c1d8998869a8c4dc5dbd49a863f70b27592"} Feb 17 12:49:40.706968 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:40.706951 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-monitoring/monitoring-plugin-69574b87bc-x78dp" Feb 17 12:49:40.711981 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:40.711960 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/monitoring-plugin-69574b87bc-x78dp" Feb 17 12:49:40.730212 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:40.730174 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" podStartSLOduration=2.173694198 podStartE2EDuration="3.730161975s" podCreationTimestamp="2026-02-17 12:49:37 +0000 UTC" firstStartedPulling="2026-02-17 12:49:37.836501238 +0000 UTC m=+184.186912041" lastFinishedPulling="2026-02-17 12:49:39.392969015 +0000 UTC m=+185.743379818" observedRunningTime="2026-02-17 12:49:39.712728058 +0000 UTC m=+186.063138869" watchObservedRunningTime="2026-02-17 12:49:40.730161975 +0000 UTC m=+187.080572791" Feb 17 12:49:40.730554 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:40.730521 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/alertmanager-main-0" podStartSLOduration=2.915362062 podStartE2EDuration="7.730515177s" podCreationTimestamp="2026-02-17 12:49:33 +0000 UTC" firstStartedPulling="2026-02-17 12:49:35.06044764 +0000 UTC m=+181.410858434" lastFinishedPulling="2026-02-17 12:49:39.875600743 +0000 UTC m=+186.226011549" observedRunningTime="2026-02-17 12:49:40.728558907 +0000 UTC m=+187.078969729" watchObservedRunningTime="2026-02-17 12:49:40.730515177 +0000 UTC m=+187.080926012" Feb 17 12:49:40.764010 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:40.763969 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/monitoring-plugin-69574b87bc-x78dp" podStartSLOduration=2.219193957 podStartE2EDuration="3.763955094s" podCreationTimestamp="2026-02-17 12:49:37 +0000 UTC" firstStartedPulling="2026-02-17 12:49:38.743980027 +0000 UTC m=+185.094390817" lastFinishedPulling="2026-02-17 12:49:40.288741162 +0000 UTC m=+186.639151954" observedRunningTime="2026-02-17 12:49:40.741077557 +0000 UTC m=+187.091488373" watchObservedRunningTime="2026-02-17 12:49:40.763955094 +0000 UTC m=+187.114365885" Feb 17 12:49:42.443162 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:42.443092 2572 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" podUID="756003d3-7d4c-4fec-b738-1d8d45bf048c" containerName="registry" containerID="cri-o://c4999988293f533ed9aad40d6664284b6887f8904a29bdfa9aa3650002705574" gracePeriod=30 Feb 17 12:49:42.714237 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:42.714151 2572 generic.go:358] "Generic (PLEG): container finished" podID="756003d3-7d4c-4fec-b738-1d8d45bf048c" containerID="c4999988293f533ed9aad40d6664284b6887f8904a29bdfa9aa3650002705574" exitCode=0 Feb 17 12:49:42.714372 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:42.714225 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" event={"ID":"756003d3-7d4c-4fec-b738-1d8d45bf048c","Type":"ContainerDied","Data":"c4999988293f533ed9aad40d6664284b6887f8904a29bdfa9aa3650002705574"} Feb 17 12:49:43.347233 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.347204 2572 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:49:43.434350 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.434319 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/756003d3-7d4c-4fec-b738-1d8d45bf048c-image-registry-private-configuration\") pod \"756003d3-7d4c-4fec-b738-1d8d45bf048c\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " Feb 17 12:49:43.434466 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.434373 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/756003d3-7d4c-4fec-b738-1d8d45bf048c-ca-trust-extracted\") pod \"756003d3-7d4c-4fec-b738-1d8d45bf048c\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " Feb 17 12:49:43.434466 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.434410 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/756003d3-7d4c-4fec-b738-1d8d45bf048c-trusted-ca\") pod \"756003d3-7d4c-4fec-b738-1d8d45bf048c\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " Feb 17 12:49:43.434466 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.434428 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/756003d3-7d4c-4fec-b738-1d8d45bf048c-registry-certificates\") pod \"756003d3-7d4c-4fec-b738-1d8d45bf048c\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " Feb 17 12:49:43.434466 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.434454 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-bound-sa-token\") pod \"756003d3-7d4c-4fec-b738-1d8d45bf048c\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " Feb 17 12:49:43.434680 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.434494 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-registry-tls\") pod \"756003d3-7d4c-4fec-b738-1d8d45bf048c\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " Feb 17 12:49:43.434680 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.434522 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cf9b\" (UniqueName: \"kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-kube-api-access-4cf9b\") pod \"756003d3-7d4c-4fec-b738-1d8d45bf048c\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " Feb 17 12:49:43.434680 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.434544 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/756003d3-7d4c-4fec-b738-1d8d45bf048c-installation-pull-secrets\") pod \"756003d3-7d4c-4fec-b738-1d8d45bf048c\" (UID: \"756003d3-7d4c-4fec-b738-1d8d45bf048c\") " Feb 17 12:49:43.434927 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.434897 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/756003d3-7d4c-4fec-b738-1d8d45bf048c-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "756003d3-7d4c-4fec-b738-1d8d45bf048c" (UID: "756003d3-7d4c-4fec-b738-1d8d45bf048c"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Feb 17 12:49:43.435303 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.435276 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/756003d3-7d4c-4fec-b738-1d8d45bf048c-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "756003d3-7d4c-4fec-b738-1d8d45bf048c" (UID: "756003d3-7d4c-4fec-b738-1d8d45bf048c"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Feb 17 12:49:43.437785 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.437743 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/756003d3-7d4c-4fec-b738-1d8d45bf048c-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "756003d3-7d4c-4fec-b738-1d8d45bf048c" (UID: "756003d3-7d4c-4fec-b738-1d8d45bf048c"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Feb 17 12:49:43.438379 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.438340 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/756003d3-7d4c-4fec-b738-1d8d45bf048c-image-registry-private-configuration" (OuterVolumeSpecName: "image-registry-private-configuration") pod "756003d3-7d4c-4fec-b738-1d8d45bf048c" (UID: "756003d3-7d4c-4fec-b738-1d8d45bf048c"). InnerVolumeSpecName "image-registry-private-configuration". PluginName "kubernetes.io/secret", VolumeGIDValue "" Feb 17 12:49:43.438477 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.438385 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "756003d3-7d4c-4fec-b738-1d8d45bf048c" (UID: "756003d3-7d4c-4fec-b738-1d8d45bf048c"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Feb 17 12:49:43.439430 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.439403 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-kube-api-access-4cf9b" (OuterVolumeSpecName: "kube-api-access-4cf9b") pod "756003d3-7d4c-4fec-b738-1d8d45bf048c" (UID: "756003d3-7d4c-4fec-b738-1d8d45bf048c"). InnerVolumeSpecName "kube-api-access-4cf9b". PluginName "kubernetes.io/projected", VolumeGIDValue "" Feb 17 12:49:43.439541 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.439518 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "756003d3-7d4c-4fec-b738-1d8d45bf048c" (UID: "756003d3-7d4c-4fec-b738-1d8d45bf048c"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGIDValue "" Feb 17 12:49:43.443929 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.443905 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/756003d3-7d4c-4fec-b738-1d8d45bf048c-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "756003d3-7d4c-4fec-b738-1d8d45bf048c" (UID: "756003d3-7d4c-4fec-b738-1d8d45bf048c"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Feb 17 12:49:43.535694 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.535599 2572 reconciler_common.go:299] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-registry-tls\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:49:43.535694 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.535630 2572 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-4cf9b\" (UniqueName: \"kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-kube-api-access-4cf9b\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:49:43.535694 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.535640 2572 reconciler_common.go:299] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/756003d3-7d4c-4fec-b738-1d8d45bf048c-installation-pull-secrets\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:49:43.535694 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.535650 2572 reconciler_common.go:299] "Volume detached for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/756003d3-7d4c-4fec-b738-1d8d45bf048c-image-registry-private-configuration\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:49:43.535694 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.535660 2572 reconciler_common.go:299] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/756003d3-7d4c-4fec-b738-1d8d45bf048c-ca-trust-extracted\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:49:43.535694 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.535669 2572 reconciler_common.go:299] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/756003d3-7d4c-4fec-b738-1d8d45bf048c-trusted-ca\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:49:43.535694 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.535677 2572 reconciler_common.go:299] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/756003d3-7d4c-4fec-b738-1d8d45bf048c-registry-certificates\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:49:43.535694 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.535688 2572 reconciler_common.go:299] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/756003d3-7d4c-4fec-b738-1d8d45bf048c-bound-sa-token\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:49:43.719173 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.719140 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"5d6ebab3-fa58-444e-8b5e-e9d52e188752","Type":"ContainerStarted","Data":"5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df"} Feb 17 12:49:43.719322 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.719181 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"5d6ebab3-fa58-444e-8b5e-e9d52e188752","Type":"ContainerStarted","Data":"0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3"} Feb 17 12:49:43.720153 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.720129 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" event={"ID":"756003d3-7d4c-4fec-b738-1d8d45bf048c","Type":"ContainerDied","Data":"cd4901d2c05a7ae08422f57471d55a04ab1d86d8d7ae104d4933efde3574e670"} Feb 17 12:49:43.720266 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.720169 2572 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-68598f8b47-r5pmz" Feb 17 12:49:43.720311 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.720170 2572 scope.go:117] "RemoveContainer" containerID="c4999988293f533ed9aad40d6664284b6887f8904a29bdfa9aa3650002705574" Feb 17 12:49:43.741807 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.741782 2572 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-68598f8b47-r5pmz"] Feb 17 12:49:43.745584 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:43.745564 2572 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-68598f8b47-r5pmz"] Feb 17 12:49:44.131772 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:44.131742 2572 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="756003d3-7d4c-4fec-b738-1d8d45bf048c" path="/var/lib/kubelet/pods/756003d3-7d4c-4fec-b738-1d8d45bf048c/volumes" Feb 17 12:49:45.731321 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:45.731287 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"5d6ebab3-fa58-444e-8b5e-e9d52e188752","Type":"ContainerStarted","Data":"3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49"} Feb 17 12:49:45.731321 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:45.731324 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"5d6ebab3-fa58-444e-8b5e-e9d52e188752","Type":"ContainerStarted","Data":"1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f"} Feb 17 12:49:45.731734 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:45.731334 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"5d6ebab3-fa58-444e-8b5e-e9d52e188752","Type":"ContainerStarted","Data":"984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832"} Feb 17 12:49:45.731734 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:45.731344 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"5d6ebab3-fa58-444e-8b5e-e9d52e188752","Type":"ContainerStarted","Data":"68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf"} Feb 17 12:49:45.765389 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:45.765329 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-k8s-0" podStartSLOduration=2.314262935 podStartE2EDuration="6.76531158s" podCreationTimestamp="2026-02-17 12:49:39 +0000 UTC" firstStartedPulling="2026-02-17 12:49:40.706596798 +0000 UTC m=+187.057007602" lastFinishedPulling="2026-02-17 12:49:45.157645448 +0000 UTC m=+191.508056247" observedRunningTime="2026-02-17 12:49:45.764782303 +0000 UTC m=+192.115193115" watchObservedRunningTime="2026-02-17 12:49:45.76531158 +0000 UTC m=+192.115722393" Feb 17 12:49:49.438899 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:49.438846 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:49:57.710649 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:57.710618 2572 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" Feb 17 12:49:57.710649 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:49:57.710658 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" Feb 17 12:50:17.716987 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:17.716957 2572 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" Feb 17 12:50:17.721303 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:17.721274 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/metrics-server-86f7684bf5-m2r9q" Feb 17 12:50:39.439383 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:39.439347 2572 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:39.454679 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:39.454650 2572 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:39.896009 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:39.895937 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:44.958077 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:44.958042 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs\") pod \"network-metrics-daemon-vmpwv\" (UID: \"685523da-d834-463a-8a1a-ef70069c19c5\") " pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:50:44.960321 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:44.960301 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/685523da-d834-463a-8a1a-ef70069c19c5-metrics-certs\") pod \"network-metrics-daemon-vmpwv\" (UID: \"685523da-d834-463a-8a1a-ef70069c19c5\") " pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:50:45.229550 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:45.229478 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-b7dtp\"" Feb 17 12:50:45.237702 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:45.237686 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vmpwv" Feb 17 12:50:45.353521 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:45.353492 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-vmpwv"] Feb 17 12:50:45.356912 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:50:45.356886 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod685523da_d834_463a_8a1a_ef70069c19c5.slice/crio-c8e2aa99038d9865b32926ae8beda5f1f9b95f7274a4c21556bbf5d9a86a6f64 WatchSource:0}: Error finding container c8e2aa99038d9865b32926ae8beda5f1f9b95f7274a4c21556bbf5d9a86a6f64: Status 404 returned error can't find the container with id c8e2aa99038d9865b32926ae8beda5f1f9b95f7274a4c21556bbf5d9a86a6f64 Feb 17 12:50:45.901191 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:45.901153 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-vmpwv" event={"ID":"685523da-d834-463a-8a1a-ef70069c19c5","Type":"ContainerStarted","Data":"c8e2aa99038d9865b32926ae8beda5f1f9b95f7274a4c21556bbf5d9a86a6f64"} Feb 17 12:50:46.905465 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:46.905431 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-vmpwv" event={"ID":"685523da-d834-463a-8a1a-ef70069c19c5","Type":"ContainerStarted","Data":"e183c3f4ac45475d11d2ade52f905224933e11477e21af6cc7be98e6cb92ca25"} Feb 17 12:50:46.905465 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:46.905469 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-vmpwv" event={"ID":"685523da-d834-463a-8a1a-ef70069c19c5","Type":"ContainerStarted","Data":"7168bd222fdd5bf1e80c2e4f93272bc09a5ff54587ca9d2b8e5cf597d87f56b8"} Feb 17 12:50:46.920396 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:46.920345 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-vmpwv" podStartSLOduration=251.945713279 podStartE2EDuration="4m12.920329272s" podCreationTimestamp="2026-02-17 12:46:34 +0000 UTC" firstStartedPulling="2026-02-17 12:50:45.358667074 +0000 UTC m=+251.709077864" lastFinishedPulling="2026-02-17 12:50:46.333283053 +0000 UTC m=+252.683693857" observedRunningTime="2026-02-17 12:50:46.91839418 +0000 UTC m=+253.268804992" watchObservedRunningTime="2026-02-17 12:50:46.920329272 +0000 UTC m=+253.270740085" Feb 17 12:50:53.235718 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:53.235684 2572 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Feb 17 12:50:53.926810 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:53.926718 2572 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID="35751e92-fed8-4a26-931d-1e9070e8181b" containerName="alertmanager" containerID="cri-o://a3b2f792f814616f4f04b46a1e85d624eeff05eb1c1c0100c30721cadf92d6e8" gracePeriod=120 Feb 17 12:50:53.926810 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:53.926778 2572 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID="35751e92-fed8-4a26-931d-1e9070e8181b" containerName="kube-rbac-proxy-metric" containerID="cri-o://b9ccfdcb7b936cf256bb6ba8141cc1d830719805be1ac916d15da1dff87968aa" gracePeriod=120 Feb 17 12:50:53.927098 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:53.926807 2572 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID="35751e92-fed8-4a26-931d-1e9070e8181b" containerName="kube-rbac-proxy-web" containerID="cri-o://fd8b3b87a83fc0ca14a6d30a829101f488a3d8c20593ee20c27bd20a02195c4f" gracePeriod=120 Feb 17 12:50:53.927098 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:53.926840 2572 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID="35751e92-fed8-4a26-931d-1e9070e8181b" containerName="config-reloader" containerID="cri-o://d1c51b0dcc092df1a64c98743fddd1c3a1961955929a955cae250827703f882c" gracePeriod=120 Feb 17 12:50:53.927098 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:53.926877 2572 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID="35751e92-fed8-4a26-931d-1e9070e8181b" containerName="prom-label-proxy" containerID="cri-o://ce42649490e3041f1dca08fb8b38628f86a6f7c94c0dc8ff19460069ff25a516" gracePeriod=120 Feb 17 12:50:53.927098 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:53.926969 2572 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID="35751e92-fed8-4a26-931d-1e9070e8181b" containerName="kube-rbac-proxy" containerID="cri-o://98fa9cd1d827c70e3d844432f4b6ddbb6a65e85fe8daf94445a1134a168ac445" gracePeriod=120 Feb 17 12:50:54.932651 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:54.932620 2572 generic.go:358] "Generic (PLEG): container finished" podID="35751e92-fed8-4a26-931d-1e9070e8181b" containerID="ce42649490e3041f1dca08fb8b38628f86a6f7c94c0dc8ff19460069ff25a516" exitCode=0 Feb 17 12:50:54.932651 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:54.932645 2572 generic.go:358] "Generic (PLEG): container finished" podID="35751e92-fed8-4a26-931d-1e9070e8181b" containerID="98fa9cd1d827c70e3d844432f4b6ddbb6a65e85fe8daf94445a1134a168ac445" exitCode=0 Feb 17 12:50:54.932651 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:54.932651 2572 generic.go:358] "Generic (PLEG): container finished" podID="35751e92-fed8-4a26-931d-1e9070e8181b" containerID="d1c51b0dcc092df1a64c98743fddd1c3a1961955929a955cae250827703f882c" exitCode=0 Feb 17 12:50:54.932651 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:54.932656 2572 generic.go:358] "Generic (PLEG): container finished" podID="35751e92-fed8-4a26-931d-1e9070e8181b" containerID="a3b2f792f814616f4f04b46a1e85d624eeff05eb1c1c0100c30721cadf92d6e8" exitCode=0 Feb 17 12:50:54.933097 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:54.932692 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"35751e92-fed8-4a26-931d-1e9070e8181b","Type":"ContainerDied","Data":"ce42649490e3041f1dca08fb8b38628f86a6f7c94c0dc8ff19460069ff25a516"} Feb 17 12:50:54.933097 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:54.932723 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"35751e92-fed8-4a26-931d-1e9070e8181b","Type":"ContainerDied","Data":"98fa9cd1d827c70e3d844432f4b6ddbb6a65e85fe8daf94445a1134a168ac445"} Feb 17 12:50:54.933097 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:54.932733 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"35751e92-fed8-4a26-931d-1e9070e8181b","Type":"ContainerDied","Data":"d1c51b0dcc092df1a64c98743fddd1c3a1961955929a955cae250827703f882c"} Feb 17 12:50:54.933097 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:54.932746 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"35751e92-fed8-4a26-931d-1e9070e8181b","Type":"ContainerDied","Data":"a3b2f792f814616f4f04b46a1e85d624eeff05eb1c1c0100c30721cadf92d6e8"} Feb 17 12:50:55.176467 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.176443 2572 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:55.348249 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.348166 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-secret-alertmanager-main-tls\") pod \"35751e92-fed8-4a26-931d-1e9070e8181b\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " Feb 17 12:50:55.348249 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.348215 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/35751e92-fed8-4a26-931d-1e9070e8181b-alertmanager-main-db\") pod \"35751e92-fed8-4a26-931d-1e9070e8181b\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " Feb 17 12:50:55.348249 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.348238 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-secret-alertmanager-kube-rbac-proxy\") pod \"35751e92-fed8-4a26-931d-1e9070e8181b\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " Feb 17 12:50:55.348510 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.348265 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-config-volume\") pod \"35751e92-fed8-4a26-931d-1e9070e8181b\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " Feb 17 12:50:55.348510 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.348282 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/35751e92-fed8-4a26-931d-1e9070e8181b-config-out\") pod \"35751e92-fed8-4a26-931d-1e9070e8181b\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " Feb 17 12:50:55.348510 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.348303 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/35751e92-fed8-4a26-931d-1e9070e8181b-alertmanager-trusted-ca-bundle\") pod \"35751e92-fed8-4a26-931d-1e9070e8181b\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " Feb 17 12:50:55.348510 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.348332 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/35751e92-fed8-4a26-931d-1e9070e8181b-metrics-client-ca\") pod \"35751e92-fed8-4a26-931d-1e9070e8181b\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " Feb 17 12:50:55.348510 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.348371 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-cluster-tls-config\") pod \"35751e92-fed8-4a26-931d-1e9070e8181b\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " Feb 17 12:50:55.348510 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.348452 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-web-config\") pod \"35751e92-fed8-4a26-931d-1e9070e8181b\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " Feb 17 12:50:55.348510 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.348484 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-secret-alertmanager-kube-rbac-proxy-web\") pod \"35751e92-fed8-4a26-931d-1e9070e8181b\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " Feb 17 12:50:55.348842 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.348533 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kgrv\" (UniqueName: \"kubernetes.io/projected/35751e92-fed8-4a26-931d-1e9070e8181b-kube-api-access-8kgrv\") pod \"35751e92-fed8-4a26-931d-1e9070e8181b\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " Feb 17 12:50:55.348842 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.348563 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-secret-alertmanager-kube-rbac-proxy-metric\") pod \"35751e92-fed8-4a26-931d-1e9070e8181b\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " Feb 17 12:50:55.348842 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.348586 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/35751e92-fed8-4a26-931d-1e9070e8181b-tls-assets\") pod \"35751e92-fed8-4a26-931d-1e9070e8181b\" (UID: \"35751e92-fed8-4a26-931d-1e9070e8181b\") " Feb 17 12:50:55.348842 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.348615 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35751e92-fed8-4a26-931d-1e9070e8181b-alertmanager-main-db" (OuterVolumeSpecName: "alertmanager-main-db") pod "35751e92-fed8-4a26-931d-1e9070e8181b" (UID: "35751e92-fed8-4a26-931d-1e9070e8181b"). InnerVolumeSpecName "alertmanager-main-db". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Feb 17 12:50:55.348842 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.348759 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35751e92-fed8-4a26-931d-1e9070e8181b-alertmanager-trusted-ca-bundle" (OuterVolumeSpecName: "alertmanager-trusted-ca-bundle") pod "35751e92-fed8-4a26-931d-1e9070e8181b" (UID: "35751e92-fed8-4a26-931d-1e9070e8181b"). InnerVolumeSpecName "alertmanager-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Feb 17 12:50:55.348842 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.348774 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35751e92-fed8-4a26-931d-1e9070e8181b-metrics-client-ca" (OuterVolumeSpecName: "metrics-client-ca") pod "35751e92-fed8-4a26-931d-1e9070e8181b" (UID: "35751e92-fed8-4a26-931d-1e9070e8181b"). InnerVolumeSpecName "metrics-client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Feb 17 12:50:55.349162 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.348899 2572 reconciler_common.go:299] "Volume detached for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/35751e92-fed8-4a26-931d-1e9070e8181b-alertmanager-main-db\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:55.349162 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.348922 2572 reconciler_common.go:299] "Volume detached for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/35751e92-fed8-4a26-931d-1e9070e8181b-alertmanager-trusted-ca-bundle\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:55.349162 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.348940 2572 reconciler_common.go:299] "Volume detached for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/35751e92-fed8-4a26-931d-1e9070e8181b-metrics-client-ca\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:55.351234 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.351192 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-secret-alertmanager-main-tls" (OuterVolumeSpecName: "secret-alertmanager-main-tls") pod "35751e92-fed8-4a26-931d-1e9070e8181b" (UID: "35751e92-fed8-4a26-931d-1e9070e8181b"). InnerVolumeSpecName "secret-alertmanager-main-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Feb 17 12:50:55.351359 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.351236 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35751e92-fed8-4a26-931d-1e9070e8181b-config-out" (OuterVolumeSpecName: "config-out") pod "35751e92-fed8-4a26-931d-1e9070e8181b" (UID: "35751e92-fed8-4a26-931d-1e9070e8181b"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Feb 17 12:50:55.351359 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.351316 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-config-volume" (OuterVolumeSpecName: "config-volume") pod "35751e92-fed8-4a26-931d-1e9070e8181b" (UID: "35751e92-fed8-4a26-931d-1e9070e8181b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/secret", VolumeGIDValue "" Feb 17 12:50:55.351470 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.351364 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-secret-alertmanager-kube-rbac-proxy-web" (OuterVolumeSpecName: "secret-alertmanager-kube-rbac-proxy-web") pod "35751e92-fed8-4a26-931d-1e9070e8181b" (UID: "35751e92-fed8-4a26-931d-1e9070e8181b"). InnerVolumeSpecName "secret-alertmanager-kube-rbac-proxy-web". PluginName "kubernetes.io/secret", VolumeGIDValue "" Feb 17 12:50:55.351755 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.351730 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-secret-alertmanager-kube-rbac-proxy" (OuterVolumeSpecName: "secret-alertmanager-kube-rbac-proxy") pod "35751e92-fed8-4a26-931d-1e9070e8181b" (UID: "35751e92-fed8-4a26-931d-1e9070e8181b"). InnerVolumeSpecName "secret-alertmanager-kube-rbac-proxy". PluginName "kubernetes.io/secret", VolumeGIDValue "" Feb 17 12:50:55.351964 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.351932 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35751e92-fed8-4a26-931d-1e9070e8181b-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "35751e92-fed8-4a26-931d-1e9070e8181b" (UID: "35751e92-fed8-4a26-931d-1e9070e8181b"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGIDValue "" Feb 17 12:50:55.352042 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.351956 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35751e92-fed8-4a26-931d-1e9070e8181b-kube-api-access-8kgrv" (OuterVolumeSpecName: "kube-api-access-8kgrv") pod "35751e92-fed8-4a26-931d-1e9070e8181b" (UID: "35751e92-fed8-4a26-931d-1e9070e8181b"). InnerVolumeSpecName "kube-api-access-8kgrv". PluginName "kubernetes.io/projected", VolumeGIDValue "" Feb 17 12:50:55.353059 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.353033 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-secret-alertmanager-kube-rbac-proxy-metric" (OuterVolumeSpecName: "secret-alertmanager-kube-rbac-proxy-metric") pod "35751e92-fed8-4a26-931d-1e9070e8181b" (UID: "35751e92-fed8-4a26-931d-1e9070e8181b"). InnerVolumeSpecName "secret-alertmanager-kube-rbac-proxy-metric". PluginName "kubernetes.io/secret", VolumeGIDValue "" Feb 17 12:50:55.356434 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.356412 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-cluster-tls-config" (OuterVolumeSpecName: "cluster-tls-config") pod "35751e92-fed8-4a26-931d-1e9070e8181b" (UID: "35751e92-fed8-4a26-931d-1e9070e8181b"). InnerVolumeSpecName "cluster-tls-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Feb 17 12:50:55.362765 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.362741 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-web-config" (OuterVolumeSpecName: "web-config") pod "35751e92-fed8-4a26-931d-1e9070e8181b" (UID: "35751e92-fed8-4a26-931d-1e9070e8181b"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Feb 17 12:50:55.450025 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.449989 2572 reconciler_common.go:299] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-web-config\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:55.450025 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.450019 2572 reconciler_common.go:299] "Volume detached for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-secret-alertmanager-kube-rbac-proxy-web\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:55.450025 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.450030 2572 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-8kgrv\" (UniqueName: \"kubernetes.io/projected/35751e92-fed8-4a26-931d-1e9070e8181b-kube-api-access-8kgrv\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:55.450261 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.450040 2572 reconciler_common.go:299] "Volume detached for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-secret-alertmanager-kube-rbac-proxy-metric\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:55.450261 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.450051 2572 reconciler_common.go:299] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/35751e92-fed8-4a26-931d-1e9070e8181b-tls-assets\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:55.450261 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.450060 2572 reconciler_common.go:299] "Volume detached for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-secret-alertmanager-main-tls\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:55.450261 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.450070 2572 reconciler_common.go:299] "Volume detached for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-secret-alertmanager-kube-rbac-proxy\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:55.450261 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.450080 2572 reconciler_common.go:299] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-config-volume\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:55.450261 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.450090 2572 reconciler_common.go:299] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/35751e92-fed8-4a26-931d-1e9070e8181b-config-out\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:55.450261 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.450099 2572 reconciler_common.go:299] "Volume detached for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/35751e92-fed8-4a26-931d-1e9070e8181b-cluster-tls-config\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:55.938363 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.938331 2572 generic.go:358] "Generic (PLEG): container finished" podID="35751e92-fed8-4a26-931d-1e9070e8181b" containerID="b9ccfdcb7b936cf256bb6ba8141cc1d830719805be1ac916d15da1dff87968aa" exitCode=0 Feb 17 12:50:55.938363 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.938359 2572 generic.go:358] "Generic (PLEG): container finished" podID="35751e92-fed8-4a26-931d-1e9070e8181b" containerID="fd8b3b87a83fc0ca14a6d30a829101f488a3d8c20593ee20c27bd20a02195c4f" exitCode=0 Feb 17 12:50:55.938800 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.938416 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"35751e92-fed8-4a26-931d-1e9070e8181b","Type":"ContainerDied","Data":"b9ccfdcb7b936cf256bb6ba8141cc1d830719805be1ac916d15da1dff87968aa"} Feb 17 12:50:55.938800 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.938436 2572 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:55.938800 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.938456 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"35751e92-fed8-4a26-931d-1e9070e8181b","Type":"ContainerDied","Data":"fd8b3b87a83fc0ca14a6d30a829101f488a3d8c20593ee20c27bd20a02195c4f"} Feb 17 12:50:55.938800 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.938468 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"35751e92-fed8-4a26-931d-1e9070e8181b","Type":"ContainerDied","Data":"64694402cdb3bdfea6e7850d714e8f6bc96bc653a2fb8ff93ffb8597fee9273e"} Feb 17 12:50:55.938800 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.938492 2572 scope.go:117] "RemoveContainer" containerID="ce42649490e3041f1dca08fb8b38628f86a6f7c94c0dc8ff19460069ff25a516" Feb 17 12:50:55.947156 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.947136 2572 scope.go:117] "RemoveContainer" containerID="b9ccfdcb7b936cf256bb6ba8141cc1d830719805be1ac916d15da1dff87968aa" Feb 17 12:50:55.954290 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.954265 2572 scope.go:117] "RemoveContainer" containerID="98fa9cd1d827c70e3d844432f4b6ddbb6a65e85fe8daf94445a1134a168ac445" Feb 17 12:50:55.961248 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.961224 2572 scope.go:117] "RemoveContainer" containerID="fd8b3b87a83fc0ca14a6d30a829101f488a3d8c20593ee20c27bd20a02195c4f" Feb 17 12:50:55.961365 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.961340 2572 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Feb 17 12:50:55.965094 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.965072 2572 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Feb 17 12:50:55.969005 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.968985 2572 scope.go:117] "RemoveContainer" containerID="d1c51b0dcc092df1a64c98743fddd1c3a1961955929a955cae250827703f882c" Feb 17 12:50:55.975596 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.975577 2572 scope.go:117] "RemoveContainer" containerID="a3b2f792f814616f4f04b46a1e85d624eeff05eb1c1c0100c30721cadf92d6e8" Feb 17 12:50:55.982401 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.982377 2572 scope.go:117] "RemoveContainer" containerID="6f5bb117d629c48ad1ffcb2dc8b5c7fc16e173a1673d071f0ec52f4b482e6b44" Feb 17 12:50:55.989080 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.988904 2572 scope.go:117] "RemoveContainer" containerID="ce42649490e3041f1dca08fb8b38628f86a6f7c94c0dc8ff19460069ff25a516" Feb 17 12:50:55.989856 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:50:55.989515 2572 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce42649490e3041f1dca08fb8b38628f86a6f7c94c0dc8ff19460069ff25a516\": container with ID starting with ce42649490e3041f1dca08fb8b38628f86a6f7c94c0dc8ff19460069ff25a516 not found: ID does not exist" containerID="ce42649490e3041f1dca08fb8b38628f86a6f7c94c0dc8ff19460069ff25a516" Feb 17 12:50:55.989856 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.989575 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce42649490e3041f1dca08fb8b38628f86a6f7c94c0dc8ff19460069ff25a516"} err="failed to get container status \"ce42649490e3041f1dca08fb8b38628f86a6f7c94c0dc8ff19460069ff25a516\": rpc error: code = NotFound desc = could not find container \"ce42649490e3041f1dca08fb8b38628f86a6f7c94c0dc8ff19460069ff25a516\": container with ID starting with ce42649490e3041f1dca08fb8b38628f86a6f7c94c0dc8ff19460069ff25a516 not found: ID does not exist" Feb 17 12:50:55.989856 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.989642 2572 scope.go:117] "RemoveContainer" containerID="b9ccfdcb7b936cf256bb6ba8141cc1d830719805be1ac916d15da1dff87968aa" Feb 17 12:50:55.990059 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:50:55.989888 2572 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9ccfdcb7b936cf256bb6ba8141cc1d830719805be1ac916d15da1dff87968aa\": container with ID starting with b9ccfdcb7b936cf256bb6ba8141cc1d830719805be1ac916d15da1dff87968aa not found: ID does not exist" containerID="b9ccfdcb7b936cf256bb6ba8141cc1d830719805be1ac916d15da1dff87968aa" Feb 17 12:50:55.990059 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.989912 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9ccfdcb7b936cf256bb6ba8141cc1d830719805be1ac916d15da1dff87968aa"} err="failed to get container status \"b9ccfdcb7b936cf256bb6ba8141cc1d830719805be1ac916d15da1dff87968aa\": rpc error: code = NotFound desc = could not find container \"b9ccfdcb7b936cf256bb6ba8141cc1d830719805be1ac916d15da1dff87968aa\": container with ID starting with b9ccfdcb7b936cf256bb6ba8141cc1d830719805be1ac916d15da1dff87968aa not found: ID does not exist" Feb 17 12:50:55.990059 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.989928 2572 scope.go:117] "RemoveContainer" containerID="98fa9cd1d827c70e3d844432f4b6ddbb6a65e85fe8daf94445a1134a168ac445" Feb 17 12:50:55.990233 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:50:55.990218 2572 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98fa9cd1d827c70e3d844432f4b6ddbb6a65e85fe8daf94445a1134a168ac445\": container with ID starting with 98fa9cd1d827c70e3d844432f4b6ddbb6a65e85fe8daf94445a1134a168ac445 not found: ID does not exist" containerID="98fa9cd1d827c70e3d844432f4b6ddbb6a65e85fe8daf94445a1134a168ac445" Feb 17 12:50:55.990286 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.990238 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98fa9cd1d827c70e3d844432f4b6ddbb6a65e85fe8daf94445a1134a168ac445"} err="failed to get container status \"98fa9cd1d827c70e3d844432f4b6ddbb6a65e85fe8daf94445a1134a168ac445\": rpc error: code = NotFound desc = could not find container \"98fa9cd1d827c70e3d844432f4b6ddbb6a65e85fe8daf94445a1134a168ac445\": container with ID starting with 98fa9cd1d827c70e3d844432f4b6ddbb6a65e85fe8daf94445a1134a168ac445 not found: ID does not exist" Feb 17 12:50:55.990286 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.990255 2572 scope.go:117] "RemoveContainer" containerID="fd8b3b87a83fc0ca14a6d30a829101f488a3d8c20593ee20c27bd20a02195c4f" Feb 17 12:50:55.990598 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:50:55.990572 2572 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd8b3b87a83fc0ca14a6d30a829101f488a3d8c20593ee20c27bd20a02195c4f\": container with ID starting with fd8b3b87a83fc0ca14a6d30a829101f488a3d8c20593ee20c27bd20a02195c4f not found: ID does not exist" containerID="fd8b3b87a83fc0ca14a6d30a829101f488a3d8c20593ee20c27bd20a02195c4f" Feb 17 12:50:55.990671 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.990608 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd8b3b87a83fc0ca14a6d30a829101f488a3d8c20593ee20c27bd20a02195c4f"} err="failed to get container status \"fd8b3b87a83fc0ca14a6d30a829101f488a3d8c20593ee20c27bd20a02195c4f\": rpc error: code = NotFound desc = could not find container \"fd8b3b87a83fc0ca14a6d30a829101f488a3d8c20593ee20c27bd20a02195c4f\": container with ID starting with fd8b3b87a83fc0ca14a6d30a829101f488a3d8c20593ee20c27bd20a02195c4f not found: ID does not exist" Feb 17 12:50:55.990671 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.990629 2572 scope.go:117] "RemoveContainer" containerID="d1c51b0dcc092df1a64c98743fddd1c3a1961955929a955cae250827703f882c" Feb 17 12:50:55.991013 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.990990 2572 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Feb 17 12:50:55.991013 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:50:55.991004 2572 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1c51b0dcc092df1a64c98743fddd1c3a1961955929a955cae250827703f882c\": container with ID starting with d1c51b0dcc092df1a64c98743fddd1c3a1961955929a955cae250827703f882c not found: ID does not exist" containerID="d1c51b0dcc092df1a64c98743fddd1c3a1961955929a955cae250827703f882c" Feb 17 12:50:55.991172 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991025 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1c51b0dcc092df1a64c98743fddd1c3a1961955929a955cae250827703f882c"} err="failed to get container status \"d1c51b0dcc092df1a64c98743fddd1c3a1961955929a955cae250827703f882c\": rpc error: code = NotFound desc = could not find container \"d1c51b0dcc092df1a64c98743fddd1c3a1961955929a955cae250827703f882c\": container with ID starting with d1c51b0dcc092df1a64c98743fddd1c3a1961955929a955cae250827703f882c not found: ID does not exist" Feb 17 12:50:55.991172 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991039 2572 scope.go:117] "RemoveContainer" containerID="a3b2f792f814616f4f04b46a1e85d624eeff05eb1c1c0100c30721cadf92d6e8" Feb 17 12:50:55.991280 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:50:55.991263 2572 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3b2f792f814616f4f04b46a1e85d624eeff05eb1c1c0100c30721cadf92d6e8\": container with ID starting with a3b2f792f814616f4f04b46a1e85d624eeff05eb1c1c0100c30721cadf92d6e8 not found: ID does not exist" containerID="a3b2f792f814616f4f04b46a1e85d624eeff05eb1c1c0100c30721cadf92d6e8" Feb 17 12:50:55.991327 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991283 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3b2f792f814616f4f04b46a1e85d624eeff05eb1c1c0100c30721cadf92d6e8"} err="failed to get container status \"a3b2f792f814616f4f04b46a1e85d624eeff05eb1c1c0100c30721cadf92d6e8\": rpc error: code = NotFound desc = could not find container \"a3b2f792f814616f4f04b46a1e85d624eeff05eb1c1c0100c30721cadf92d6e8\": container with ID starting with a3b2f792f814616f4f04b46a1e85d624eeff05eb1c1c0100c30721cadf92d6e8 not found: ID does not exist" Feb 17 12:50:55.991327 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991302 2572 scope.go:117] "RemoveContainer" containerID="6f5bb117d629c48ad1ffcb2dc8b5c7fc16e173a1673d071f0ec52f4b482e6b44" Feb 17 12:50:55.991402 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991384 2572 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="756003d3-7d4c-4fec-b738-1d8d45bf048c" containerName="registry" Feb 17 12:50:55.991449 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991401 2572 state_mem.go:107] "Deleted CPUSet assignment" podUID="756003d3-7d4c-4fec-b738-1d8d45bf048c" containerName="registry" Feb 17 12:50:55.991449 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991416 2572 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="35751e92-fed8-4a26-931d-1e9070e8181b" containerName="alertmanager" Feb 17 12:50:55.991449 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991425 2572 state_mem.go:107] "Deleted CPUSet assignment" podUID="35751e92-fed8-4a26-931d-1e9070e8181b" containerName="alertmanager" Feb 17 12:50:55.991449 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991443 2572 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="35751e92-fed8-4a26-931d-1e9070e8181b" containerName="kube-rbac-proxy-metric" Feb 17 12:50:55.991583 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991452 2572 state_mem.go:107] "Deleted CPUSet assignment" podUID="35751e92-fed8-4a26-931d-1e9070e8181b" containerName="kube-rbac-proxy-metric" Feb 17 12:50:55.991583 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991463 2572 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="35751e92-fed8-4a26-931d-1e9070e8181b" containerName="config-reloader" Feb 17 12:50:55.991583 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991471 2572 state_mem.go:107] "Deleted CPUSet assignment" podUID="35751e92-fed8-4a26-931d-1e9070e8181b" containerName="config-reloader" Feb 17 12:50:55.991583 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991481 2572 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="35751e92-fed8-4a26-931d-1e9070e8181b" containerName="kube-rbac-proxy-web" Feb 17 12:50:55.991583 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991490 2572 state_mem.go:107] "Deleted CPUSet assignment" podUID="35751e92-fed8-4a26-931d-1e9070e8181b" containerName="kube-rbac-proxy-web" Feb 17 12:50:55.991583 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991504 2572 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="35751e92-fed8-4a26-931d-1e9070e8181b" containerName="prom-label-proxy" Feb 17 12:50:55.991583 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991509 2572 state_mem.go:107] "Deleted CPUSet assignment" podUID="35751e92-fed8-4a26-931d-1e9070e8181b" containerName="prom-label-proxy" Feb 17 12:50:55.991583 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991517 2572 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="35751e92-fed8-4a26-931d-1e9070e8181b" containerName="kube-rbac-proxy" Feb 17 12:50:55.991583 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991524 2572 state_mem.go:107] "Deleted CPUSet assignment" podUID="35751e92-fed8-4a26-931d-1e9070e8181b" containerName="kube-rbac-proxy" Feb 17 12:50:55.991583 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991535 2572 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="35751e92-fed8-4a26-931d-1e9070e8181b" containerName="init-config-reloader" Feb 17 12:50:55.991583 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991540 2572 state_mem.go:107] "Deleted CPUSet assignment" podUID="35751e92-fed8-4a26-931d-1e9070e8181b" containerName="init-config-reloader" Feb 17 12:50:55.991583 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:50:55.991554 2572 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f5bb117d629c48ad1ffcb2dc8b5c7fc16e173a1673d071f0ec52f4b482e6b44\": container with ID starting with 6f5bb117d629c48ad1ffcb2dc8b5c7fc16e173a1673d071f0ec52f4b482e6b44 not found: ID does not exist" containerID="6f5bb117d629c48ad1ffcb2dc8b5c7fc16e173a1673d071f0ec52f4b482e6b44" Feb 17 12:50:55.991999 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991582 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f5bb117d629c48ad1ffcb2dc8b5c7fc16e173a1673d071f0ec52f4b482e6b44"} err="failed to get container status \"6f5bb117d629c48ad1ffcb2dc8b5c7fc16e173a1673d071f0ec52f4b482e6b44\": rpc error: code = NotFound desc = could not find container \"6f5bb117d629c48ad1ffcb2dc8b5c7fc16e173a1673d071f0ec52f4b482e6b44\": container with ID starting with 6f5bb117d629c48ad1ffcb2dc8b5c7fc16e173a1673d071f0ec52f4b482e6b44 not found: ID does not exist" Feb 17 12:50:55.991999 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991597 2572 memory_manager.go:356] "RemoveStaleState removing state" podUID="35751e92-fed8-4a26-931d-1e9070e8181b" containerName="kube-rbac-proxy-metric" Feb 17 12:50:55.991999 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991603 2572 scope.go:117] "RemoveContainer" containerID="ce42649490e3041f1dca08fb8b38628f86a6f7c94c0dc8ff19460069ff25a516" Feb 17 12:50:55.991999 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991607 2572 memory_manager.go:356] "RemoveStaleState removing state" podUID="756003d3-7d4c-4fec-b738-1d8d45bf048c" containerName="registry" Feb 17 12:50:55.991999 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991615 2572 memory_manager.go:356] "RemoveStaleState removing state" podUID="35751e92-fed8-4a26-931d-1e9070e8181b" containerName="alertmanager" Feb 17 12:50:55.991999 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991623 2572 memory_manager.go:356] "RemoveStaleState removing state" podUID="35751e92-fed8-4a26-931d-1e9070e8181b" containerName="kube-rbac-proxy-web" Feb 17 12:50:55.991999 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991630 2572 memory_manager.go:356] "RemoveStaleState removing state" podUID="35751e92-fed8-4a26-931d-1e9070e8181b" containerName="prom-label-proxy" Feb 17 12:50:55.991999 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991636 2572 memory_manager.go:356] "RemoveStaleState removing state" podUID="35751e92-fed8-4a26-931d-1e9070e8181b" containerName="config-reloader" Feb 17 12:50:55.991999 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991644 2572 memory_manager.go:356] "RemoveStaleState removing state" podUID="35751e92-fed8-4a26-931d-1e9070e8181b" containerName="kube-rbac-proxy" Feb 17 12:50:55.991999 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991850 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce42649490e3041f1dca08fb8b38628f86a6f7c94c0dc8ff19460069ff25a516"} err="failed to get container status \"ce42649490e3041f1dca08fb8b38628f86a6f7c94c0dc8ff19460069ff25a516\": rpc error: code = NotFound desc = could not find container \"ce42649490e3041f1dca08fb8b38628f86a6f7c94c0dc8ff19460069ff25a516\": container with ID starting with ce42649490e3041f1dca08fb8b38628f86a6f7c94c0dc8ff19460069ff25a516 not found: ID does not exist" Feb 17 12:50:55.991999 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.991875 2572 scope.go:117] "RemoveContainer" containerID="b9ccfdcb7b936cf256bb6ba8141cc1d830719805be1ac916d15da1dff87968aa" Feb 17 12:50:55.992436 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.992156 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9ccfdcb7b936cf256bb6ba8141cc1d830719805be1ac916d15da1dff87968aa"} err="failed to get container status \"b9ccfdcb7b936cf256bb6ba8141cc1d830719805be1ac916d15da1dff87968aa\": rpc error: code = NotFound desc = could not find container \"b9ccfdcb7b936cf256bb6ba8141cc1d830719805be1ac916d15da1dff87968aa\": container with ID starting with b9ccfdcb7b936cf256bb6ba8141cc1d830719805be1ac916d15da1dff87968aa not found: ID does not exist" Feb 17 12:50:55.992436 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.992173 2572 scope.go:117] "RemoveContainer" containerID="98fa9cd1d827c70e3d844432f4b6ddbb6a65e85fe8daf94445a1134a168ac445" Feb 17 12:50:55.992436 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.992402 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98fa9cd1d827c70e3d844432f4b6ddbb6a65e85fe8daf94445a1134a168ac445"} err="failed to get container status \"98fa9cd1d827c70e3d844432f4b6ddbb6a65e85fe8daf94445a1134a168ac445\": rpc error: code = NotFound desc = could not find container \"98fa9cd1d827c70e3d844432f4b6ddbb6a65e85fe8daf94445a1134a168ac445\": container with ID starting with 98fa9cd1d827c70e3d844432f4b6ddbb6a65e85fe8daf94445a1134a168ac445 not found: ID does not exist" Feb 17 12:50:55.992436 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.992417 2572 scope.go:117] "RemoveContainer" containerID="fd8b3b87a83fc0ca14a6d30a829101f488a3d8c20593ee20c27bd20a02195c4f" Feb 17 12:50:55.992608 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.992593 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd8b3b87a83fc0ca14a6d30a829101f488a3d8c20593ee20c27bd20a02195c4f"} err="failed to get container status \"fd8b3b87a83fc0ca14a6d30a829101f488a3d8c20593ee20c27bd20a02195c4f\": rpc error: code = NotFound desc = could not find container \"fd8b3b87a83fc0ca14a6d30a829101f488a3d8c20593ee20c27bd20a02195c4f\": container with ID starting with fd8b3b87a83fc0ca14a6d30a829101f488a3d8c20593ee20c27bd20a02195c4f not found: ID does not exist" Feb 17 12:50:55.992656 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.992607 2572 scope.go:117] "RemoveContainer" containerID="d1c51b0dcc092df1a64c98743fddd1c3a1961955929a955cae250827703f882c" Feb 17 12:50:55.992836 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.992814 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1c51b0dcc092df1a64c98743fddd1c3a1961955929a955cae250827703f882c"} err="failed to get container status \"d1c51b0dcc092df1a64c98743fddd1c3a1961955929a955cae250827703f882c\": rpc error: code = NotFound desc = could not find container \"d1c51b0dcc092df1a64c98743fddd1c3a1961955929a955cae250827703f882c\": container with ID starting with d1c51b0dcc092df1a64c98743fddd1c3a1961955929a955cae250827703f882c not found: ID does not exist" Feb 17 12:50:55.992881 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.992836 2572 scope.go:117] "RemoveContainer" containerID="a3b2f792f814616f4f04b46a1e85d624eeff05eb1c1c0100c30721cadf92d6e8" Feb 17 12:50:55.993151 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.993093 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3b2f792f814616f4f04b46a1e85d624eeff05eb1c1c0100c30721cadf92d6e8"} err="failed to get container status \"a3b2f792f814616f4f04b46a1e85d624eeff05eb1c1c0100c30721cadf92d6e8\": rpc error: code = NotFound desc = could not find container \"a3b2f792f814616f4f04b46a1e85d624eeff05eb1c1c0100c30721cadf92d6e8\": container with ID starting with a3b2f792f814616f4f04b46a1e85d624eeff05eb1c1c0100c30721cadf92d6e8 not found: ID does not exist" Feb 17 12:50:55.993206 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.993152 2572 scope.go:117] "RemoveContainer" containerID="6f5bb117d629c48ad1ffcb2dc8b5c7fc16e173a1673d071f0ec52f4b482e6b44" Feb 17 12:50:55.993428 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.993409 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f5bb117d629c48ad1ffcb2dc8b5c7fc16e173a1673d071f0ec52f4b482e6b44"} err="failed to get container status \"6f5bb117d629c48ad1ffcb2dc8b5c7fc16e173a1673d071f0ec52f4b482e6b44\": rpc error: code = NotFound desc = could not find container \"6f5bb117d629c48ad1ffcb2dc8b5c7fc16e173a1673d071f0ec52f4b482e6b44\": container with ID starting with 6f5bb117d629c48ad1ffcb2dc8b5c7fc16e173a1673d071f0ec52f4b482e6b44 not found: ID does not exist" Feb 17 12:50:55.996880 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.996862 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:55.998678 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.998658 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-cluster-tls-config\"" Feb 17 12:50:55.998772 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.998753 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy\"" Feb 17 12:50:55.999034 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.998998 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-web-config\"" Feb 17 12:50:55.999034 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.999013 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy-web\"" Feb 17 12:50:55.999034 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.999004 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-generated\"" Feb 17 12:50:55.999252 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.999085 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-tls-assets-0\"" Feb 17 12:50:55.999252 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.999016 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-tls\"" Feb 17 12:50:55.999356 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.999299 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-kube-rbac-proxy-metric\"" Feb 17 12:50:55.999408 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:55.999381 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"alertmanager-main-dockercfg-6qj8b\"" Feb 17 12:50:56.004649 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.004619 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"alertmanager-trusted-ca-bundle\"" Feb 17 12:50:56.004903 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.004884 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Feb 17 12:50:56.130205 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.130166 2572 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35751e92-fed8-4a26-931d-1e9070e8181b" path="/var/lib/kubelet/pods/35751e92-fed8-4a26-931d-1e9070e8181b/volumes" Feb 17 12:50:56.155597 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.155565 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/49f08cb5-241b-4020-87e6-08cfd5bec0cc-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.155721 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.155611 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49f08cb5-241b-4020-87e6-08cfd5bec0cc-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.155721 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.155642 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/49f08cb5-241b-4020-87e6-08cfd5bec0cc-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.155721 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.155667 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/49f08cb5-241b-4020-87e6-08cfd5bec0cc-web-config\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.155721 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.155688 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/49f08cb5-241b-4020-87e6-08cfd5bec0cc-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.155721 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.155707 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwmnc\" (UniqueName: \"kubernetes.io/projected/49f08cb5-241b-4020-87e6-08cfd5bec0cc-kube-api-access-lwmnc\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.155874 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.155759 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/49f08cb5-241b-4020-87e6-08cfd5bec0cc-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.155874 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.155813 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/49f08cb5-241b-4020-87e6-08cfd5bec0cc-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.155874 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.155838 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/49f08cb5-241b-4020-87e6-08cfd5bec0cc-config-out\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.155960 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.155901 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/49f08cb5-241b-4020-87e6-08cfd5bec0cc-tls-assets\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.155960 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.155936 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/49f08cb5-241b-4020-87e6-08cfd5bec0cc-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.155960 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.155956 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/49f08cb5-241b-4020-87e6-08cfd5bec0cc-cluster-tls-config\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.156055 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.155975 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/49f08cb5-241b-4020-87e6-08cfd5bec0cc-config-volume\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.256439 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.256343 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/49f08cb5-241b-4020-87e6-08cfd5bec0cc-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.256439 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.256388 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-lwmnc\" (UniqueName: \"kubernetes.io/projected/49f08cb5-241b-4020-87e6-08cfd5bec0cc-kube-api-access-lwmnc\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.256439 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.256414 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/49f08cb5-241b-4020-87e6-08cfd5bec0cc-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.256719 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.256455 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/49f08cb5-241b-4020-87e6-08cfd5bec0cc-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.256719 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.256487 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/49f08cb5-241b-4020-87e6-08cfd5bec0cc-config-out\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.256719 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.256522 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/49f08cb5-241b-4020-87e6-08cfd5bec0cc-tls-assets\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.256719 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.256556 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/49f08cb5-241b-4020-87e6-08cfd5bec0cc-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.256719 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.256578 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/49f08cb5-241b-4020-87e6-08cfd5bec0cc-cluster-tls-config\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.256719 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.256602 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/49f08cb5-241b-4020-87e6-08cfd5bec0cc-config-volume\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.256719 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.256636 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/49f08cb5-241b-4020-87e6-08cfd5bec0cc-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.256719 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.256664 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49f08cb5-241b-4020-87e6-08cfd5bec0cc-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.256719 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.256699 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/49f08cb5-241b-4020-87e6-08cfd5bec0cc-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.257185 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.256740 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/49f08cb5-241b-4020-87e6-08cfd5bec0cc-web-config\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.259420 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.259392 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/49f08cb5-241b-4020-87e6-08cfd5bec0cc-config-out\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.259693 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.259668 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/49f08cb5-241b-4020-87e6-08cfd5bec0cc-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.259806 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.259709 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/49f08cb5-241b-4020-87e6-08cfd5bec0cc-config-volume\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.259806 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.259728 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/49f08cb5-241b-4020-87e6-08cfd5bec0cc-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.259806 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.259668 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/49f08cb5-241b-4020-87e6-08cfd5bec0cc-web-config\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.259968 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.259857 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/49f08cb5-241b-4020-87e6-08cfd5bec0cc-cluster-tls-config\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.260022 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.259986 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/49f08cb5-241b-4020-87e6-08cfd5bec0cc-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.260122 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.260085 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/49f08cb5-241b-4020-87e6-08cfd5bec0cc-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.260329 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.260313 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/49f08cb5-241b-4020-87e6-08cfd5bec0cc-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.260403 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.260341 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/49f08cb5-241b-4020-87e6-08cfd5bec0cc-tls-assets\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.260730 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.260704 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49f08cb5-241b-4020-87e6-08cfd5bec0cc-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.261845 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.261825 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/49f08cb5-241b-4020-87e6-08cfd5bec0cc-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.263181 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.263159 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwmnc\" (UniqueName: \"kubernetes.io/projected/49f08cb5-241b-4020-87e6-08cfd5bec0cc-kube-api-access-lwmnc\") pod \"alertmanager-main-0\" (UID: \"49f08cb5-241b-4020-87e6-08cfd5bec0cc\") " pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.307438 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.307393 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Feb 17 12:50:56.432652 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.432629 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Feb 17 12:50:56.435593 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:50:56.435557 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49f08cb5_241b_4020_87e6_08cfd5bec0cc.slice/crio-986dd9092043c37af752d60c68f718b5723757ece74f460c6dd699ccd887933a WatchSource:0}: Error finding container 986dd9092043c37af752d60c68f718b5723757ece74f460c6dd699ccd887933a: Status 404 returned error can't find the container with id 986dd9092043c37af752d60c68f718b5723757ece74f460c6dd699ccd887933a Feb 17 12:50:56.942578 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.942547 2572 generic.go:358] "Generic (PLEG): container finished" podID="49f08cb5-241b-4020-87e6-08cfd5bec0cc" containerID="f87b915fae1743f5fe6f9da289794b2eecedb4013cddfda7b57726d36aeb2c28" exitCode=0 Feb 17 12:50:56.943004 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.942632 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"49f08cb5-241b-4020-87e6-08cfd5bec0cc","Type":"ContainerDied","Data":"f87b915fae1743f5fe6f9da289794b2eecedb4013cddfda7b57726d36aeb2c28"} Feb 17 12:50:56.943004 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:56.942667 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"49f08cb5-241b-4020-87e6-08cfd5bec0cc","Type":"ContainerStarted","Data":"986dd9092043c37af752d60c68f718b5723757ece74f460c6dd699ccd887933a"} Feb 17 12:50:57.469919 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.469725 2572 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Feb 17 12:50:57.470520 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.470253 2572 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-0" podUID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerName="prometheus" containerID="cri-o://0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3" gracePeriod=600 Feb 17 12:50:57.470520 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.470253 2572 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-0" podUID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerName="kube-rbac-proxy" containerID="cri-o://1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f" gracePeriod=600 Feb 17 12:50:57.470520 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.470298 2572 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-0" podUID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerName="config-reloader" containerID="cri-o://5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df" gracePeriod=600 Feb 17 12:50:57.470520 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.470298 2572 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-0" podUID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerName="thanos-sidecar" containerID="cri-o://68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf" gracePeriod=600 Feb 17 12:50:57.470520 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.470396 2572 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-0" podUID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerName="kube-rbac-proxy-thanos" containerID="cri-o://3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49" gracePeriod=600 Feb 17 12:50:57.470520 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.470410 2572 kuberuntime_container.go:864] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-0" podUID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerName="kube-rbac-proxy-web" containerID="cri-o://984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832" gracePeriod=600 Feb 17 12:50:57.716751 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.716726 2572 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:57.870223 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.870124 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " Feb 17 12:50:57.870223 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.870162 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-prometheus-k8s-tls\") pod \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " Feb 17 12:50:57.870223 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.870183 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/5d6ebab3-fa58-444e-8b5e-e9d52e188752-prometheus-k8s-db\") pod \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " Feb 17 12:50:57.870223 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.870200 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stdxg\" (UniqueName: \"kubernetes.io/projected/5d6ebab3-fa58-444e-8b5e-e9d52e188752-kube-api-access-stdxg\") pod \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " Feb 17 12:50:57.870223 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.870222 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5d6ebab3-fa58-444e-8b5e-e9d52e188752-config-out\") pod \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " Feb 17 12:50:57.870552 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.870242 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-prometheus-trusted-ca-bundle\") pod \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " Feb 17 12:50:57.870552 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.870272 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-thanos-prometheus-http-client-file\") pod \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " Feb 17 12:50:57.870552 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.870451 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-configmap-metrics-client-ca\") pod \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " Feb 17 12:50:57.870552 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.870509 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-web-config\") pod \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " Feb 17 12:50:57.870552 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.870535 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-config\") pod \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " Feb 17 12:50:57.870754 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.870601 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-prometheus-k8s-rulefiles-0\") pod \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " Feb 17 12:50:57.870754 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.870633 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5d6ebab3-fa58-444e-8b5e-e9d52e188752-tls-assets\") pod \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " Feb 17 12:50:57.870754 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.870681 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-grpc-tls\") pod \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " Feb 17 12:50:57.870754 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.870709 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " Feb 17 12:50:57.870962 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.870759 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-configmap-kubelet-serving-ca-bundle\") pod \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " Feb 17 12:50:57.870962 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.870788 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-kube-rbac-proxy\") pod \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " Feb 17 12:50:57.870962 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.870842 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-configmap-serving-certs-ca-bundle\") pod \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " Feb 17 12:50:57.870962 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.870867 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-metrics-client-certs\") pod \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\" (UID: \"5d6ebab3-fa58-444e-8b5e-e9d52e188752\") " Feb 17 12:50:57.870962 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.870941 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-prometheus-trusted-ca-bundle" (OuterVolumeSpecName: "prometheus-trusted-ca-bundle") pod "5d6ebab3-fa58-444e-8b5e-e9d52e188752" (UID: "5d6ebab3-fa58-444e-8b5e-e9d52e188752"). InnerVolumeSpecName "prometheus-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Feb 17 12:50:57.871221 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.871185 2572 reconciler_common.go:299] "Volume detached for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-prometheus-trusted-ca-bundle\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:57.872280 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.871616 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d6ebab3-fa58-444e-8b5e-e9d52e188752-prometheus-k8s-db" (OuterVolumeSpecName: "prometheus-k8s-db") pod "5d6ebab3-fa58-444e-8b5e-e9d52e188752" (UID: "5d6ebab3-fa58-444e-8b5e-e9d52e188752"). InnerVolumeSpecName "prometheus-k8s-db". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Feb 17 12:50:57.872280 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.871916 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-prometheus-k8s-rulefiles-0" (OuterVolumeSpecName: "prometheus-k8s-rulefiles-0") pod "5d6ebab3-fa58-444e-8b5e-e9d52e188752" (UID: "5d6ebab3-fa58-444e-8b5e-e9d52e188752"). InnerVolumeSpecName "prometheus-k8s-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Feb 17 12:50:57.872280 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.872015 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-configmap-metrics-client-ca" (OuterVolumeSpecName: "configmap-metrics-client-ca") pod "5d6ebab3-fa58-444e-8b5e-e9d52e188752" (UID: "5d6ebab3-fa58-444e-8b5e-e9d52e188752"). InnerVolumeSpecName "configmap-metrics-client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Feb 17 12:50:57.872616 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.872559 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-configmap-serving-certs-ca-bundle" (OuterVolumeSpecName: "configmap-serving-certs-ca-bundle") pod "5d6ebab3-fa58-444e-8b5e-e9d52e188752" (UID: "5d6ebab3-fa58-444e-8b5e-e9d52e188752"). InnerVolumeSpecName "configmap-serving-certs-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Feb 17 12:50:57.873616 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.873584 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-configmap-kubelet-serving-ca-bundle" (OuterVolumeSpecName: "configmap-kubelet-serving-ca-bundle") pod "5d6ebab3-fa58-444e-8b5e-e9d52e188752" (UID: "5d6ebab3-fa58-444e-8b5e-e9d52e188752"). InnerVolumeSpecName "configmap-kubelet-serving-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Feb 17 12:50:57.873729 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.873671 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-metrics-client-certs" (OuterVolumeSpecName: "secret-metrics-client-certs") pod "5d6ebab3-fa58-444e-8b5e-e9d52e188752" (UID: "5d6ebab3-fa58-444e-8b5e-e9d52e188752"). InnerVolumeSpecName "secret-metrics-client-certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" Feb 17 12:50:57.873790 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.873754 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "5d6ebab3-fa58-444e-8b5e-e9d52e188752" (UID: "5d6ebab3-fa58-444e-8b5e-e9d52e188752"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGIDValue "" Feb 17 12:50:57.873916 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.873884 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-prometheus-k8s-thanos-sidecar-tls" (OuterVolumeSpecName: "secret-prometheus-k8s-thanos-sidecar-tls") pod "5d6ebab3-fa58-444e-8b5e-e9d52e188752" (UID: "5d6ebab3-fa58-444e-8b5e-e9d52e188752"). InnerVolumeSpecName "secret-prometheus-k8s-thanos-sidecar-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Feb 17 12:50:57.874255 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.874201 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d6ebab3-fa58-444e-8b5e-e9d52e188752-config-out" (OuterVolumeSpecName: "config-out") pod "5d6ebab3-fa58-444e-8b5e-e9d52e188752" (UID: "5d6ebab3-fa58-444e-8b5e-e9d52e188752"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Feb 17 12:50:57.874584 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.874554 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-prometheus-k8s-kube-rbac-proxy-web" (OuterVolumeSpecName: "secret-prometheus-k8s-kube-rbac-proxy-web") pod "5d6ebab3-fa58-444e-8b5e-e9d52e188752" (UID: "5d6ebab3-fa58-444e-8b5e-e9d52e188752"). InnerVolumeSpecName "secret-prometheus-k8s-kube-rbac-proxy-web". PluginName "kubernetes.io/secret", VolumeGIDValue "" Feb 17 12:50:57.874981 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.874957 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-prometheus-k8s-tls" (OuterVolumeSpecName: "secret-prometheus-k8s-tls") pod "5d6ebab3-fa58-444e-8b5e-e9d52e188752" (UID: "5d6ebab3-fa58-444e-8b5e-e9d52e188752"). InnerVolumeSpecName "secret-prometheus-k8s-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Feb 17 12:50:57.875142 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.875119 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d6ebab3-fa58-444e-8b5e-e9d52e188752-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "5d6ebab3-fa58-444e-8b5e-e9d52e188752" (UID: "5d6ebab3-fa58-444e-8b5e-e9d52e188752"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGIDValue "" Feb 17 12:50:57.875357 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.875340 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-config" (OuterVolumeSpecName: "config") pod "5d6ebab3-fa58-444e-8b5e-e9d52e188752" (UID: "5d6ebab3-fa58-444e-8b5e-e9d52e188752"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Feb 17 12:50:57.875701 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.875673 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d6ebab3-fa58-444e-8b5e-e9d52e188752-kube-api-access-stdxg" (OuterVolumeSpecName: "kube-api-access-stdxg") pod "5d6ebab3-fa58-444e-8b5e-e9d52e188752" (UID: "5d6ebab3-fa58-444e-8b5e-e9d52e188752"). InnerVolumeSpecName "kube-api-access-stdxg". PluginName "kubernetes.io/projected", VolumeGIDValue "" Feb 17 12:50:57.875792 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.875707 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-kube-rbac-proxy" (OuterVolumeSpecName: "secret-kube-rbac-proxy") pod "5d6ebab3-fa58-444e-8b5e-e9d52e188752" (UID: "5d6ebab3-fa58-444e-8b5e-e9d52e188752"). InnerVolumeSpecName "secret-kube-rbac-proxy". PluginName "kubernetes.io/secret", VolumeGIDValue "" Feb 17 12:50:57.875837 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.875793 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-grpc-tls" (OuterVolumeSpecName: "secret-grpc-tls") pod "5d6ebab3-fa58-444e-8b5e-e9d52e188752" (UID: "5d6ebab3-fa58-444e-8b5e-e9d52e188752"). InnerVolumeSpecName "secret-grpc-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Feb 17 12:50:57.884757 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.884736 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-web-config" (OuterVolumeSpecName: "web-config") pod "5d6ebab3-fa58-444e-8b5e-e9d52e188752" (UID: "5d6ebab3-fa58-444e-8b5e-e9d52e188752"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Feb 17 12:50:57.948958 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.948925 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"49f08cb5-241b-4020-87e6-08cfd5bec0cc","Type":"ContainerStarted","Data":"b5e941750ff5596b36cf5611de2dc33e74bca83e51317b8e6e9a903b2af9a072"} Feb 17 12:50:57.948958 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.948966 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"49f08cb5-241b-4020-87e6-08cfd5bec0cc","Type":"ContainerStarted","Data":"0853c1fe91cf0a2a555ba7799d20f7814b87bbcd4bc23e38c87b084dea5b056a"} Feb 17 12:50:57.949471 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.948978 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"49f08cb5-241b-4020-87e6-08cfd5bec0cc","Type":"ContainerStarted","Data":"aa6fb35a1dfbc7746c7823e9c8cad6658fcb4c60d893a0c47ae9ce6cfacde52b"} Feb 17 12:50:57.949471 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.948986 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"49f08cb5-241b-4020-87e6-08cfd5bec0cc","Type":"ContainerStarted","Data":"64c96a00bf206178f8f8319a06d1dd1a3541c11af1e297078caacd3c200cd697"} Feb 17 12:50:57.949471 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.948995 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"49f08cb5-241b-4020-87e6-08cfd5bec0cc","Type":"ContainerStarted","Data":"86f1985cbc2fc95425cdc7da3c7208261de7e03ae80f89c7905a6ad2c4e2db3a"} Feb 17 12:50:57.949471 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.949002 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"49f08cb5-241b-4020-87e6-08cfd5bec0cc","Type":"ContainerStarted","Data":"0d2ab20ac97dd00f7718e299d7b241da4361d4e5a61a83a5838379ea37d48fee"} Feb 17 12:50:57.951610 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.951586 2572 generic.go:358] "Generic (PLEG): container finished" podID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerID="3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49" exitCode=0 Feb 17 12:50:57.951610 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.951607 2572 generic.go:358] "Generic (PLEG): container finished" podID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerID="1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f" exitCode=0 Feb 17 12:50:57.951778 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.951615 2572 generic.go:358] "Generic (PLEG): container finished" podID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerID="984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832" exitCode=0 Feb 17 12:50:57.951778 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.951622 2572 generic.go:358] "Generic (PLEG): container finished" podID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerID="68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf" exitCode=0 Feb 17 12:50:57.951778 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.951627 2572 generic.go:358] "Generic (PLEG): container finished" podID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerID="5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df" exitCode=0 Feb 17 12:50:57.951778 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.951632 2572 generic.go:358] "Generic (PLEG): container finished" podID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerID="0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3" exitCode=0 Feb 17 12:50:57.951778 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.951669 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"5d6ebab3-fa58-444e-8b5e-e9d52e188752","Type":"ContainerDied","Data":"3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49"} Feb 17 12:50:57.951778 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.951709 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"5d6ebab3-fa58-444e-8b5e-e9d52e188752","Type":"ContainerDied","Data":"1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f"} Feb 17 12:50:57.951778 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.951708 2572 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:57.951778 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.951720 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"5d6ebab3-fa58-444e-8b5e-e9d52e188752","Type":"ContainerDied","Data":"984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832"} Feb 17 12:50:57.951778 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.951731 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"5d6ebab3-fa58-444e-8b5e-e9d52e188752","Type":"ContainerDied","Data":"68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf"} Feb 17 12:50:57.951778 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.951740 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"5d6ebab3-fa58-444e-8b5e-e9d52e188752","Type":"ContainerDied","Data":"5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df"} Feb 17 12:50:57.951778 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.951751 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"5d6ebab3-fa58-444e-8b5e-e9d52e188752","Type":"ContainerDied","Data":"0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3"} Feb 17 12:50:57.951778 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.951761 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"5d6ebab3-fa58-444e-8b5e-e9d52e188752","Type":"ContainerDied","Data":"eb814d2a53086156ca7779bf67372e992ce49afeb4608f64f1d0f216d0d98e1b"} Feb 17 12:50:57.951778 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.951776 2572 scope.go:117] "RemoveContainer" containerID="3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49" Feb 17 12:50:57.959739 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.959715 2572 scope.go:117] "RemoveContainer" containerID="1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f" Feb 17 12:50:57.966904 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.966877 2572 scope.go:117] "RemoveContainer" containerID="984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832" Feb 17 12:50:57.970817 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.970734 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/alertmanager-main-0" podStartSLOduration=2.9707176669999997 podStartE2EDuration="2.970717667s" podCreationTimestamp="2026-02-17 12:50:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 12:50:57.970159336 +0000 UTC m=+264.320570150" watchObservedRunningTime="2026-02-17 12:50:57.970717667 +0000 UTC m=+264.321128477" Feb 17 12:50:57.971985 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.971801 2572 reconciler_common.go:299] "Volume detached for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-configmap-kubelet-serving-ca-bundle\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:57.971985 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.971835 2572 reconciler_common.go:299] "Volume detached for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-kube-rbac-proxy\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:57.971985 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.971853 2572 reconciler_common.go:299] "Volume detached for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-configmap-serving-certs-ca-bundle\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:57.971985 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.971870 2572 reconciler_common.go:299] "Volume detached for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-metrics-client-certs\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:57.971985 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.971887 2572 reconciler_common.go:299] "Volume detached for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-prometheus-k8s-thanos-sidecar-tls\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:57.971985 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.971903 2572 reconciler_common.go:299] "Volume detached for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-prometheus-k8s-tls\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:57.971985 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.971916 2572 reconciler_common.go:299] "Volume detached for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/5d6ebab3-fa58-444e-8b5e-e9d52e188752-prometheus-k8s-db\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:57.971985 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.971929 2572 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-stdxg\" (UniqueName: \"kubernetes.io/projected/5d6ebab3-fa58-444e-8b5e-e9d52e188752-kube-api-access-stdxg\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:57.971985 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.971942 2572 reconciler_common.go:299] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5d6ebab3-fa58-444e-8b5e-e9d52e188752-config-out\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:57.971985 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.971957 2572 reconciler_common.go:299] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-thanos-prometheus-http-client-file\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:57.971985 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.971970 2572 reconciler_common.go:299] "Volume detached for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-configmap-metrics-client-ca\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:57.971985 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.971985 2572 reconciler_common.go:299] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-web-config\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:57.972400 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.971999 2572 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-config\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:57.972400 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.972012 2572 reconciler_common.go:299] "Volume detached for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/5d6ebab3-fa58-444e-8b5e-e9d52e188752-prometheus-k8s-rulefiles-0\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:57.972400 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.972024 2572 reconciler_common.go:299] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5d6ebab3-fa58-444e-8b5e-e9d52e188752-tls-assets\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:57.972400 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.972037 2572 reconciler_common.go:299] "Volume detached for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-grpc-tls\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:57.972400 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.972049 2572 reconciler_common.go:299] "Volume detached for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/5d6ebab3-fa58-444e-8b5e-e9d52e188752-secret-prometheus-k8s-kube-rbac-proxy-web\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:50:57.974641 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.974616 2572 scope.go:117] "RemoveContainer" containerID="68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf" Feb 17 12:50:57.982077 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.982050 2572 scope.go:117] "RemoveContainer" containerID="5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df" Feb 17 12:50:57.988630 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.988519 2572 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Feb 17 12:50:57.991950 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.991917 2572 scope.go:117] "RemoveContainer" containerID="0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3" Feb 17 12:50:57.993385 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:57.993364 2572 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Feb 17 12:50:58.000923 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.000900 2572 scope.go:117] "RemoveContainer" containerID="6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d" Feb 17 12:50:58.008846 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.008827 2572 scope.go:117] "RemoveContainer" containerID="3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49" Feb 17 12:50:58.009185 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:50:58.009131 2572 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49\": container with ID starting with 3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49 not found: ID does not exist" containerID="3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49" Feb 17 12:50:58.009185 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.009168 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49"} err="failed to get container status \"3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49\": rpc error: code = NotFound desc = could not find container \"3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49\": container with ID starting with 3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49 not found: ID does not exist" Feb 17 12:50:58.009317 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.009189 2572 scope.go:117] "RemoveContainer" containerID="1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f" Feb 17 12:50:58.009455 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:50:58.009437 2572 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f\": container with ID starting with 1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f not found: ID does not exist" containerID="1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f" Feb 17 12:50:58.009495 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.009461 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f"} err="failed to get container status \"1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f\": rpc error: code = NotFound desc = could not find container \"1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f\": container with ID starting with 1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f not found: ID does not exist" Feb 17 12:50:58.009495 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.009484 2572 scope.go:117] "RemoveContainer" containerID="984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832" Feb 17 12:50:58.009735 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:50:58.009715 2572 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832\": container with ID starting with 984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832 not found: ID does not exist" containerID="984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832" Feb 17 12:50:58.009788 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.009740 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832"} err="failed to get container status \"984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832\": rpc error: code = NotFound desc = could not find container \"984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832\": container with ID starting with 984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832 not found: ID does not exist" Feb 17 12:50:58.009788 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.009755 2572 scope.go:117] "RemoveContainer" containerID="68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf" Feb 17 12:50:58.010018 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:50:58.010002 2572 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf\": container with ID starting with 68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf not found: ID does not exist" containerID="68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf" Feb 17 12:50:58.010054 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.010031 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf"} err="failed to get container status \"68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf\": rpc error: code = NotFound desc = could not find container \"68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf\": container with ID starting with 68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf not found: ID does not exist" Feb 17 12:50:58.010054 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.010046 2572 scope.go:117] "RemoveContainer" containerID="5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df" Feb 17 12:50:58.010308 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:50:58.010289 2572 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df\": container with ID starting with 5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df not found: ID does not exist" containerID="5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df" Feb 17 12:50:58.010386 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.010311 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df"} err="failed to get container status \"5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df\": rpc error: code = NotFound desc = could not find container \"5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df\": container with ID starting with 5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df not found: ID does not exist" Feb 17 12:50:58.010386 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.010327 2572 scope.go:117] "RemoveContainer" containerID="0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3" Feb 17 12:50:58.010581 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:50:58.010558 2572 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3\": container with ID starting with 0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3 not found: ID does not exist" containerID="0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3" Feb 17 12:50:58.010620 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.010590 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3"} err="failed to get container status \"0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3\": rpc error: code = NotFound desc = could not find container \"0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3\": container with ID starting with 0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3 not found: ID does not exist" Feb 17 12:50:58.010620 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.010612 2572 scope.go:117] "RemoveContainer" containerID="6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d" Feb 17 12:50:58.010858 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:50:58.010830 2572 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d\": container with ID starting with 6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d not found: ID does not exist" containerID="6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d" Feb 17 12:50:58.010946 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.010863 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d"} err="failed to get container status \"6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d\": rpc error: code = NotFound desc = could not find container \"6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d\": container with ID starting with 6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d not found: ID does not exist" Feb 17 12:50:58.010946 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.010893 2572 scope.go:117] "RemoveContainer" containerID="3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49" Feb 17 12:50:58.011167 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.011149 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49"} err="failed to get container status \"3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49\": rpc error: code = NotFound desc = could not find container \"3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49\": container with ID starting with 3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49 not found: ID does not exist" Feb 17 12:50:58.011218 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.011168 2572 scope.go:117] "RemoveContainer" containerID="1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f" Feb 17 12:50:58.011391 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.011372 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f"} err="failed to get container status \"1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f\": rpc error: code = NotFound desc = could not find container \"1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f\": container with ID starting with 1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f not found: ID does not exist" Feb 17 12:50:58.011391 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.011390 2572 scope.go:117] "RemoveContainer" containerID="984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832" Feb 17 12:50:58.011732 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.011712 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832"} err="failed to get container status \"984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832\": rpc error: code = NotFound desc = could not find container \"984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832\": container with ID starting with 984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832 not found: ID does not exist" Feb 17 12:50:58.011802 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.011734 2572 scope.go:117] "RemoveContainer" containerID="68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf" Feb 17 12:50:58.011960 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.011943 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf"} err="failed to get container status \"68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf\": rpc error: code = NotFound desc = could not find container \"68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf\": container with ID starting with 68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf not found: ID does not exist" Feb 17 12:50:58.012010 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.011960 2572 scope.go:117] "RemoveContainer" containerID="5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df" Feb 17 12:50:58.012195 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.012172 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df"} err="failed to get container status \"5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df\": rpc error: code = NotFound desc = could not find container \"5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df\": container with ID starting with 5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df not found: ID does not exist" Feb 17 12:50:58.012267 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.012196 2572 scope.go:117] "RemoveContainer" containerID="0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3" Feb 17 12:50:58.012439 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.012417 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3"} err="failed to get container status \"0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3\": rpc error: code = NotFound desc = could not find container \"0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3\": container with ID starting with 0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3 not found: ID does not exist" Feb 17 12:50:58.012482 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.012442 2572 scope.go:117] "RemoveContainer" containerID="6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d" Feb 17 12:50:58.012692 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.012673 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d"} err="failed to get container status \"6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d\": rpc error: code = NotFound desc = could not find container \"6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d\": container with ID starting with 6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d not found: ID does not exist" Feb 17 12:50:58.012692 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.012691 2572 scope.go:117] "RemoveContainer" containerID="3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49" Feb 17 12:50:58.012880 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.012866 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49"} err="failed to get container status \"3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49\": rpc error: code = NotFound desc = could not find container \"3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49\": container with ID starting with 3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49 not found: ID does not exist" Feb 17 12:50:58.012926 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.012881 2572 scope.go:117] "RemoveContainer" containerID="1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f" Feb 17 12:50:58.013089 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.013068 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f"} err="failed to get container status \"1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f\": rpc error: code = NotFound desc = could not find container \"1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f\": container with ID starting with 1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f not found: ID does not exist" Feb 17 12:50:58.013143 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.013090 2572 scope.go:117] "RemoveContainer" containerID="984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832" Feb 17 12:50:58.013343 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.013328 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832"} err="failed to get container status \"984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832\": rpc error: code = NotFound desc = could not find container \"984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832\": container with ID starting with 984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832 not found: ID does not exist" Feb 17 12:50:58.013343 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.013342 2572 scope.go:117] "RemoveContainer" containerID="68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf" Feb 17 12:50:58.013530 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.013514 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf"} err="failed to get container status \"68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf\": rpc error: code = NotFound desc = could not find container \"68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf\": container with ID starting with 68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf not found: ID does not exist" Feb 17 12:50:58.013577 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.013531 2572 scope.go:117] "RemoveContainer" containerID="5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df" Feb 17 12:50:58.013723 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.013704 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df"} err="failed to get container status \"5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df\": rpc error: code = NotFound desc = could not find container \"5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df\": container with ID starting with 5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df not found: ID does not exist" Feb 17 12:50:58.013765 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.013725 2572 scope.go:117] "RemoveContainer" containerID="0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3" Feb 17 12:50:58.013948 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.013924 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3"} err="failed to get container status \"0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3\": rpc error: code = NotFound desc = could not find container \"0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3\": container with ID starting with 0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3 not found: ID does not exist" Feb 17 12:50:58.013995 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.013950 2572 scope.go:117] "RemoveContainer" containerID="6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d" Feb 17 12:50:58.014197 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.014177 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d"} err="failed to get container status \"6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d\": rpc error: code = NotFound desc = could not find container \"6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d\": container with ID starting with 6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d not found: ID does not exist" Feb 17 12:50:58.014247 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.014201 2572 scope.go:117] "RemoveContainer" containerID="3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49" Feb 17 12:50:58.014421 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.014404 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49"} err="failed to get container status \"3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49\": rpc error: code = NotFound desc = could not find container \"3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49\": container with ID starting with 3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49 not found: ID does not exist" Feb 17 12:50:58.014473 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.014423 2572 scope.go:117] "RemoveContainer" containerID="1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f" Feb 17 12:50:58.014637 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.014619 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f"} err="failed to get container status \"1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f\": rpc error: code = NotFound desc = could not find container \"1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f\": container with ID starting with 1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f not found: ID does not exist" Feb 17 12:50:58.014682 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.014641 2572 scope.go:117] "RemoveContainer" containerID="984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832" Feb 17 12:50:58.014863 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.014842 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832"} err="failed to get container status \"984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832\": rpc error: code = NotFound desc = could not find container \"984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832\": container with ID starting with 984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832 not found: ID does not exist" Feb 17 12:50:58.014863 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.014863 2572 scope.go:117] "RemoveContainer" containerID="68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf" Feb 17 12:50:58.015073 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.015056 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf"} err="failed to get container status \"68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf\": rpc error: code = NotFound desc = could not find container \"68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf\": container with ID starting with 68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf not found: ID does not exist" Feb 17 12:50:58.015073 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.015072 2572 scope.go:117] "RemoveContainer" containerID="5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df" Feb 17 12:50:58.015272 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.015256 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df"} err="failed to get container status \"5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df\": rpc error: code = NotFound desc = could not find container \"5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df\": container with ID starting with 5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df not found: ID does not exist" Feb 17 12:50:58.015325 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.015273 2572 scope.go:117] "RemoveContainer" containerID="0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3" Feb 17 12:50:58.015497 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.015477 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3"} err="failed to get container status \"0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3\": rpc error: code = NotFound desc = could not find container \"0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3\": container with ID starting with 0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3 not found: ID does not exist" Feb 17 12:50:58.015552 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.015507 2572 scope.go:117] "RemoveContainer" containerID="6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d" Feb 17 12:50:58.015734 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.015717 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d"} err="failed to get container status \"6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d\": rpc error: code = NotFound desc = could not find container \"6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d\": container with ID starting with 6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d not found: ID does not exist" Feb 17 12:50:58.015782 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.015736 2572 scope.go:117] "RemoveContainer" containerID="3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49" Feb 17 12:50:58.015999 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.015981 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49"} err="failed to get container status \"3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49\": rpc error: code = NotFound desc = could not find container \"3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49\": container with ID starting with 3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49 not found: ID does not exist" Feb 17 12:50:58.016047 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.016000 2572 scope.go:117] "RemoveContainer" containerID="1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f" Feb 17 12:50:58.016243 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.016225 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f"} err="failed to get container status \"1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f\": rpc error: code = NotFound desc = could not find container \"1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f\": container with ID starting with 1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f not found: ID does not exist" Feb 17 12:50:58.016243 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.016241 2572 scope.go:117] "RemoveContainer" containerID="984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832" Feb 17 12:50:58.016445 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.016425 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832"} err="failed to get container status \"984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832\": rpc error: code = NotFound desc = could not find container \"984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832\": container with ID starting with 984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832 not found: ID does not exist" Feb 17 12:50:58.016517 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.016447 2572 scope.go:117] "RemoveContainer" containerID="68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf" Feb 17 12:50:58.016686 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.016670 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf"} err="failed to get container status \"68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf\": rpc error: code = NotFound desc = could not find container \"68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf\": container with ID starting with 68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf not found: ID does not exist" Feb 17 12:50:58.016686 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.016686 2572 scope.go:117] "RemoveContainer" containerID="5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df" Feb 17 12:50:58.016911 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.016894 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df"} err="failed to get container status \"5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df\": rpc error: code = NotFound desc = could not find container \"5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df\": container with ID starting with 5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df not found: ID does not exist" Feb 17 12:50:58.016911 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.016910 2572 scope.go:117] "RemoveContainer" containerID="0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3" Feb 17 12:50:58.017140 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.017096 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3"} err="failed to get container status \"0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3\": rpc error: code = NotFound desc = could not find container \"0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3\": container with ID starting with 0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3 not found: ID does not exist" Feb 17 12:50:58.017140 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.017139 2572 scope.go:117] "RemoveContainer" containerID="6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d" Feb 17 12:50:58.017304 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.017289 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d"} err="failed to get container status \"6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d\": rpc error: code = NotFound desc = could not find container \"6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d\": container with ID starting with 6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d not found: ID does not exist" Feb 17 12:50:58.017367 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.017304 2572 scope.go:117] "RemoveContainer" containerID="3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49" Feb 17 12:50:58.017504 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.017485 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49"} err="failed to get container status \"3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49\": rpc error: code = NotFound desc = could not find container \"3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49\": container with ID starting with 3f0f403c18e8d7cb812a53a467acdf3f6f14434a6919751c5ab5b4e6edd10c49 not found: ID does not exist" Feb 17 12:50:58.017570 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.017507 2572 scope.go:117] "RemoveContainer" containerID="1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f" Feb 17 12:50:58.017740 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.017709 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f"} err="failed to get container status \"1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f\": rpc error: code = NotFound desc = could not find container \"1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f\": container with ID starting with 1d7f8f8769d8f8cf4da3bee44c978e23445b8440379f10d0f48ca06a9a020d7f not found: ID does not exist" Feb 17 12:50:58.017740 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.017738 2572 scope.go:117] "RemoveContainer" containerID="984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832" Feb 17 12:50:58.018025 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.017997 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832"} err="failed to get container status \"984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832\": rpc error: code = NotFound desc = could not find container \"984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832\": container with ID starting with 984f9883bfe4c87654526754349db62e1bd5727a4b4e26b4fd8512078d3ee832 not found: ID does not exist" Feb 17 12:50:58.018025 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.018025 2572 scope.go:117] "RemoveContainer" containerID="68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf" Feb 17 12:50:58.018416 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.018309 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf"} err="failed to get container status \"68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf\": rpc error: code = NotFound desc = could not find container \"68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf\": container with ID starting with 68b324c87797bdb0b3101dbcb89af414822b491a70e332c6941c03d585fd3dbf not found: ID does not exist" Feb 17 12:50:58.018416 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.018346 2572 scope.go:117] "RemoveContainer" containerID="5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df" Feb 17 12:50:58.018944 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.018844 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df"} err="failed to get container status \"5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df\": rpc error: code = NotFound desc = could not find container \"5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df\": container with ID starting with 5d745d110c2fc9595c4cc3cd64413eb187ced7b4a6cf845df9b0ef1a5ebd78df not found: ID does not exist" Feb 17 12:50:58.018944 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.018868 2572 scope.go:117] "RemoveContainer" containerID="0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3" Feb 17 12:50:58.019167 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.019137 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3"} err="failed to get container status \"0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3\": rpc error: code = NotFound desc = could not find container \"0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3\": container with ID starting with 0bf104d953aa3f7501ebb99615486349cf3caa33baa75c8ea9f6a1f253ac0df3 not found: ID does not exist" Feb 17 12:50:58.019167 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.019163 2572 scope.go:117] "RemoveContainer" containerID="6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d" Feb 17 12:50:58.019432 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.019410 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d"} err="failed to get container status \"6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d\": rpc error: code = NotFound desc = could not find container \"6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d\": container with ID starting with 6df5199066881073d0d984d8d2579621d94016752845d0de1740fb9577d1f40d not found: ID does not exist" Feb 17 12:50:58.019805 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.019788 2572 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Feb 17 12:50:58.020150 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.020136 2572 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerName="kube-rbac-proxy-web" Feb 17 12:50:58.020194 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.020152 2572 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerName="kube-rbac-proxy-web" Feb 17 12:50:58.020194 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.020165 2572 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerName="init-config-reloader" Feb 17 12:50:58.020194 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.020170 2572 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerName="init-config-reloader" Feb 17 12:50:58.020194 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.020178 2572 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerName="thanos-sidecar" Feb 17 12:50:58.020194 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.020183 2572 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerName="thanos-sidecar" Feb 17 12:50:58.020194 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.020191 2572 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerName="kube-rbac-proxy" Feb 17 12:50:58.020360 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.020199 2572 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerName="kube-rbac-proxy" Feb 17 12:50:58.020360 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.020208 2572 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerName="kube-rbac-proxy-thanos" Feb 17 12:50:58.020360 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.020214 2572 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerName="kube-rbac-proxy-thanos" Feb 17 12:50:58.020360 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.020221 2572 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerName="config-reloader" Feb 17 12:50:58.020360 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.020226 2572 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerName="config-reloader" Feb 17 12:50:58.020360 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.020235 2572 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerName="prometheus" Feb 17 12:50:58.020360 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.020239 2572 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerName="prometheus" Feb 17 12:50:58.020360 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.020300 2572 memory_manager.go:356] "RemoveStaleState removing state" podUID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerName="kube-rbac-proxy" Feb 17 12:50:58.020360 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.020309 2572 memory_manager.go:356] "RemoveStaleState removing state" podUID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerName="kube-rbac-proxy-thanos" Feb 17 12:50:58.020360 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.020316 2572 memory_manager.go:356] "RemoveStaleState removing state" podUID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerName="kube-rbac-proxy-web" Feb 17 12:50:58.020360 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.020322 2572 memory_manager.go:356] "RemoveStaleState removing state" podUID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerName="prometheus" Feb 17 12:50:58.020360 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.020327 2572 memory_manager.go:356] "RemoveStaleState removing state" podUID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerName="thanos-sidecar" Feb 17 12:50:58.020360 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.020334 2572 memory_manager.go:356] "RemoveStaleState removing state" podUID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" containerName="config-reloader" Feb 17 12:50:58.026812 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.026360 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.028563 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.028541 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-web-config\"" Feb 17 12:50:58.028669 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.028660 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"serving-certs-ca-bundle\"" Feb 17 12:50:58.028744 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.028728 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-thanos-sidecar-tls\"" Feb 17 12:50:58.028848 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.028813 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s\"" Feb 17 12:50:58.028910 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.028876 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-grpc-tls-5l1v479posc61\"" Feb 17 12:50:58.029183 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.029164 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-thanos-prometheus-http-client-file\"" Feb 17 12:50:58.029314 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.029292 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-tls-assets-0\"" Feb 17 12:50:58.029762 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.029743 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-tls\"" Feb 17 12:50:58.029861 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.029830 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"metrics-client-certs\"" Feb 17 12:50:58.029924 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.029883 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"kube-rbac-proxy\"" Feb 17 12:50:58.029924 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.029883 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-kube-rbac-proxy-web\"" Feb 17 12:50:58.030360 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.030339 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-dockercfg-wdjcs\"" Feb 17 12:50:58.031743 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.031723 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"prometheus-k8s-rulefiles-0\"" Feb 17 12:50:58.034799 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.034524 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Feb 17 12:50:58.035695 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.035674 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-monitoring\"/\"prometheus-trusted-ca-bundle\"" Feb 17 12:50:58.131989 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.131902 2572 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d6ebab3-fa58-444e-8b5e-e9d52e188752" path="/var/lib/kubelet/pods/5d6ebab3-fa58-444e-8b5e-e9d52e188752/volumes" Feb 17 12:50:58.173765 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.173730 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.173765 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.173767 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.173981 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.173788 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.173981 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.173806 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.173981 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.173836 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.173981 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.173853 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-config\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.173981 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.173867 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-web-config\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.173981 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.173972 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.174298 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.174044 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.174298 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.174077 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.174298 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.174133 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb9ct\" (UniqueName: \"kubernetes.io/projected/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-kube-api-access-gb9ct\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.174298 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.174167 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.174298 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.174195 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.174298 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.174217 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.174298 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.174248 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-config-out\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.174298 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.174275 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.174298 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.174299 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.174581 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.174321 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.274850 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.274812 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.275027 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.274858 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.275027 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.274899 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gb9ct\" (UniqueName: \"kubernetes.io/projected/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-kube-api-access-gb9ct\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.275027 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.274926 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.275027 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.274950 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.275242 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.275199 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.275288 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.275251 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-config-out\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.275334 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.275297 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.275334 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.275324 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.275425 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.275351 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.275425 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.275401 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.275511 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.275426 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.275511 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.275466 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.275511 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.275492 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.275658 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.275547 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.275658 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.275577 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-config\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.275658 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.275604 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-web-config\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.275802 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.275685 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.276563 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.275902 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.276563 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.276060 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.283681 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.277493 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.283681 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.280340 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-config-out\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.283681 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.280927 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.283681 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.281061 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.283681 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.281517 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.283681 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.281581 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.283681 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.281925 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.283681 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.282083 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.283681 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.282405 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.283681 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.282408 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.283681 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.283172 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.284313 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.283718 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb9ct\" (UniqueName: \"kubernetes.io/projected/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-kube-api-access-gb9ct\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.284377 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.284315 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.286279 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.286135 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-web-config\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.286642 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.286619 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-config\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.287840 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.287743 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232\") " pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.337852 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.337823 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:50:58.458818 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.458761 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Feb 17 12:50:58.462431 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:50:58.462404 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8fc65e67_6fd7_44fd_a8e8_ad1e3ada3232.slice/crio-ff5eff22899faae00b20520ac10a7cf9a5ae5822400abf394a1c605d2ee337d9 WatchSource:0}: Error finding container ff5eff22899faae00b20520ac10a7cf9a5ae5822400abf394a1c605d2ee337d9: Status 404 returned error can't find the container with id ff5eff22899faae00b20520ac10a7cf9a5ae5822400abf394a1c605d2ee337d9 Feb 17 12:50:58.956001 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.955963 2572 generic.go:358] "Generic (PLEG): container finished" podID="8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232" containerID="35dba6db2a0128a939d9644c3aa51ff4330ed8a06b671b81d4f27dfb910fe9d5" exitCode=0 Feb 17 12:50:58.956459 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.956037 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232","Type":"ContainerDied","Data":"35dba6db2a0128a939d9644c3aa51ff4330ed8a06b671b81d4f27dfb910fe9d5"} Feb 17 12:50:58.956459 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:58.956068 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232","Type":"ContainerStarted","Data":"ff5eff22899faae00b20520ac10a7cf9a5ae5822400abf394a1c605d2ee337d9"} Feb 17 12:50:59.963559 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:59.963523 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232","Type":"ContainerStarted","Data":"588bbdae6248b0cd83f32b997ec62adaa8ed97737d5fb2131a0b7c2615fad269"} Feb 17 12:50:59.963936 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:59.963566 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232","Type":"ContainerStarted","Data":"b198fc6e8acd41292fd2fd5cabd9f0388324dfc6fb84e69071be3882c4c1a6ea"} Feb 17 12:50:59.963936 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:59.963580 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232","Type":"ContainerStarted","Data":"6108212691d30f0a49c9e5f4cf74d99b8e06f5443fa0b0cb9dd41d83b0452d7c"} Feb 17 12:50:59.963936 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:59.963594 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232","Type":"ContainerStarted","Data":"239ad18053d4d65e2387c32ca01a687470db0cf4f56768acad5588a7fe0d3ea4"} Feb 17 12:50:59.963936 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:59.963606 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232","Type":"ContainerStarted","Data":"dbad2c5283ca475d01613546d059452314d5e6b80e6cb2e70f1d4e57bd86d89a"} Feb 17 12:50:59.963936 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:59.963618 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232","Type":"ContainerStarted","Data":"e53e5f41a2b41795bdca38a78892a2da83683949d195ed9449666b9fa1be400f"} Feb 17 12:50:59.986694 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:50:59.986652 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-k8s-0" podStartSLOduration=1.986637545 podStartE2EDuration="1.986637545s" podCreationTimestamp="2026-02-17 12:50:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 12:50:59.984400028 +0000 UTC m=+266.334810840" watchObservedRunningTime="2026-02-17 12:50:59.986637545 +0000 UTC m=+266.337048401" Feb 17 12:51:03.338166 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:51:03.338136 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:51:34.088574 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:51:34.088542 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lb8m9_01eb7d37-a76a-4933-87df-44cf8a8d3152/ovn-acl-logging/0.log" Feb 17 12:51:34.089977 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:51:34.089946 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lb8m9_01eb7d37-a76a-4933-87df-44cf8a8d3152/ovn-acl-logging/0.log" Feb 17 12:51:34.092796 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:51:34.092779 2572 kubelet.go:1628] "Image garbage collection succeeded" Feb 17 12:51:40.324532 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:51:40.324494 2572 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["kube-system/global-pull-secret-syncer-b4nbx"] Feb 17 12:51:40.327623 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:51:40.327606 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-b4nbx" Feb 17 12:51:40.329556 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:51:40.329540 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"kube-system\"/\"original-pull-secret\"" Feb 17 12:51:40.332951 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:51:40.332928 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kube-system/global-pull-secret-syncer-b4nbx"] Feb 17 12:51:40.427511 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:51:40.427475 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/fb40dc60-7177-4037-a5d5-7cb6434ea4f5-dbus\") pod \"global-pull-secret-syncer-b4nbx\" (UID: \"fb40dc60-7177-4037-a5d5-7cb6434ea4f5\") " pod="kube-system/global-pull-secret-syncer-b4nbx" Feb 17 12:51:40.427671 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:51:40.427526 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/fb40dc60-7177-4037-a5d5-7cb6434ea4f5-original-pull-secret\") pod \"global-pull-secret-syncer-b4nbx\" (UID: \"fb40dc60-7177-4037-a5d5-7cb6434ea4f5\") " pod="kube-system/global-pull-secret-syncer-b4nbx" Feb 17 12:51:40.427671 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:51:40.427659 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/fb40dc60-7177-4037-a5d5-7cb6434ea4f5-kubelet-config\") pod \"global-pull-secret-syncer-b4nbx\" (UID: \"fb40dc60-7177-4037-a5d5-7cb6434ea4f5\") " pod="kube-system/global-pull-secret-syncer-b4nbx" Feb 17 12:51:40.528238 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:51:40.528204 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/fb40dc60-7177-4037-a5d5-7cb6434ea4f5-kubelet-config\") pod \"global-pull-secret-syncer-b4nbx\" (UID: \"fb40dc60-7177-4037-a5d5-7cb6434ea4f5\") " pod="kube-system/global-pull-secret-syncer-b4nbx" Feb 17 12:51:40.528238 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:51:40.528242 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/fb40dc60-7177-4037-a5d5-7cb6434ea4f5-dbus\") pod \"global-pull-secret-syncer-b4nbx\" (UID: \"fb40dc60-7177-4037-a5d5-7cb6434ea4f5\") " pod="kube-system/global-pull-secret-syncer-b4nbx" Feb 17 12:51:40.528443 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:51:40.528266 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/fb40dc60-7177-4037-a5d5-7cb6434ea4f5-original-pull-secret\") pod \"global-pull-secret-syncer-b4nbx\" (UID: \"fb40dc60-7177-4037-a5d5-7cb6434ea4f5\") " pod="kube-system/global-pull-secret-syncer-b4nbx" Feb 17 12:51:40.528443 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:51:40.528341 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-config\" (UniqueName: \"kubernetes.io/host-path/fb40dc60-7177-4037-a5d5-7cb6434ea4f5-kubelet-config\") pod \"global-pull-secret-syncer-b4nbx\" (UID: \"fb40dc60-7177-4037-a5d5-7cb6434ea4f5\") " pod="kube-system/global-pull-secret-syncer-b4nbx" Feb 17 12:51:40.528443 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:51:40.528389 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"dbus\" (UniqueName: \"kubernetes.io/host-path/fb40dc60-7177-4037-a5d5-7cb6434ea4f5-dbus\") pod \"global-pull-secret-syncer-b4nbx\" (UID: \"fb40dc60-7177-4037-a5d5-7cb6434ea4f5\") " pod="kube-system/global-pull-secret-syncer-b4nbx" Feb 17 12:51:40.530566 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:51:40.530545 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"original-pull-secret\" (UniqueName: \"kubernetes.io/secret/fb40dc60-7177-4037-a5d5-7cb6434ea4f5-original-pull-secret\") pod \"global-pull-secret-syncer-b4nbx\" (UID: \"fb40dc60-7177-4037-a5d5-7cb6434ea4f5\") " pod="kube-system/global-pull-secret-syncer-b4nbx" Feb 17 12:51:40.637498 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:51:40.637406 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/global-pull-secret-syncer-b4nbx" Feb 17 12:51:40.753619 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:51:40.753584 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["kube-system/global-pull-secret-syncer-b4nbx"] Feb 17 12:51:40.756807 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:51:40.756780 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb40dc60_7177_4037_a5d5_7cb6434ea4f5.slice/crio-5d0bc481b04e936d0a74ae032b02aa5dcd5b567f67dfee8eb50d582ebf42d537 WatchSource:0}: Error finding container 5d0bc481b04e936d0a74ae032b02aa5dcd5b567f67dfee8eb50d582ebf42d537: Status 404 returned error can't find the container with id 5d0bc481b04e936d0a74ae032b02aa5dcd5b567f67dfee8eb50d582ebf42d537 Feb 17 12:51:40.758435 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:51:40.758411 2572 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 17 12:51:41.082071 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:51:41.082033 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/global-pull-secret-syncer-b4nbx" event={"ID":"fb40dc60-7177-4037-a5d5-7cb6434ea4f5","Type":"ContainerStarted","Data":"5d0bc481b04e936d0a74ae032b02aa5dcd5b567f67dfee8eb50d582ebf42d537"} Feb 17 12:51:45.094986 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:51:45.094952 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="kube-system/global-pull-secret-syncer-b4nbx" event={"ID":"fb40dc60-7177-4037-a5d5-7cb6434ea4f5","Type":"ContainerStarted","Data":"3d462e07b9927f7e9723e379314c9619e42587f24bf0845951e9f729afb86625"} Feb 17 12:51:45.110332 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:51:45.110281 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/global-pull-secret-syncer-b4nbx" podStartSLOduration=1.719181179 podStartE2EDuration="5.110267319s" podCreationTimestamp="2026-02-17 12:51:40 +0000 UTC" firstStartedPulling="2026-02-17 12:51:40.75853375 +0000 UTC m=+307.108944540" lastFinishedPulling="2026-02-17 12:51:44.149619887 +0000 UTC m=+310.500030680" observedRunningTime="2026-02-17 12:51:45.108221009 +0000 UTC m=+311.458631820" watchObservedRunningTime="2026-02-17 12:51:45.110267319 +0000 UTC m=+311.460678131" Feb 17 12:51:58.338534 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:51:58.338500 2572 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:51:58.354125 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:51:58.354081 2572 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:51:59.148603 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:51:59.148573 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-0" Feb 17 12:56:34.112247 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:56:34.112218 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lb8m9_01eb7d37-a76a-4933-87df-44cf8a8d3152/ovn-acl-logging/0.log" Feb 17 12:56:34.113286 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:56:34.113264 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lb8m9_01eb7d37-a76a-4933-87df-44cf8a8d3152/ovn-acl-logging/0.log" Feb 17 12:57:41.126629 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:57:41.126595 2572 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["test-ns-feast/redis-6f5dff5888-l84dj"] Feb 17 12:57:41.131012 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:57:41.130991 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="test-ns-feast/redis-6f5dff5888-l84dj" Feb 17 12:57:41.133802 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:57:41.133778 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"test-ns-feast\"/\"openshift-service-ca.crt\"" Feb 17 12:57:41.134051 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:57:41.133788 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"test-ns-feast\"/\"kube-root-ca.crt\"" Feb 17 12:57:41.134236 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:57:41.133846 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"test-ns-feast\"/\"default-dockercfg-tgwxj\"" Feb 17 12:57:41.135194 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:57:41.135169 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["test-ns-feast/redis-6f5dff5888-l84dj"] Feb 17 12:57:41.285599 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:57:41.285551 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2wsd\" (UniqueName: \"kubernetes.io/projected/16490934-5213-4454-a1fb-3710756af56b-kube-api-access-l2wsd\") pod \"redis-6f5dff5888-l84dj\" (UID: \"16490934-5213-4454-a1fb-3710756af56b\") " pod="test-ns-feast/redis-6f5dff5888-l84dj" Feb 17 12:57:41.386956 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:57:41.386871 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l2wsd\" (UniqueName: \"kubernetes.io/projected/16490934-5213-4454-a1fb-3710756af56b-kube-api-access-l2wsd\") pod \"redis-6f5dff5888-l84dj\" (UID: \"16490934-5213-4454-a1fb-3710756af56b\") " pod="test-ns-feast/redis-6f5dff5888-l84dj" Feb 17 12:57:41.393625 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:57:41.393593 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2wsd\" (UniqueName: \"kubernetes.io/projected/16490934-5213-4454-a1fb-3710756af56b-kube-api-access-l2wsd\") pod \"redis-6f5dff5888-l84dj\" (UID: \"16490934-5213-4454-a1fb-3710756af56b\") " pod="test-ns-feast/redis-6f5dff5888-l84dj" Feb 17 12:57:41.451148 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:57:41.451114 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="test-ns-feast/redis-6f5dff5888-l84dj" Feb 17 12:57:41.563873 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:57:41.563833 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["test-ns-feast/redis-6f5dff5888-l84dj"] Feb 17 12:57:41.568706 ip-10-0-131-118 kubenswrapper[2572]: W0217 12:57:41.568677 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16490934_5213_4454_a1fb_3710756af56b.slice/crio-1c5b8dc974443e69ffe29205c6f984a239e1279ddc7daf96f2daa2f0f4e43fbc WatchSource:0}: Error finding container 1c5b8dc974443e69ffe29205c6f984a239e1279ddc7daf96f2daa2f0f4e43fbc: Status 404 returned error can't find the container with id 1c5b8dc974443e69ffe29205c6f984a239e1279ddc7daf96f2daa2f0f4e43fbc Feb 17 12:57:41.570842 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:57:41.570826 2572 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 17 12:57:42.084464 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:57:42.084433 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="test-ns-feast/redis-6f5dff5888-l84dj" event={"ID":"16490934-5213-4454-a1fb-3710756af56b","Type":"ContainerStarted","Data":"1c5b8dc974443e69ffe29205c6f984a239e1279ddc7daf96f2daa2f0f4e43fbc"} Feb 17 12:57:46.099099 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:57:46.099060 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="test-ns-feast/redis-6f5dff5888-l84dj" event={"ID":"16490934-5213-4454-a1fb-3710756af56b","Type":"ContainerStarted","Data":"513a8e08b1f9616bc438335c77ffd87f66000cb4b58223c4173dea6fcd556036"} Feb 17 12:57:46.112515 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:57:46.112462 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="test-ns-feast/redis-6f5dff5888-l84dj" podStartSLOduration=1.106576978 podStartE2EDuration="5.112446903s" podCreationTimestamp="2026-02-17 12:57:41 +0000 UTC" firstStartedPulling="2026-02-17 12:57:41.57094897 +0000 UTC m=+667.921359761" lastFinishedPulling="2026-02-17 12:57:45.576818896 +0000 UTC m=+671.927229686" observedRunningTime="2026-02-17 12:57:46.111340116 +0000 UTC m=+672.461750929" watchObservedRunningTime="2026-02-17 12:57:46.112446903 +0000 UTC m=+672.462857715" Feb 17 12:58:46.702507 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:58:46.702418 2572 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["test-ns-feast/redis-6f5dff5888-l84dj"] Feb 17 12:58:46.703069 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:58:46.702680 2572 kuberuntime_container.go:864] "Killing container with a grace period" pod="test-ns-feast/redis-6f5dff5888-l84dj" podUID="16490934-5213-4454-a1fb-3710756af56b" containerName="redis" containerID="cri-o://513a8e08b1f9616bc438335c77ffd87f66000cb4b58223c4173dea6fcd556036" gracePeriod=30 Feb 17 12:58:46.951396 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:58:46.951372 2572 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="test-ns-feast/redis-6f5dff5888-l84dj" Feb 17 12:58:47.032440 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:58:47.032346 2572 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2wsd\" (UniqueName: \"kubernetes.io/projected/16490934-5213-4454-a1fb-3710756af56b-kube-api-access-l2wsd\") pod \"16490934-5213-4454-a1fb-3710756af56b\" (UID: \"16490934-5213-4454-a1fb-3710756af56b\") " Feb 17 12:58:47.034480 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:58:47.034457 2572 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16490934-5213-4454-a1fb-3710756af56b-kube-api-access-l2wsd" (OuterVolumeSpecName: "kube-api-access-l2wsd") pod "16490934-5213-4454-a1fb-3710756af56b" (UID: "16490934-5213-4454-a1fb-3710756af56b"). InnerVolumeSpecName "kube-api-access-l2wsd". PluginName "kubernetes.io/projected", VolumeGIDValue "" Feb 17 12:58:47.133817 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:58:47.133700 2572 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-l2wsd\" (UniqueName: \"kubernetes.io/projected/16490934-5213-4454-a1fb-3710756af56b-kube-api-access-l2wsd\") on node \"ip-10-0-131-118.ec2.internal\" DevicePath \"\"" Feb 17 12:58:47.271937 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:58:47.271901 2572 generic.go:358] "Generic (PLEG): container finished" podID="16490934-5213-4454-a1fb-3710756af56b" containerID="513a8e08b1f9616bc438335c77ffd87f66000cb4b58223c4173dea6fcd556036" exitCode=0 Feb 17 12:58:47.272122 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:58:47.271960 2572 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="test-ns-feast/redis-6f5dff5888-l84dj" Feb 17 12:58:47.272122 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:58:47.271983 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="test-ns-feast/redis-6f5dff5888-l84dj" event={"ID":"16490934-5213-4454-a1fb-3710756af56b","Type":"ContainerDied","Data":"513a8e08b1f9616bc438335c77ffd87f66000cb4b58223c4173dea6fcd556036"} Feb 17 12:58:47.272122 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:58:47.272022 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="test-ns-feast/redis-6f5dff5888-l84dj" event={"ID":"16490934-5213-4454-a1fb-3710756af56b","Type":"ContainerDied","Data":"1c5b8dc974443e69ffe29205c6f984a239e1279ddc7daf96f2daa2f0f4e43fbc"} Feb 17 12:58:47.272122 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:58:47.272039 2572 scope.go:117] "RemoveContainer" containerID="513a8e08b1f9616bc438335c77ffd87f66000cb4b58223c4173dea6fcd556036" Feb 17 12:58:47.280781 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:58:47.280755 2572 scope.go:117] "RemoveContainer" containerID="513a8e08b1f9616bc438335c77ffd87f66000cb4b58223c4173dea6fcd556036" Feb 17 12:58:47.281083 ip-10-0-131-118 kubenswrapper[2572]: E0217 12:58:47.281060 2572 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"513a8e08b1f9616bc438335c77ffd87f66000cb4b58223c4173dea6fcd556036\": container with ID starting with 513a8e08b1f9616bc438335c77ffd87f66000cb4b58223c4173dea6fcd556036 not found: ID does not exist" containerID="513a8e08b1f9616bc438335c77ffd87f66000cb4b58223c4173dea6fcd556036" Feb 17 12:58:47.281190 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:58:47.281094 2572 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"513a8e08b1f9616bc438335c77ffd87f66000cb4b58223c4173dea6fcd556036"} err="failed to get container status \"513a8e08b1f9616bc438335c77ffd87f66000cb4b58223c4173dea6fcd556036\": rpc error: code = NotFound desc = could not find container \"513a8e08b1f9616bc438335c77ffd87f66000cb4b58223c4173dea6fcd556036\": container with ID starting with 513a8e08b1f9616bc438335c77ffd87f66000cb4b58223c4173dea6fcd556036 not found: ID does not exist" Feb 17 12:58:47.292002 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:58:47.291943 2572 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["test-ns-feast/redis-6f5dff5888-l84dj"] Feb 17 12:58:47.294124 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:58:47.294082 2572 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["test-ns-feast/redis-6f5dff5888-l84dj"] Feb 17 12:58:48.129650 ip-10-0-131-118 kubenswrapper[2572]: I0217 12:58:48.129610 2572 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16490934-5213-4454-a1fb-3710756af56b" path="/var/lib/kubelet/pods/16490934-5213-4454-a1fb-3710756af56b/volumes" Feb 17 13:01:34.140520 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:01:34.140447 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lb8m9_01eb7d37-a76a-4933-87df-44cf8a8d3152/ovn-acl-logging/0.log" Feb 17 13:01:34.141529 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:01:34.141505 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lb8m9_01eb7d37-a76a-4933-87df-44cf8a8d3152/ovn-acl-logging/0.log" Feb 17 13:06:34.162893 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:06:34.162865 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lb8m9_01eb7d37-a76a-4933-87df-44cf8a8d3152/ovn-acl-logging/0.log" Feb 17 13:06:34.163391 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:06:34.162872 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lb8m9_01eb7d37-a76a-4933-87df-44cf8a8d3152/ovn-acl-logging/0.log" Feb 17 13:09:32.732742 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:32.732694 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_global-pull-secret-syncer-b4nbx_fb40dc60-7177-4037-a5d5-7cb6434ea4f5/global-pull-secret-syncer/0.log" Feb 17 13:09:32.856577 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:32.856542 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_konnectivity-agent-kmsrd_30392cb9-7527-44af-833b-567cd508520e/konnectivity-agent/0.log" Feb 17 13:09:32.908481 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:32.908449 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_kube-apiserver-proxy-ip-10-0-131-118.ec2.internal_cde915e910705d98ba55e7806902bce3/haproxy/0.log" Feb 17 13:09:35.615900 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:35.615876 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_49f08cb5-241b-4020-87e6-08cfd5bec0cc/alertmanager/0.log" Feb 17 13:09:35.646863 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:35.646827 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_49f08cb5-241b-4020-87e6-08cfd5bec0cc/config-reloader/0.log" Feb 17 13:09:35.676615 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:35.676593 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_49f08cb5-241b-4020-87e6-08cfd5bec0cc/kube-rbac-proxy-web/0.log" Feb 17 13:09:35.705029 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:35.705006 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_49f08cb5-241b-4020-87e6-08cfd5bec0cc/kube-rbac-proxy/0.log" Feb 17 13:09:35.730939 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:35.730916 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_49f08cb5-241b-4020-87e6-08cfd5bec0cc/kube-rbac-proxy-metric/0.log" Feb 17 13:09:35.758423 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:35.758398 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_49f08cb5-241b-4020-87e6-08cfd5bec0cc/prom-label-proxy/0.log" Feb 17 13:09:35.785901 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:35.785872 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_49f08cb5-241b-4020-87e6-08cfd5bec0cc/init-config-reloader/0.log" Feb 17 13:09:35.943889 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:35.943862 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_metrics-server-86f7684bf5-m2r9q_2766fd9f-4249-4473-8285-5dcd03d2b37c/metrics-server/0.log" Feb 17 13:09:35.987550 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:35.987526 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_monitoring-plugin-69574b87bc-x78dp_b0db52e3-5296-4f94-95e3-b3867884108a/monitoring-plugin/0.log" Feb 17 13:09:36.034333 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:36.034300 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-6r5m2_c32ad535-2861-4d6f-83a7-8852563190a5/node-exporter/0.log" Feb 17 13:09:36.058895 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:36.058873 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-6r5m2_c32ad535-2861-4d6f-83a7-8852563190a5/kube-rbac-proxy/0.log" Feb 17 13:09:36.080822 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:36.080802 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-6r5m2_c32ad535-2861-4d6f-83a7-8852563190a5/init-textfile/0.log" Feb 17 13:09:36.336897 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:36.336872 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232/prometheus/0.log" Feb 17 13:09:36.359428 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:36.359397 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232/config-reloader/0.log" Feb 17 13:09:36.393313 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:36.393290 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232/thanos-sidecar/0.log" Feb 17 13:09:36.416324 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:36.416302 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232/kube-rbac-proxy-web/0.log" Feb 17 13:09:36.440760 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:36.440733 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232/kube-rbac-proxy/0.log" Feb 17 13:09:36.467521 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:36.467499 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232/kube-rbac-proxy-thanos/0.log" Feb 17 13:09:36.489828 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:36.489757 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_8fc65e67-6fd7-44fd-a8e8-ad1e3ada3232/init-config-reloader/0.log" Feb 17 13:09:36.518383 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:36.518347 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-operator-6567ccfb9-dxxms_e928aa3b-de14-4d54-aad7-4f8bb32ef910/prometheus-operator/0.log" Feb 17 13:09:36.538403 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:36.538377 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-operator-6567ccfb9-dxxms_e928aa3b-de14-4d54-aad7-4f8bb32ef910/kube-rbac-proxy/0.log" Feb 17 13:09:36.562858 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:36.562827 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-operator-admission-webhook-7b7b594589-sk6tq_37490693-a051-4312-8ffc-772adf00e51b/prometheus-operator-admission-webhook/0.log" Feb 17 13:09:39.482330 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.482300 2572 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-must-gather-7x4hp/perf-node-gather-daemonset-jrpkc"] Feb 17 13:09:39.482794 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.482607 2572 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="16490934-5213-4454-a1fb-3710756af56b" containerName="redis" Feb 17 13:09:39.482794 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.482619 2572 state_mem.go:107] "Deleted CPUSet assignment" podUID="16490934-5213-4454-a1fb-3710756af56b" containerName="redis" Feb 17 13:09:39.482794 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.482676 2572 memory_manager.go:356] "RemoveStaleState removing state" podUID="16490934-5213-4454-a1fb-3710756af56b" containerName="redis" Feb 17 13:09:39.485637 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.485621 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7x4hp/perf-node-gather-daemonset-jrpkc" Feb 17 13:09:39.487332 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.487308 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-7x4hp\"/\"kube-root-ca.crt\"" Feb 17 13:09:39.487451 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.487321 2572 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-7x4hp\"/\"openshift-service-ca.crt\"" Feb 17 13:09:39.487620 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.487606 2572 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-must-gather-7x4hp\"/\"default-dockercfg-vnmgq\"" Feb 17 13:09:39.493941 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.493917 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-7x4hp/perf-node-gather-daemonset-jrpkc"] Feb 17 13:09:39.550739 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.550702 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bc567b78-3827-479d-b473-4efea943c8ca-sys\") pod \"perf-node-gather-daemonset-jrpkc\" (UID: \"bc567b78-3827-479d-b473-4efea943c8ca\") " pod="openshift-must-gather-7x4hp/perf-node-gather-daemonset-jrpkc" Feb 17 13:09:39.550912 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.550749 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/bc567b78-3827-479d-b473-4efea943c8ca-podres\") pod \"perf-node-gather-daemonset-jrpkc\" (UID: \"bc567b78-3827-479d-b473-4efea943c8ca\") " pod="openshift-must-gather-7x4hp/perf-node-gather-daemonset-jrpkc" Feb 17 13:09:39.550912 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.550769 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/bc567b78-3827-479d-b473-4efea943c8ca-proc\") pod \"perf-node-gather-daemonset-jrpkc\" (UID: \"bc567b78-3827-479d-b473-4efea943c8ca\") " pod="openshift-must-gather-7x4hp/perf-node-gather-daemonset-jrpkc" Feb 17 13:09:39.550912 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.550788 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8tq7\" (UniqueName: \"kubernetes.io/projected/bc567b78-3827-479d-b473-4efea943c8ca-kube-api-access-q8tq7\") pod \"perf-node-gather-daemonset-jrpkc\" (UID: \"bc567b78-3827-479d-b473-4efea943c8ca\") " pod="openshift-must-gather-7x4hp/perf-node-gather-daemonset-jrpkc" Feb 17 13:09:39.550912 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.550835 2572 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bc567b78-3827-479d-b473-4efea943c8ca-lib-modules\") pod \"perf-node-gather-daemonset-jrpkc\" (UID: \"bc567b78-3827-479d-b473-4efea943c8ca\") " pod="openshift-must-gather-7x4hp/perf-node-gather-daemonset-jrpkc" Feb 17 13:09:39.613094 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.613072 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_dns-default-b6hth_a01f1f0a-fa4d-4b3c-a85e-053603fecfc1/dns/0.log" Feb 17 13:09:39.634199 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.634174 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_dns-default-b6hth_a01f1f0a-fa4d-4b3c-a85e-053603fecfc1/kube-rbac-proxy/0.log" Feb 17 13:09:39.651793 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.651763 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/bc567b78-3827-479d-b473-4efea943c8ca-proc\") pod \"perf-node-gather-daemonset-jrpkc\" (UID: \"bc567b78-3827-479d-b473-4efea943c8ca\") " pod="openshift-must-gather-7x4hp/perf-node-gather-daemonset-jrpkc" Feb 17 13:09:39.651972 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.651801 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-q8tq7\" (UniqueName: \"kubernetes.io/projected/bc567b78-3827-479d-b473-4efea943c8ca-kube-api-access-q8tq7\") pod \"perf-node-gather-daemonset-jrpkc\" (UID: \"bc567b78-3827-479d-b473-4efea943c8ca\") " pod="openshift-must-gather-7x4hp/perf-node-gather-daemonset-jrpkc" Feb 17 13:09:39.651972 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.651892 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/bc567b78-3827-479d-b473-4efea943c8ca-proc\") pod \"perf-node-gather-daemonset-jrpkc\" (UID: \"bc567b78-3827-479d-b473-4efea943c8ca\") " pod="openshift-must-gather-7x4hp/perf-node-gather-daemonset-jrpkc" Feb 17 13:09:39.652063 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.651975 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bc567b78-3827-479d-b473-4efea943c8ca-lib-modules\") pod \"perf-node-gather-daemonset-jrpkc\" (UID: \"bc567b78-3827-479d-b473-4efea943c8ca\") " pod="openshift-must-gather-7x4hp/perf-node-gather-daemonset-jrpkc" Feb 17 13:09:39.652063 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.652042 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bc567b78-3827-479d-b473-4efea943c8ca-sys\") pod \"perf-node-gather-daemonset-jrpkc\" (UID: \"bc567b78-3827-479d-b473-4efea943c8ca\") " pod="openshift-must-gather-7x4hp/perf-node-gather-daemonset-jrpkc" Feb 17 13:09:39.652175 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.652077 2572 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/bc567b78-3827-479d-b473-4efea943c8ca-podres\") pod \"perf-node-gather-daemonset-jrpkc\" (UID: \"bc567b78-3827-479d-b473-4efea943c8ca\") " pod="openshift-must-gather-7x4hp/perf-node-gather-daemonset-jrpkc" Feb 17 13:09:39.652175 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.652147 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bc567b78-3827-479d-b473-4efea943c8ca-lib-modules\") pod \"perf-node-gather-daemonset-jrpkc\" (UID: \"bc567b78-3827-479d-b473-4efea943c8ca\") " pod="openshift-must-gather-7x4hp/perf-node-gather-daemonset-jrpkc" Feb 17 13:09:39.652252 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.652176 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/bc567b78-3827-479d-b473-4efea943c8ca-sys\") pod \"perf-node-gather-daemonset-jrpkc\" (UID: \"bc567b78-3827-479d-b473-4efea943c8ca\") " pod="openshift-must-gather-7x4hp/perf-node-gather-daemonset-jrpkc" Feb 17 13:09:39.652252 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.652191 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/bc567b78-3827-479d-b473-4efea943c8ca-podres\") pod \"perf-node-gather-daemonset-jrpkc\" (UID: \"bc567b78-3827-479d-b473-4efea943c8ca\") " pod="openshift-must-gather-7x4hp/perf-node-gather-daemonset-jrpkc" Feb 17 13:09:39.658565 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.658543 2572 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8tq7\" (UniqueName: \"kubernetes.io/projected/bc567b78-3827-479d-b473-4efea943c8ca-kube-api-access-q8tq7\") pod \"perf-node-gather-daemonset-jrpkc\" (UID: \"bc567b78-3827-479d-b473-4efea943c8ca\") " pod="openshift-must-gather-7x4hp/perf-node-gather-daemonset-jrpkc" Feb 17 13:09:39.788756 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.788676 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_node-resolver-jqfj5_22f3475b-a5d2-4d66-818f-5010e57aac8e/dns-node-resolver/0.log" Feb 17 13:09:39.796420 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.796401 2572 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-7x4hp/perf-node-gather-daemonset-jrpkc" Feb 17 13:09:39.913765 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.913719 2572 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-7x4hp/perf-node-gather-daemonset-jrpkc"] Feb 17 13:09:39.916664 ip-10-0-131-118 kubenswrapper[2572]: W0217 13:09:39.916623 2572 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podbc567b78_3827_479d_b473_4efea943c8ca.slice/crio-271855d26736664d6713e7e713577e827d228eb39fa6551e2151b9697f000cc0 WatchSource:0}: Error finding container 271855d26736664d6713e7e713577e827d228eb39fa6551e2151b9697f000cc0: Status 404 returned error can't find the container with id 271855d26736664d6713e7e713577e827d228eb39fa6551e2151b9697f000cc0 Feb 17 13:09:39.918306 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:39.918288 2572 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 17 13:09:40.109917 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:40.109826 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7x4hp/perf-node-gather-daemonset-jrpkc" event={"ID":"bc567b78-3827-479d-b473-4efea943c8ca","Type":"ContainerStarted","Data":"2f97214c1745f397db92a0db9149e1aeed829aa971b9a5f832ec0d367343c086"} Feb 17 13:09:40.109917 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:40.109866 2572 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-7x4hp/perf-node-gather-daemonset-jrpkc" event={"ID":"bc567b78-3827-479d-b473-4efea943c8ca","Type":"ContainerStarted","Data":"271855d26736664d6713e7e713577e827d228eb39fa6551e2151b9697f000cc0"} Feb 17 13:09:40.110123 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:40.109981 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-must-gather-7x4hp/perf-node-gather-daemonset-jrpkc" Feb 17 13:09:40.122524 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:40.122483 2572 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-7x4hp/perf-node-gather-daemonset-jrpkc" podStartSLOduration=1.122470964 podStartE2EDuration="1.122470964s" podCreationTimestamp="2026-02-17 13:09:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-17 13:09:40.121466667 +0000 UTC m=+1386.471877490" watchObservedRunningTime="2026-02-17 13:09:40.122470964 +0000 UTC m=+1386.472881775" Feb 17 13:09:40.258609 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:40.258561 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_node-ca-qjw6p_8738b60c-920d-4bc5-933f-01d0fb8d68d3/node-ca/0.log" Feb 17 13:09:40.885613 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:40.885583 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_router-default-7764898fbb-x52mf_34367000-c224-43d7-9e8b-916d67860ee4/router/0.log" Feb 17 13:09:41.208946 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:41.208921 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-canary_ingress-canary-prcdr_51d19194-106c-4b3f-ba37-a0a7eedbc75a/serve-healthcheck-canary/0.log" Feb 17 13:09:41.606823 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:41.606747 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-hgpw7_72276741-8f08-4c0c-98ca-d1614d018232/kube-rbac-proxy/0.log" Feb 17 13:09:41.628195 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:41.628174 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-hgpw7_72276741-8f08-4c0c-98ca-d1614d018232/exporter/0.log" Feb 17 13:09:41.649123 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:41.649088 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-runtime-extractor-hgpw7_72276741-8f08-4c0c-98ca-d1614d018232/extractor/0.log" Feb 17 13:09:45.485259 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:45.485233 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator_migrator-75f6d74b65-gtbxg_3dfbc9a4-8a6f-4466-88ee-8a276d190f2e/migrator/0.log" Feb 17 13:09:45.505726 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:45.505708 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator_migrator-75f6d74b65-gtbxg_3dfbc9a4-8a6f-4466-88ee-8a276d190f2e/graceful-termination/0.log" Feb 17 13:09:46.121860 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:46.121835 2572 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-must-gather-7x4hp/perf-node-gather-daemonset-jrpkc" Feb 17 13:09:46.796443 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:46.796420 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-wfszv_9751ddaa-e412-4e4e-b97a-f4e5a7a050a9/kube-multus-additional-cni-plugins/0.log" Feb 17 13:09:46.816967 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:46.816941 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-wfszv_9751ddaa-e412-4e4e-b97a-f4e5a7a050a9/egress-router-binary-copy/0.log" Feb 17 13:09:46.837250 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:46.837229 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-wfszv_9751ddaa-e412-4e4e-b97a-f4e5a7a050a9/cni-plugins/0.log" Feb 17 13:09:46.856760 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:46.856738 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-wfszv_9751ddaa-e412-4e4e-b97a-f4e5a7a050a9/bond-cni-plugin/0.log" Feb 17 13:09:46.878478 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:46.878459 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-wfszv_9751ddaa-e412-4e4e-b97a-f4e5a7a050a9/routeoverride-cni/0.log" Feb 17 13:09:46.900017 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:46.899996 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-wfszv_9751ddaa-e412-4e4e-b97a-f4e5a7a050a9/whereabouts-cni-bincopy/0.log" Feb 17 13:09:46.920851 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:46.920833 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-additional-cni-plugins-wfszv_9751ddaa-e412-4e4e-b97a-f4e5a7a050a9/whereabouts-cni/0.log" Feb 17 13:09:46.974560 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:46.974531 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-tsb7s_30e805d5-0857-42ef-a870-6492d444e008/kube-multus/0.log" Feb 17 13:09:47.112025 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:47.111959 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-vmpwv_685523da-d834-463a-8a1a-ef70069c19c5/network-metrics-daemon/0.log" Feb 17 13:09:47.132432 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:47.132390 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_network-metrics-daemon-vmpwv_685523da-d834-463a-8a1a-ef70069c19c5/kube-rbac-proxy/0.log" Feb 17 13:09:48.238257 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:48.238227 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lb8m9_01eb7d37-a76a-4933-87df-44cf8a8d3152/ovn-controller/0.log" Feb 17 13:09:48.261215 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:48.261173 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lb8m9_01eb7d37-a76a-4933-87df-44cf8a8d3152/ovn-acl-logging/0.log" Feb 17 13:09:48.267175 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:48.267156 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lb8m9_01eb7d37-a76a-4933-87df-44cf8a8d3152/ovn-acl-logging/1.log" Feb 17 13:09:48.288361 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:48.288329 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lb8m9_01eb7d37-a76a-4933-87df-44cf8a8d3152/kube-rbac-proxy-node/0.log" Feb 17 13:09:48.312838 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:48.312811 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lb8m9_01eb7d37-a76a-4933-87df-44cf8a8d3152/kube-rbac-proxy-ovn-metrics/0.log" Feb 17 13:09:48.331278 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:48.331256 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lb8m9_01eb7d37-a76a-4933-87df-44cf8a8d3152/northd/0.log" Feb 17 13:09:48.353211 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:48.353142 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lb8m9_01eb7d37-a76a-4933-87df-44cf8a8d3152/nbdb/0.log" Feb 17 13:09:48.375065 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:48.375039 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lb8m9_01eb7d37-a76a-4933-87df-44cf8a8d3152/sbdb/0.log" Feb 17 13:09:48.487381 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:48.487357 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lb8m9_01eb7d37-a76a-4933-87df-44cf8a8d3152/ovnkube-controller/0.log" Feb 17 13:09:49.738820 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:49.738790 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-target-f2tlm_f1e10a0f-13dd-44a0-9d79-e6375e12ac24/network-check-target-container/0.log" Feb 17 13:09:50.709909 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:50.709885 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-operator_iptables-alerter-fm5zs_ba1edd82-a8f5-4e73-ac84-987cbbec0939/iptables-alerter/0.log" Feb 17 13:09:51.276149 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:51.276086 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-node-tuning-operator_tuned-48j6f_e35ae435-3708-4b40-93e1-40c8b34c3fa4/tuned/0.log" Feb 17 13:09:52.970824 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:52.970797 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-samples-operator_cluster-samples-operator-7bf5dd6d69-nwb99_1ac4bab6-4158-4d4e-91f1-ff2c43225f7d/cluster-samples-operator/0.log" Feb 17 13:09:52.987897 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:52.987874 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-samples-operator_cluster-samples-operator-7bf5dd6d69-nwb99_1ac4bab6-4158-4d4e-91f1-ff2c43225f7d/cluster-samples-operator-watch/0.log" Feb 17 13:09:54.480865 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:54.480840 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-csi-drivers_aws-ebs-csi-driver-node-ghtvv_3d7ab4ec-e6ac-454b-8724-36a22d414dae/csi-driver/0.log" Feb 17 13:09:54.501540 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:54.501515 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-csi-drivers_aws-ebs-csi-driver-node-ghtvv_3d7ab4ec-e6ac-454b-8724-36a22d414dae/csi-node-driver-registrar/0.log" Feb 17 13:09:54.521495 ip-10-0-131-118 kubenswrapper[2572]: I0217 13:09:54.521470 2572 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-csi-drivers_aws-ebs-csi-driver-node-ghtvv_3d7ab4ec-e6ac-454b-8724-36a22d414dae/csi-liveness-probe/0.log"